var/home/core/zuul-output/0000755000175000017500000000000015114722304014524 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015114734160015473 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005152664315114734152017714 0ustar rootrootDec 06 03:40:06 crc systemd[1]: Starting Kubernetes Kubelet... Dec 06 03:40:06 crc restorecon[4683]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:06 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 03:40:07 crc restorecon[4683]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 03:40:07 crc restorecon[4683]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 06 03:40:07 crc kubenswrapper[4802]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 06 03:40:07 crc kubenswrapper[4802]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 06 03:40:07 crc kubenswrapper[4802]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 06 03:40:07 crc kubenswrapper[4802]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 06 03:40:07 crc kubenswrapper[4802]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 06 03:40:07 crc kubenswrapper[4802]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.267148 4802 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.271836 4802 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.271858 4802 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.271864 4802 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.271871 4802 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.271877 4802 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.271883 4802 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.271902 4802 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.271909 4802 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.271916 4802 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.271932 4802 feature_gate.go:330] unrecognized feature gate: Example Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.271938 4802 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.271944 4802 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.271949 4802 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.271954 4802 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.271960 4802 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.271965 4802 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.271970 4802 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.271975 4802 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.271981 4802 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.271986 4802 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.271991 4802 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.271996 4802 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.272002 4802 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.272007 4802 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.272013 4802 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.272018 4802 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.272023 4802 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.272030 4802 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.272037 4802 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.272043 4802 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.272049 4802 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.272055 4802 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.272061 4802 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.272066 4802 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.272071 4802 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.272077 4802 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.272082 4802 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.272088 4802 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.272093 4802 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.272099 4802 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.272105 4802 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.272113 4802 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.272119 4802 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.272125 4802 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.272131 4802 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.272137 4802 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.272143 4802 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.272153 4802 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.272160 4802 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.272166 4802 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.272172 4802 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.272178 4802 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.272185 4802 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.272192 4802 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.272197 4802 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.272202 4802 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.272208 4802 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.272213 4802 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.272219 4802 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.272225 4802 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.272231 4802 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.272236 4802 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.272242 4802 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.272248 4802 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.272253 4802 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.272260 4802 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.272265 4802 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.272270 4802 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.272277 4802 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.272285 4802 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.272291 4802 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272401 4802 flags.go:64] FLAG: --address="0.0.0.0" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272412 4802 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272422 4802 flags.go:64] FLAG: --anonymous-auth="true" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272431 4802 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272438 4802 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272445 4802 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272454 4802 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272462 4802 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272469 4802 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272476 4802 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272482 4802 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272489 4802 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272496 4802 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272502 4802 flags.go:64] FLAG: --cgroup-root="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272508 4802 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272514 4802 flags.go:64] FLAG: --client-ca-file="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272520 4802 flags.go:64] FLAG: --cloud-config="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272526 4802 flags.go:64] FLAG: --cloud-provider="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272533 4802 flags.go:64] FLAG: --cluster-dns="[]" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272540 4802 flags.go:64] FLAG: --cluster-domain="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272546 4802 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272552 4802 flags.go:64] FLAG: --config-dir="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272557 4802 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272565 4802 flags.go:64] FLAG: --container-log-max-files="5" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272573 4802 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272579 4802 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272585 4802 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272591 4802 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272597 4802 flags.go:64] FLAG: --contention-profiling="false" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272604 4802 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272610 4802 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272617 4802 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272624 4802 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272631 4802 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272638 4802 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272644 4802 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272650 4802 flags.go:64] FLAG: --enable-load-reader="false" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272656 4802 flags.go:64] FLAG: --enable-server="true" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272662 4802 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272670 4802 flags.go:64] FLAG: --event-burst="100" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272676 4802 flags.go:64] FLAG: --event-qps="50" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272682 4802 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272693 4802 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272699 4802 flags.go:64] FLAG: --eviction-hard="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272708 4802 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272714 4802 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272721 4802 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272728 4802 flags.go:64] FLAG: --eviction-soft="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272735 4802 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272742 4802 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272770 4802 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272777 4802 flags.go:64] FLAG: --experimental-mounter-path="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272783 4802 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272789 4802 flags.go:64] FLAG: --fail-swap-on="true" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272795 4802 flags.go:64] FLAG: --feature-gates="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272803 4802 flags.go:64] FLAG: --file-check-frequency="20s" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272809 4802 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272815 4802 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272822 4802 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272828 4802 flags.go:64] FLAG: --healthz-port="10248" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272835 4802 flags.go:64] FLAG: --help="false" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272841 4802 flags.go:64] FLAG: --hostname-override="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272847 4802 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272854 4802 flags.go:64] FLAG: --http-check-frequency="20s" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272862 4802 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272868 4802 flags.go:64] FLAG: --image-credential-provider-config="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272875 4802 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272882 4802 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272888 4802 flags.go:64] FLAG: --image-service-endpoint="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272894 4802 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272900 4802 flags.go:64] FLAG: --kube-api-burst="100" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272906 4802 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272912 4802 flags.go:64] FLAG: --kube-api-qps="50" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272918 4802 flags.go:64] FLAG: --kube-reserved="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272925 4802 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272931 4802 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272938 4802 flags.go:64] FLAG: --kubelet-cgroups="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272944 4802 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272950 4802 flags.go:64] FLAG: --lock-file="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272956 4802 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272962 4802 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272969 4802 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272978 4802 flags.go:64] FLAG: --log-json-split-stream="false" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272984 4802 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272990 4802 flags.go:64] FLAG: --log-text-split-stream="false" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.272995 4802 flags.go:64] FLAG: --logging-format="text" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273002 4802 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273008 4802 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273014 4802 flags.go:64] FLAG: --manifest-url="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273020 4802 flags.go:64] FLAG: --manifest-url-header="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273028 4802 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273035 4802 flags.go:64] FLAG: --max-open-files="1000000" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273042 4802 flags.go:64] FLAG: --max-pods="110" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273048 4802 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273054 4802 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273060 4802 flags.go:64] FLAG: --memory-manager-policy="None" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273071 4802 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273077 4802 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273084 4802 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273090 4802 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273103 4802 flags.go:64] FLAG: --node-status-max-images="50" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273110 4802 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273116 4802 flags.go:64] FLAG: --oom-score-adj="-999" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273122 4802 flags.go:64] FLAG: --pod-cidr="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273128 4802 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273137 4802 flags.go:64] FLAG: --pod-manifest-path="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273143 4802 flags.go:64] FLAG: --pod-max-pids="-1" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273149 4802 flags.go:64] FLAG: --pods-per-core="0" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273155 4802 flags.go:64] FLAG: --port="10250" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273162 4802 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273168 4802 flags.go:64] FLAG: --provider-id="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273174 4802 flags.go:64] FLAG: --qos-reserved="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273180 4802 flags.go:64] FLAG: --read-only-port="10255" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273186 4802 flags.go:64] FLAG: --register-node="true" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273193 4802 flags.go:64] FLAG: --register-schedulable="true" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273199 4802 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273209 4802 flags.go:64] FLAG: --registry-burst="10" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273215 4802 flags.go:64] FLAG: --registry-qps="5" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273221 4802 flags.go:64] FLAG: --reserved-cpus="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273227 4802 flags.go:64] FLAG: --reserved-memory="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273235 4802 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273241 4802 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273247 4802 flags.go:64] FLAG: --rotate-certificates="false" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273253 4802 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273259 4802 flags.go:64] FLAG: --runonce="false" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273265 4802 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273271 4802 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273277 4802 flags.go:64] FLAG: --seccomp-default="false" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273287 4802 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273293 4802 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273299 4802 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273305 4802 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273311 4802 flags.go:64] FLAG: --storage-driver-password="root" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273318 4802 flags.go:64] FLAG: --storage-driver-secure="false" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273326 4802 flags.go:64] FLAG: --storage-driver-table="stats" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273332 4802 flags.go:64] FLAG: --storage-driver-user="root" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273338 4802 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273345 4802 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273353 4802 flags.go:64] FLAG: --system-cgroups="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273359 4802 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273368 4802 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273374 4802 flags.go:64] FLAG: --tls-cert-file="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273380 4802 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273388 4802 flags.go:64] FLAG: --tls-min-version="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273393 4802 flags.go:64] FLAG: --tls-private-key-file="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273399 4802 flags.go:64] FLAG: --topology-manager-policy="none" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273405 4802 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273411 4802 flags.go:64] FLAG: --topology-manager-scope="container" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273418 4802 flags.go:64] FLAG: --v="2" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273425 4802 flags.go:64] FLAG: --version="false" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273434 4802 flags.go:64] FLAG: --vmodule="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273441 4802 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.273448 4802 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273620 4802 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273628 4802 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273634 4802 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273639 4802 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273646 4802 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273653 4802 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273659 4802 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273672 4802 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273678 4802 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273684 4802 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273689 4802 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273694 4802 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273699 4802 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273704 4802 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273709 4802 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273715 4802 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273720 4802 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273725 4802 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273730 4802 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273735 4802 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273741 4802 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273746 4802 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273769 4802 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273774 4802 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273779 4802 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273785 4802 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273790 4802 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273795 4802 feature_gate.go:330] unrecognized feature gate: Example Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273800 4802 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273805 4802 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273811 4802 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273816 4802 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273821 4802 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273827 4802 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273833 4802 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273838 4802 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273843 4802 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273849 4802 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273855 4802 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273866 4802 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273873 4802 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273881 4802 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273887 4802 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273894 4802 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273900 4802 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273906 4802 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273912 4802 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273918 4802 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273924 4802 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273930 4802 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273936 4802 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273942 4802 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273948 4802 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273954 4802 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273960 4802 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273968 4802 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273975 4802 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273983 4802 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273991 4802 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.273997 4802 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.274004 4802 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.274012 4802 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.274021 4802 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.274028 4802 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.274035 4802 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.274042 4802 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.274047 4802 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.274054 4802 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.274060 4802 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.274069 4802 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.274076 4802 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.274092 4802 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.282745 4802 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.282799 4802 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.282892 4802 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.282902 4802 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.282907 4802 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.282911 4802 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.282915 4802 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.282919 4802 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.282925 4802 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.282932 4802 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.282937 4802 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.282942 4802 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.282946 4802 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.282950 4802 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.282955 4802 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.282959 4802 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.282963 4802 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.282967 4802 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.282971 4802 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.282975 4802 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.282979 4802 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.282982 4802 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.283015 4802 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.283018 4802 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.283023 4802 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.283027 4802 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.283030 4802 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.283036 4802 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.283041 4802 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.283046 4802 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.283051 4802 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.283054 4802 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.283058 4802 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.283061 4802 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.283065 4802 feature_gate.go:330] unrecognized feature gate: Example Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.283069 4802 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.283075 4802 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.283079 4802 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.283082 4802 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.283087 4802 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.283090 4802 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.283094 4802 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.283097 4802 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.283102 4802 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.283105 4802 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.283109 4802 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.283113 4802 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.283117 4802 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.283120 4802 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.283125 4802 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.283128 4802 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.283132 4802 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.283136 4802 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.283139 4802 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.283143 4802 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.283147 4802 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.283151 4802 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.283154 4802 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.283158 4802 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.283161 4802 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.283165 4802 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.283168 4802 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.283172 4802 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.283175 4802 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.283179 4802 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.283182 4802 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.283186 4802 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.283190 4802 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.283193 4802 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.283198 4802 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.283203 4802 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.283207 4802 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.283211 4802 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.283219 4802 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284137 4802 feature_gate.go:330] unrecognized feature gate: Example Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284172 4802 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284181 4802 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284187 4802 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284193 4802 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284198 4802 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284204 4802 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284209 4802 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284214 4802 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284219 4802 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284223 4802 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284229 4802 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284240 4802 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284245 4802 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284254 4802 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284260 4802 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284266 4802 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284272 4802 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284277 4802 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284282 4802 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284290 4802 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284296 4802 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284301 4802 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284306 4802 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284325 4802 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284330 4802 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284336 4802 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284341 4802 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284346 4802 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284352 4802 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284357 4802 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284361 4802 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284367 4802 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284373 4802 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284379 4802 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284384 4802 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284391 4802 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284399 4802 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284404 4802 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284409 4802 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284414 4802 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284418 4802 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284423 4802 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284428 4802 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284433 4802 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284439 4802 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284445 4802 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284450 4802 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284455 4802 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284464 4802 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284469 4802 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284475 4802 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284480 4802 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284486 4802 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284491 4802 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284496 4802 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284501 4802 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284506 4802 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284511 4802 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284517 4802 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284522 4802 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284528 4802 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284536 4802 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284542 4802 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284548 4802 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284555 4802 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284561 4802 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284567 4802 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284574 4802 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284581 4802 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.284586 4802 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.284597 4802 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.285808 4802 server.go:940] "Client rotation is on, will bootstrap in background" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.292467 4802 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.292671 4802 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.293551 4802 server.go:997] "Starting client certificate rotation" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.293599 4802 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.293789 4802 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-11 20:55:29.946614912 +0000 UTC Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.293908 4802 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.301412 4802 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 06 03:40:07 crc kubenswrapper[4802]: E1206 03:40:07.303388 4802 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.150:6443: connect: connection refused" logger="UnhandledError" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.304945 4802 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.314731 4802 log.go:25] "Validated CRI v1 runtime API" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.335847 4802 log.go:25] "Validated CRI v1 image API" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.337723 4802 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.340771 4802 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-06-03-35-38-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.340835 4802 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.359013 4802 manager.go:217] Machine: {Timestamp:2025-12-06 03:40:07.357582324 +0000 UTC m=+0.229491506 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654124544 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:c7c815e4-d2a4-4244-b035-f988fc95e215 BootID:0e01ca3a-758f-44bc-b4a2-30c1bc04fda0 Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:d5:10:b1 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:d5:10:b1 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:ec:d6:30 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:a7:d5:a2 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:4d:56:00 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:ea:97:ae Speed:-1 Mtu:1496} {Name:eth10 MacAddress:66:1a:6a:49:e4:6b Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:be:6f:7a:80:46:97 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654124544 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.359381 4802 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.359667 4802 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.360477 4802 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.360731 4802 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.360830 4802 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.361159 4802 topology_manager.go:138] "Creating topology manager with none policy" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.361176 4802 container_manager_linux.go:303] "Creating device plugin manager" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.361707 4802 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.361784 4802 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.362112 4802 state_mem.go:36] "Initialized new in-memory state store" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.362257 4802 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.363061 4802 kubelet.go:418] "Attempting to sync node with API server" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.363088 4802 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.363122 4802 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.363145 4802 kubelet.go:324] "Adding apiserver pod source" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.363163 4802 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.365660 4802 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.366380 4802 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.366360 4802 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.150:6443: connect: connection refused Dec 06 03:40:07 crc kubenswrapper[4802]: E1206 03:40:07.366461 4802 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.150:6443: connect: connection refused" logger="UnhandledError" Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.366447 4802 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.150:6443: connect: connection refused Dec 06 03:40:07 crc kubenswrapper[4802]: E1206 03:40:07.366531 4802 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.150:6443: connect: connection refused" logger="UnhandledError" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.367615 4802 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.368481 4802 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.368534 4802 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.368550 4802 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.368564 4802 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.368588 4802 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.368604 4802 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.368619 4802 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.368641 4802 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.368660 4802 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.368674 4802 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.368696 4802 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.368709 4802 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.369039 4802 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.369962 4802 server.go:1280] "Started kubelet" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.370800 4802 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.370901 4802 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.150:6443: connect: connection refused Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.370974 4802 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 06 03:40:07 crc systemd[1]: Started Kubernetes Kubelet. Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.372450 4802 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.373565 4802 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.374354 4802 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.374420 4802 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-02 20:27:29.040018797 +0000 UTC Dec 06 03:40:07 crc kubenswrapper[4802]: E1206 03:40:07.374558 4802 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.375592 4802 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.375669 4802 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.375737 4802 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 06 03:40:07 crc kubenswrapper[4802]: E1206 03:40:07.377452 4802 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.150:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187e833e408b45a6 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-06 03:40:07.369893286 +0000 UTC m=+0.241802478,LastTimestamp:2025-12-06 03:40:07.369893286 +0000 UTC m=+0.241802478,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 06 03:40:07 crc kubenswrapper[4802]: E1206 03:40:07.377013 4802 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.150:6443: connect: connection refused" interval="200ms" Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.379440 4802 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.150:6443: connect: connection refused Dec 06 03:40:07 crc kubenswrapper[4802]: E1206 03:40:07.379729 4802 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.150:6443: connect: connection refused" logger="UnhandledError" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.380224 4802 server.go:460] "Adding debug handlers to kubelet server" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.380384 4802 factory.go:55] Registering systemd factory Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.380415 4802 factory.go:221] Registration of the systemd container factory successfully Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.385725 4802 factory.go:153] Registering CRI-O factory Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.385825 4802 factory.go:221] Registration of the crio container factory successfully Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.385982 4802 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.386027 4802 factory.go:103] Registering Raw factory Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.386063 4802 manager.go:1196] Started watching for new ooms in manager Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.387505 4802 manager.go:319] Starting recovery of all containers Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.395477 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.395530 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.395547 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.395562 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.395575 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.395588 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.395604 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.395620 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.395639 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.395657 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.395672 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.395719 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.395732 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.395805 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.395823 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.395837 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.395857 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.395871 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.395887 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.395901 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.395914 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.395928 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.395943 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.395957 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.395971 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.395994 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.396011 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.396051 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.396067 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.396081 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.396098 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.396113 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.396137 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.396149 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.396163 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.396186 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.396202 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.396216 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.396231 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.396248 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.396262 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.396276 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.396294 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.396311 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.396333 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.396351 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.396370 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.396392 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.396417 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.396433 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.396458 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.396476 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.396511 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.396527 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.396546 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.396566 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.396588 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.396611 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.396631 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.396645 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.396662 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.396676 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.396691 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.396711 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.396733 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.396784 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.396802 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.396824 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.396837 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.396850 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.396927 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.397010 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.397025 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.397039 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.397054 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.398367 4802 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.398398 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.398414 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.398428 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.398442 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.398459 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.398472 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.398486 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.398497 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.398510 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.398523 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.398536 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.398551 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.398565 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.398578 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.398592 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.398605 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.398618 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.398631 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.398646 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.398660 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.398675 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.398689 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.398708 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.398721 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.398734 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.398781 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.398796 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.398810 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.398823 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.398844 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.398859 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.398876 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.398892 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.398913 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.398929 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.398949 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.398965 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.398980 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.398995 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.399010 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.399023 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.399044 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.399058 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.399074 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.399086 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.399107 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.399119 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.399134 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.399154 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.399167 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.399182 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.399198 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.399218 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.399232 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.399249 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.399263 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.399282 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.399296 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.399309 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.399322 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.399345 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.399358 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.399371 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.399385 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.399397 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.399411 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.399431 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.399444 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.399457 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.399470 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.399482 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.399503 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.399517 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.399536 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.399550 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.399565 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.399577 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.399590 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.399609 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.399621 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.399633 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.399646 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.399668 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.399686 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.399703 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.399723 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.399738 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.399800 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.399825 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.399846 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.399865 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.399878 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.399934 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.399956 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.399969 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.399989 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.400003 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.400016 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.400031 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.400044 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.400061 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.400075 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.400088 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.400101 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.400121 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.400135 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.400154 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.400171 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.400184 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.400199 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.400212 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.400230 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.400244 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.400257 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.400272 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.400285 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.400301 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.400314 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.400327 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.400345 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.400357 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.400369 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.400383 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.400402 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.400425 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.400443 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.400456 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.400475 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.400488 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.400505 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.400518 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.400531 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.400544 4802 reconstruct.go:97] "Volume reconstruction finished" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.400554 4802 reconciler.go:26] "Reconciler: start to sync state" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.412402 4802 manager.go:324] Recovery completed Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.424527 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.426861 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.427014 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.427100 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.429469 4802 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.429565 4802 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.429662 4802 state_mem.go:36] "Initialized new in-memory state store" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.440857 4802 policy_none.go:49] "None policy: Start" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.442840 4802 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.442904 4802 state_mem.go:35] "Initializing new in-memory state store" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.446200 4802 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.448688 4802 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.448783 4802 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.448826 4802 kubelet.go:2335] "Starting kubelet main sync loop" Dec 06 03:40:07 crc kubenswrapper[4802]: E1206 03:40:07.448890 4802 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.450989 4802 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.150:6443: connect: connection refused Dec 06 03:40:07 crc kubenswrapper[4802]: E1206 03:40:07.451098 4802 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.150:6443: connect: connection refused" logger="UnhandledError" Dec 06 03:40:07 crc kubenswrapper[4802]: E1206 03:40:07.474794 4802 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.512285 4802 manager.go:334] "Starting Device Plugin manager" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.512399 4802 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.512443 4802 server.go:79] "Starting device plugin registration server" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.513198 4802 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.513221 4802 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.513860 4802 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.513962 4802 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.513976 4802 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 06 03:40:07 crc kubenswrapper[4802]: E1206 03:40:07.527487 4802 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.550006 4802 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.550175 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.551730 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.551822 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.551837 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.552251 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.552530 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.552568 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.553516 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.553542 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.553554 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.553684 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.553804 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.553830 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.553855 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.553877 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.553862 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.554517 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.554552 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.554566 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.554732 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.554944 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.555009 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.556222 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.556256 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.556268 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.556528 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.556556 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.556567 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.556674 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.556884 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.556957 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.556970 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.557047 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.556996 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.557634 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.557673 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.557691 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.557958 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.558007 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.558118 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.558178 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.558203 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.559246 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.559312 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.559335 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:07 crc kubenswrapper[4802]: E1206 03:40:07.581047 4802 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.150:6443: connect: connection refused" interval="400ms" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.605436 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.605527 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.605565 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.605675 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.605982 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.606010 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.606080 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.606108 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.606130 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.606158 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.606246 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.606364 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.606419 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.606516 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.606610 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.613387 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.615198 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.615243 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.615254 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.615287 4802 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 06 03:40:07 crc kubenswrapper[4802]: E1206 03:40:07.615663 4802 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.150:6443: connect: connection refused" node="crc" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.708287 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.708353 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.708393 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.708428 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.708468 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.708500 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.708531 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.708556 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.708567 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.708597 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.708615 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.708519 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.708700 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.708841 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.708858 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.708872 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.708879 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.708893 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.708941 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.708939 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.708940 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.708938 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.708965 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.709011 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.709135 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.709194 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.709202 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.709242 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.709326 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.709387 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.816662 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.818560 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.818638 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.818657 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.818714 4802 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 06 03:40:07 crc kubenswrapper[4802]: E1206 03:40:07.819350 4802 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.150:6443: connect: connection refused" node="crc" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.908554 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.933203 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.939984 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.948779 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-8abc71477ae008d8e07c9306ab6649d96e9fc48ee2df46a653f67108ace32201 WatchSource:0}: Error finding container 8abc71477ae008d8e07c9306ab6649d96e9fc48ee2df46a653f67108ace32201: Status 404 returned error can't find the container with id 8abc71477ae008d8e07c9306ab6649d96e9fc48ee2df46a653f67108ace32201 Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.957338 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 03:40:07 crc kubenswrapper[4802]: I1206 03:40:07.962688 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.963684 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-40bd783864d57afa9c90fe445811b35dfcf9e1ab8da849c53b5bcb37ef58d691 WatchSource:0}: Error finding container 40bd783864d57afa9c90fe445811b35dfcf9e1ab8da849c53b5bcb37ef58d691: Status 404 returned error can't find the container with id 40bd783864d57afa9c90fe445811b35dfcf9e1ab8da849c53b5bcb37ef58d691 Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.964840 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-85038e49ae134abd6863213a301321f9cfdc6ee76482529ff81ae788154bbfd0 WatchSource:0}: Error finding container 85038e49ae134abd6863213a301321f9cfdc6ee76482529ff81ae788154bbfd0: Status 404 returned error can't find the container with id 85038e49ae134abd6863213a301321f9cfdc6ee76482529ff81ae788154bbfd0 Dec 06 03:40:07 crc kubenswrapper[4802]: E1206 03:40:07.982368 4802 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.150:6443: connect: connection refused" interval="800ms" Dec 06 03:40:07 crc kubenswrapper[4802]: W1206 03:40:07.998534 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-f0e606b33eff31df8966b0bb885078430f80db8ba202976207945cde4b478b47 WatchSource:0}: Error finding container f0e606b33eff31df8966b0bb885078430f80db8ba202976207945cde4b478b47: Status 404 returned error can't find the container with id f0e606b33eff31df8966b0bb885078430f80db8ba202976207945cde4b478b47 Dec 06 03:40:08 crc kubenswrapper[4802]: W1206 03:40:08.000848 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-1fb90ffbf249a4a6d27ae776a91358222d81a6965ef11f0d55613dcb31f4f0e3 WatchSource:0}: Error finding container 1fb90ffbf249a4a6d27ae776a91358222d81a6965ef11f0d55613dcb31f4f0e3: Status 404 returned error can't find the container with id 1fb90ffbf249a4a6d27ae776a91358222d81a6965ef11f0d55613dcb31f4f0e3 Dec 06 03:40:08 crc kubenswrapper[4802]: I1206 03:40:08.219877 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:40:08 crc kubenswrapper[4802]: I1206 03:40:08.221226 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:08 crc kubenswrapper[4802]: I1206 03:40:08.221280 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:08 crc kubenswrapper[4802]: I1206 03:40:08.221293 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:08 crc kubenswrapper[4802]: I1206 03:40:08.221324 4802 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 06 03:40:08 crc kubenswrapper[4802]: E1206 03:40:08.221682 4802 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.150:6443: connect: connection refused" node="crc" Dec 06 03:40:08 crc kubenswrapper[4802]: W1206 03:40:08.240441 4802 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.150:6443: connect: connection refused Dec 06 03:40:08 crc kubenswrapper[4802]: E1206 03:40:08.240552 4802 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.150:6443: connect: connection refused" logger="UnhandledError" Dec 06 03:40:08 crc kubenswrapper[4802]: I1206 03:40:08.372454 4802 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.150:6443: connect: connection refused Dec 06 03:40:08 crc kubenswrapper[4802]: I1206 03:40:08.375550 4802 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-26 01:37:56.188330008 +0000 UTC Dec 06 03:40:08 crc kubenswrapper[4802]: I1206 03:40:08.375609 4802 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 477h57m47.812724976s for next certificate rotation Dec 06 03:40:08 crc kubenswrapper[4802]: I1206 03:40:08.456422 4802 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="adad8cbf6f6508ca0ad781d9e59b9d90566c02acaeb3a4b3208725c75bfda226" exitCode=0 Dec 06 03:40:08 crc kubenswrapper[4802]: I1206 03:40:08.456572 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"adad8cbf6f6508ca0ad781d9e59b9d90566c02acaeb3a4b3208725c75bfda226"} Dec 06 03:40:08 crc kubenswrapper[4802]: I1206 03:40:08.456780 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"40bd783864d57afa9c90fe445811b35dfcf9e1ab8da849c53b5bcb37ef58d691"} Dec 06 03:40:08 crc kubenswrapper[4802]: I1206 03:40:08.457000 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:40:08 crc kubenswrapper[4802]: I1206 03:40:08.458443 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:08 crc kubenswrapper[4802]: I1206 03:40:08.458493 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:08 crc kubenswrapper[4802]: I1206 03:40:08.458514 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:08 crc kubenswrapper[4802]: I1206 03:40:08.459121 4802 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="18a08a15442f04e37a5aa71afbcc293e80b876477d60cbe71d562c75d141a9ba" exitCode=0 Dec 06 03:40:08 crc kubenswrapper[4802]: I1206 03:40:08.459192 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"18a08a15442f04e37a5aa71afbcc293e80b876477d60cbe71d562c75d141a9ba"} Dec 06 03:40:08 crc kubenswrapper[4802]: I1206 03:40:08.459248 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"8abc71477ae008d8e07c9306ab6649d96e9fc48ee2df46a653f67108ace32201"} Dec 06 03:40:08 crc kubenswrapper[4802]: I1206 03:40:08.459360 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:40:08 crc kubenswrapper[4802]: I1206 03:40:08.460461 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:08 crc kubenswrapper[4802]: I1206 03:40:08.460492 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:08 crc kubenswrapper[4802]: I1206 03:40:08.460509 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:08 crc kubenswrapper[4802]: I1206 03:40:08.462383 4802 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="c2d359853c738e3e0e264f1a6a919b17afc71a362821b3ea170cda6695e3593f" exitCode=0 Dec 06 03:40:08 crc kubenswrapper[4802]: I1206 03:40:08.462488 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"c2d359853c738e3e0e264f1a6a919b17afc71a362821b3ea170cda6695e3593f"} Dec 06 03:40:08 crc kubenswrapper[4802]: I1206 03:40:08.462593 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"1fb90ffbf249a4a6d27ae776a91358222d81a6965ef11f0d55613dcb31f4f0e3"} Dec 06 03:40:08 crc kubenswrapper[4802]: I1206 03:40:08.462737 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:40:08 crc kubenswrapper[4802]: I1206 03:40:08.463959 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:08 crc kubenswrapper[4802]: I1206 03:40:08.464010 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:08 crc kubenswrapper[4802]: I1206 03:40:08.464024 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:08 crc kubenswrapper[4802]: I1206 03:40:08.466081 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"03004a29aefe7fe2452fb9f6663c0807afccb055d7ddbf6b907f305e6a19c81c"} Dec 06 03:40:08 crc kubenswrapper[4802]: I1206 03:40:08.466135 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"f0e606b33eff31df8966b0bb885078430f80db8ba202976207945cde4b478b47"} Dec 06 03:40:08 crc kubenswrapper[4802]: I1206 03:40:08.469162 4802 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce" exitCode=0 Dec 06 03:40:08 crc kubenswrapper[4802]: I1206 03:40:08.469212 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce"} Dec 06 03:40:08 crc kubenswrapper[4802]: I1206 03:40:08.469242 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"85038e49ae134abd6863213a301321f9cfdc6ee76482529ff81ae788154bbfd0"} Dec 06 03:40:08 crc kubenswrapper[4802]: I1206 03:40:08.469404 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:40:08 crc kubenswrapper[4802]: I1206 03:40:08.470666 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:08 crc kubenswrapper[4802]: I1206 03:40:08.470703 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:08 crc kubenswrapper[4802]: I1206 03:40:08.470713 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:08 crc kubenswrapper[4802]: I1206 03:40:08.472154 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:40:08 crc kubenswrapper[4802]: I1206 03:40:08.473141 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:08 crc kubenswrapper[4802]: I1206 03:40:08.473170 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:08 crc kubenswrapper[4802]: I1206 03:40:08.473182 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:08 crc kubenswrapper[4802]: W1206 03:40:08.706447 4802 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.150:6443: connect: connection refused Dec 06 03:40:08 crc kubenswrapper[4802]: E1206 03:40:08.706533 4802 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.150:6443: connect: connection refused" logger="UnhandledError" Dec 06 03:40:08 crc kubenswrapper[4802]: W1206 03:40:08.710224 4802 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.150:6443: connect: connection refused Dec 06 03:40:08 crc kubenswrapper[4802]: E1206 03:40:08.710311 4802 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.150:6443: connect: connection refused" logger="UnhandledError" Dec 06 03:40:08 crc kubenswrapper[4802]: E1206 03:40:08.784085 4802 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.150:6443: connect: connection refused" interval="1.6s" Dec 06 03:40:08 crc kubenswrapper[4802]: E1206 03:40:08.821282 4802 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.150:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187e833e408b45a6 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-06 03:40:07.369893286 +0000 UTC m=+0.241802478,LastTimestamp:2025-12-06 03:40:07.369893286 +0000 UTC m=+0.241802478,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 06 03:40:09 crc kubenswrapper[4802]: I1206 03:40:09.022448 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:40:09 crc kubenswrapper[4802]: I1206 03:40:09.024581 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:09 crc kubenswrapper[4802]: I1206 03:40:09.024784 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:09 crc kubenswrapper[4802]: I1206 03:40:09.024808 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:09 crc kubenswrapper[4802]: I1206 03:40:09.024866 4802 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 06 03:40:09 crc kubenswrapper[4802]: E1206 03:40:09.025490 4802 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.150:6443: connect: connection refused" node="crc" Dec 06 03:40:09 crc kubenswrapper[4802]: W1206 03:40:09.039207 4802 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.150:6443: connect: connection refused Dec 06 03:40:09 crc kubenswrapper[4802]: E1206 03:40:09.039295 4802 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.150:6443: connect: connection refused" logger="UnhandledError" Dec 06 03:40:09 crc kubenswrapper[4802]: I1206 03:40:09.325348 4802 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 06 03:40:09 crc kubenswrapper[4802]: E1206 03:40:09.328089 4802 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.150:6443: connect: connection refused" logger="UnhandledError" Dec 06 03:40:09 crc kubenswrapper[4802]: I1206 03:40:09.371946 4802 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.150:6443: connect: connection refused Dec 06 03:40:09 crc kubenswrapper[4802]: I1206 03:40:09.474290 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"16a92d501c9e626fe91e4a6793e426d10eb292cfbab766db8b8cc72e5dae94ef"} Dec 06 03:40:09 crc kubenswrapper[4802]: I1206 03:40:09.474333 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"9a21a97385f856ec996105b597d0ae81c56ed28b625ef3f2905e1af6ec8f6a45"} Dec 06 03:40:09 crc kubenswrapper[4802]: I1206 03:40:09.474346 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"646e01a58908ba2e703aae5eef53e1b981b0f0c3d1d3202487f2a8b94ee3f55d"} Dec 06 03:40:09 crc kubenswrapper[4802]: I1206 03:40:09.474358 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a895e0cc04880c3287ba725ac5484835235f1e066a94ce98fcf4b6133c5cbd2f"} Dec 06 03:40:09 crc kubenswrapper[4802]: I1206 03:40:09.476192 4802 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="84ce35f5ad94653c857069ade3fcdf326b1941c3b4fae4b75a695a5b1917f9aa" exitCode=0 Dec 06 03:40:09 crc kubenswrapper[4802]: I1206 03:40:09.476274 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"84ce35f5ad94653c857069ade3fcdf326b1941c3b4fae4b75a695a5b1917f9aa"} Dec 06 03:40:09 crc kubenswrapper[4802]: I1206 03:40:09.476492 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:40:09 crc kubenswrapper[4802]: I1206 03:40:09.477424 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:09 crc kubenswrapper[4802]: I1206 03:40:09.477463 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:09 crc kubenswrapper[4802]: I1206 03:40:09.477475 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:09 crc kubenswrapper[4802]: I1206 03:40:09.480061 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"990943a2a48d79458297f4e5cf4178bbd6e9eafbcadc154a803846c589474e85"} Dec 06 03:40:09 crc kubenswrapper[4802]: I1206 03:40:09.480157 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:40:09 crc kubenswrapper[4802]: I1206 03:40:09.481379 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:09 crc kubenswrapper[4802]: I1206 03:40:09.481406 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:09 crc kubenswrapper[4802]: I1206 03:40:09.481417 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:09 crc kubenswrapper[4802]: I1206 03:40:09.487027 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"e9e6b48b5701dd330ef91712cdcd160afe9831a3881038bfb1e0397a9c8a3aa4"} Dec 06 03:40:09 crc kubenswrapper[4802]: I1206 03:40:09.487099 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"af83814e6b2c5ed803157308ab895551e0826fd02f1ecc8b2751145c09d359da"} Dec 06 03:40:09 crc kubenswrapper[4802]: I1206 03:40:09.487111 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"ce321001135d0a2b878316f8003bf1bc54cba9a36f6a654ba376e1e7c0b4da71"} Dec 06 03:40:09 crc kubenswrapper[4802]: I1206 03:40:09.487251 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:40:09 crc kubenswrapper[4802]: I1206 03:40:09.488201 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:09 crc kubenswrapper[4802]: I1206 03:40:09.488248 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:09 crc kubenswrapper[4802]: I1206 03:40:09.488263 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:09 crc kubenswrapper[4802]: I1206 03:40:09.489077 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"c6e454507cadbb982a566a3692995aaa3c733ff1cd7fb4870476602190e1eeba"} Dec 06 03:40:09 crc kubenswrapper[4802]: I1206 03:40:09.489103 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"1801ac64db4ba91c2e8a249e3651d3e50ab62e9f2c87312ce9bf7f38518a82cd"} Dec 06 03:40:09 crc kubenswrapper[4802]: I1206 03:40:09.489116 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"8f0d6e582161f0d9126126e03b27211be22f6b5f797154928c92fa13a8118f52"} Dec 06 03:40:09 crc kubenswrapper[4802]: I1206 03:40:09.489168 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:40:09 crc kubenswrapper[4802]: I1206 03:40:09.489993 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:09 crc kubenswrapper[4802]: I1206 03:40:09.490015 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:09 crc kubenswrapper[4802]: I1206 03:40:09.490023 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:09 crc kubenswrapper[4802]: I1206 03:40:09.960104 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 03:40:09 crc kubenswrapper[4802]: I1206 03:40:09.964450 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 03:40:10 crc kubenswrapper[4802]: I1206 03:40:10.495644 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b602a9bcbfd0f984f1778c0fcdf44058ce528b0d21af6563e76be589971384b5"} Dec 06 03:40:10 crc kubenswrapper[4802]: I1206 03:40:10.495841 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:40:10 crc kubenswrapper[4802]: I1206 03:40:10.497462 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:10 crc kubenswrapper[4802]: I1206 03:40:10.497514 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:10 crc kubenswrapper[4802]: I1206 03:40:10.497533 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:10 crc kubenswrapper[4802]: I1206 03:40:10.498509 4802 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="17f78cae7920f1f196ec9baf053ff98df6935df8078f8114a2b443aa2111d50f" exitCode=0 Dec 06 03:40:10 crc kubenswrapper[4802]: I1206 03:40:10.498640 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"17f78cae7920f1f196ec9baf053ff98df6935df8078f8114a2b443aa2111d50f"} Dec 06 03:40:10 crc kubenswrapper[4802]: I1206 03:40:10.498700 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:40:10 crc kubenswrapper[4802]: I1206 03:40:10.498930 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:40:10 crc kubenswrapper[4802]: I1206 03:40:10.500098 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:10 crc kubenswrapper[4802]: I1206 03:40:10.500160 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:10 crc kubenswrapper[4802]: I1206 03:40:10.500178 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:10 crc kubenswrapper[4802]: I1206 03:40:10.500255 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:10 crc kubenswrapper[4802]: I1206 03:40:10.500296 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:10 crc kubenswrapper[4802]: I1206 03:40:10.500316 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:10 crc kubenswrapper[4802]: I1206 03:40:10.626520 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:40:10 crc kubenswrapper[4802]: I1206 03:40:10.627829 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:10 crc kubenswrapper[4802]: I1206 03:40:10.627864 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:10 crc kubenswrapper[4802]: I1206 03:40:10.627876 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:10 crc kubenswrapper[4802]: I1206 03:40:10.627900 4802 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 06 03:40:10 crc kubenswrapper[4802]: I1206 03:40:10.976638 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 03:40:11 crc kubenswrapper[4802]: I1206 03:40:11.147980 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:40:11 crc kubenswrapper[4802]: I1206 03:40:11.508031 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"648ea0415efa4e81ad24f8af337ed2c8c0ba0a041aaf3ad7bb4e471eef14a877"} Dec 06 03:40:11 crc kubenswrapper[4802]: I1206 03:40:11.508102 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"a7ee33d59413614773a292125b39f656b7a5962fa254a217966e625bd239ba7c"} Dec 06 03:40:11 crc kubenswrapper[4802]: I1206 03:40:11.508125 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"4aa3d0fabd0ec4495a3b66418d59d266828d55d591782797864a9a0b261dd9a0"} Dec 06 03:40:11 crc kubenswrapper[4802]: I1206 03:40:11.508129 4802 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 03:40:11 crc kubenswrapper[4802]: I1206 03:40:11.508176 4802 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 03:40:11 crc kubenswrapper[4802]: I1206 03:40:11.508221 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:40:11 crc kubenswrapper[4802]: I1206 03:40:11.508240 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:40:11 crc kubenswrapper[4802]: I1206 03:40:11.508143 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"2ab8a6728c484ad31865d6c09bc04fb41e5cc612d64b7d90e6b4edb4fa317236"} Dec 06 03:40:11 crc kubenswrapper[4802]: I1206 03:40:11.509702 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:11 crc kubenswrapper[4802]: I1206 03:40:11.509791 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:11 crc kubenswrapper[4802]: I1206 03:40:11.509811 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:11 crc kubenswrapper[4802]: I1206 03:40:11.509975 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:11 crc kubenswrapper[4802]: I1206 03:40:11.510050 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:11 crc kubenswrapper[4802]: I1206 03:40:11.510069 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:12 crc kubenswrapper[4802]: I1206 03:40:12.033675 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 03:40:12 crc kubenswrapper[4802]: I1206 03:40:12.033997 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:40:12 crc kubenswrapper[4802]: I1206 03:40:12.035502 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:12 crc kubenswrapper[4802]: I1206 03:40:12.035562 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:12 crc kubenswrapper[4802]: I1206 03:40:12.035588 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:12 crc kubenswrapper[4802]: I1206 03:40:12.511473 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:40:12 crc kubenswrapper[4802]: I1206 03:40:12.516579 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"4fc93fc85f068f1dc9e3c410331b403e582a0cce3828d53e4887ada0bdd0a77e"} Dec 06 03:40:12 crc kubenswrapper[4802]: I1206 03:40:12.516647 4802 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 03:40:12 crc kubenswrapper[4802]: I1206 03:40:12.516709 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:40:12 crc kubenswrapper[4802]: I1206 03:40:12.516747 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:40:12 crc kubenswrapper[4802]: I1206 03:40:12.516799 4802 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 03:40:12 crc kubenswrapper[4802]: I1206 03:40:12.516858 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:40:12 crc kubenswrapper[4802]: I1206 03:40:12.519276 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:12 crc kubenswrapper[4802]: I1206 03:40:12.519316 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:12 crc kubenswrapper[4802]: I1206 03:40:12.519387 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:12 crc kubenswrapper[4802]: I1206 03:40:12.519416 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:12 crc kubenswrapper[4802]: I1206 03:40:12.519349 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:12 crc kubenswrapper[4802]: I1206 03:40:12.519587 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:12 crc kubenswrapper[4802]: I1206 03:40:12.522981 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:12 crc kubenswrapper[4802]: I1206 03:40:12.523701 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:12 crc kubenswrapper[4802]: I1206 03:40:12.523724 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:13 crc kubenswrapper[4802]: I1206 03:40:13.467067 4802 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 06 03:40:13 crc kubenswrapper[4802]: I1206 03:40:13.508239 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 03:40:13 crc kubenswrapper[4802]: I1206 03:40:13.511367 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:40:13 crc kubenswrapper[4802]: I1206 03:40:13.518561 4802 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 03:40:13 crc kubenswrapper[4802]: I1206 03:40:13.518606 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:40:13 crc kubenswrapper[4802]: I1206 03:40:13.518720 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:40:13 crc kubenswrapper[4802]: I1206 03:40:13.518879 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:40:13 crc kubenswrapper[4802]: I1206 03:40:13.519587 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:13 crc kubenswrapper[4802]: I1206 03:40:13.519685 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:13 crc kubenswrapper[4802]: I1206 03:40:13.519855 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:13 crc kubenswrapper[4802]: I1206 03:40:13.520364 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:13 crc kubenswrapper[4802]: I1206 03:40:13.520388 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:13 crc kubenswrapper[4802]: I1206 03:40:13.520396 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:13 crc kubenswrapper[4802]: I1206 03:40:13.520500 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:13 crc kubenswrapper[4802]: I1206 03:40:13.520551 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:13 crc kubenswrapper[4802]: I1206 03:40:13.520570 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:13 crc kubenswrapper[4802]: I1206 03:40:13.977242 4802 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 03:40:13 crc kubenswrapper[4802]: I1206 03:40:13.977312 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 03:40:14 crc kubenswrapper[4802]: I1206 03:40:14.164951 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 06 03:40:14 crc kubenswrapper[4802]: I1206 03:40:14.185472 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 06 03:40:14 crc kubenswrapper[4802]: I1206 03:40:14.521123 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:40:14 crc kubenswrapper[4802]: I1206 03:40:14.521130 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:40:14 crc kubenswrapper[4802]: I1206 03:40:14.522163 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:14 crc kubenswrapper[4802]: I1206 03:40:14.522213 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:14 crc kubenswrapper[4802]: I1206 03:40:14.522233 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:14 crc kubenswrapper[4802]: I1206 03:40:14.522528 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:14 crc kubenswrapper[4802]: I1206 03:40:14.522609 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:14 crc kubenswrapper[4802]: I1206 03:40:14.522636 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:15 crc kubenswrapper[4802]: I1206 03:40:15.523440 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:40:15 crc kubenswrapper[4802]: I1206 03:40:15.524414 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:15 crc kubenswrapper[4802]: I1206 03:40:15.524493 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:15 crc kubenswrapper[4802]: I1206 03:40:15.524507 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:16 crc kubenswrapper[4802]: I1206 03:40:16.668570 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 03:40:16 crc kubenswrapper[4802]: I1206 03:40:16.668844 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:40:16 crc kubenswrapper[4802]: I1206 03:40:16.670440 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:16 crc kubenswrapper[4802]: I1206 03:40:16.670509 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:16 crc kubenswrapper[4802]: I1206 03:40:16.670532 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:17 crc kubenswrapper[4802]: E1206 03:40:17.527833 4802 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 06 03:40:20 crc kubenswrapper[4802]: I1206 03:40:20.372977 4802 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Dec 06 03:40:20 crc kubenswrapper[4802]: E1206 03:40:20.385412 4802 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" interval="3.2s" Dec 06 03:40:20 crc kubenswrapper[4802]: I1206 03:40:20.526264 4802 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 06 03:40:20 crc kubenswrapper[4802]: I1206 03:40:20.526340 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 06 03:40:20 crc kubenswrapper[4802]: I1206 03:40:20.540570 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 06 03:40:20 crc kubenswrapper[4802]: I1206 03:40:20.542856 4802 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 06 03:40:20 crc kubenswrapper[4802]: I1206 03:40:20.542916 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 06 03:40:20 crc kubenswrapper[4802]: I1206 03:40:20.543770 4802 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="b602a9bcbfd0f984f1778c0fcdf44058ce528b0d21af6563e76be589971384b5" exitCode=255 Dec 06 03:40:20 crc kubenswrapper[4802]: I1206 03:40:20.543804 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"b602a9bcbfd0f984f1778c0fcdf44058ce528b0d21af6563e76be589971384b5"} Dec 06 03:40:20 crc kubenswrapper[4802]: I1206 03:40:20.543963 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:40:20 crc kubenswrapper[4802]: I1206 03:40:20.545116 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:20 crc kubenswrapper[4802]: I1206 03:40:20.545201 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:20 crc kubenswrapper[4802]: I1206 03:40:20.545229 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:20 crc kubenswrapper[4802]: I1206 03:40:20.546427 4802 scope.go:117] "RemoveContainer" containerID="b602a9bcbfd0f984f1778c0fcdf44058ce528b0d21af6563e76be589971384b5" Dec 06 03:40:21 crc kubenswrapper[4802]: I1206 03:40:21.549347 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 06 03:40:21 crc kubenswrapper[4802]: I1206 03:40:21.552634 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc"} Dec 06 03:40:21 crc kubenswrapper[4802]: I1206 03:40:21.552886 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:40:21 crc kubenswrapper[4802]: I1206 03:40:21.553955 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:21 crc kubenswrapper[4802]: I1206 03:40:21.554002 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:21 crc kubenswrapper[4802]: I1206 03:40:21.554016 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:22 crc kubenswrapper[4802]: I1206 03:40:22.519696 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:40:22 crc kubenswrapper[4802]: I1206 03:40:22.555647 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:40:22 crc kubenswrapper[4802]: I1206 03:40:22.555887 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:40:22 crc kubenswrapper[4802]: I1206 03:40:22.557223 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:22 crc kubenswrapper[4802]: I1206 03:40:22.557275 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:22 crc kubenswrapper[4802]: I1206 03:40:22.557295 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:22 crc kubenswrapper[4802]: I1206 03:40:22.562920 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:40:23 crc kubenswrapper[4802]: I1206 03:40:23.558803 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:40:23 crc kubenswrapper[4802]: I1206 03:40:23.560671 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:23 crc kubenswrapper[4802]: I1206 03:40:23.560722 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:23 crc kubenswrapper[4802]: I1206 03:40:23.560740 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:23 crc kubenswrapper[4802]: I1206 03:40:23.977713 4802 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 03:40:23 crc kubenswrapper[4802]: I1206 03:40:23.977844 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 03:40:24 crc kubenswrapper[4802]: I1206 03:40:24.222637 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 06 03:40:24 crc kubenswrapper[4802]: I1206 03:40:24.222932 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:40:24 crc kubenswrapper[4802]: I1206 03:40:24.224703 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:24 crc kubenswrapper[4802]: I1206 03:40:24.224777 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:24 crc kubenswrapper[4802]: I1206 03:40:24.224796 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:24 crc kubenswrapper[4802]: I1206 03:40:24.242918 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 06 03:40:24 crc kubenswrapper[4802]: I1206 03:40:24.560864 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:40:24 crc kubenswrapper[4802]: I1206 03:40:24.560978 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:40:24 crc kubenswrapper[4802]: I1206 03:40:24.562110 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:24 crc kubenswrapper[4802]: I1206 03:40:24.562150 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:24 crc kubenswrapper[4802]: I1206 03:40:24.562163 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:24 crc kubenswrapper[4802]: I1206 03:40:24.562351 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:24 crc kubenswrapper[4802]: I1206 03:40:24.562418 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:24 crc kubenswrapper[4802]: I1206 03:40:24.562437 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:25 crc kubenswrapper[4802]: I1206 03:40:25.535297 4802 trace.go:236] Trace[1403377243]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (06-Dec-2025 03:40:11.162) (total time: 14372ms): Dec 06 03:40:25 crc kubenswrapper[4802]: Trace[1403377243]: ---"Objects listed" error: 14372ms (03:40:25.535) Dec 06 03:40:25 crc kubenswrapper[4802]: Trace[1403377243]: [14.37233503s] [14.37233503s] END Dec 06 03:40:25 crc kubenswrapper[4802]: I1206 03:40:25.535326 4802 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 06 03:40:25 crc kubenswrapper[4802]: I1206 03:40:25.535782 4802 trace.go:236] Trace[410472501]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (06-Dec-2025 03:40:11.312) (total time: 14222ms): Dec 06 03:40:25 crc kubenswrapper[4802]: Trace[410472501]: ---"Objects listed" error: 14222ms (03:40:25.535) Dec 06 03:40:25 crc kubenswrapper[4802]: Trace[410472501]: [14.222787804s] [14.222787804s] END Dec 06 03:40:25 crc kubenswrapper[4802]: I1206 03:40:25.535812 4802 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 06 03:40:25 crc kubenswrapper[4802]: I1206 03:40:25.536341 4802 trace.go:236] Trace[1341375043]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (06-Dec-2025 03:40:11.793) (total time: 13742ms): Dec 06 03:40:25 crc kubenswrapper[4802]: Trace[1341375043]: ---"Objects listed" error: 13742ms (03:40:25.536) Dec 06 03:40:25 crc kubenswrapper[4802]: Trace[1341375043]: [13.742284946s] [13.742284946s] END Dec 06 03:40:25 crc kubenswrapper[4802]: I1206 03:40:25.536379 4802 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 06 03:40:25 crc kubenswrapper[4802]: I1206 03:40:25.537072 4802 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Dec 06 03:40:25 crc kubenswrapper[4802]: E1206 03:40:25.538502 4802 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 06 03:40:25 crc kubenswrapper[4802]: I1206 03:40:25.541451 4802 trace.go:236] Trace[871446993]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (06-Dec-2025 03:40:11.400) (total time: 14140ms): Dec 06 03:40:25 crc kubenswrapper[4802]: Trace[871446993]: ---"Objects listed" error: 14140ms (03:40:25.541) Dec 06 03:40:25 crc kubenswrapper[4802]: Trace[871446993]: [14.140657998s] [14.140657998s] END Dec 06 03:40:25 crc kubenswrapper[4802]: I1206 03:40:25.541489 4802 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 06 03:40:25 crc kubenswrapper[4802]: I1206 03:40:25.547126 4802 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 06 03:40:25 crc kubenswrapper[4802]: I1206 03:40:25.571376 4802 csr.go:261] certificate signing request csr-bp7nx is approved, waiting to be issued Dec 06 03:40:25 crc kubenswrapper[4802]: I1206 03:40:25.579841 4802 csr.go:257] certificate signing request csr-bp7nx is issued Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.375639 4802 apiserver.go:52] "Watching apiserver" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.378542 4802 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.378859 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-b2knv","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.379446 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-b2knv" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.379643 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.379732 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:40:26 crc kubenswrapper[4802]: E1206 03:40:26.379791 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.380091 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.380273 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.380564 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:40:26 crc kubenswrapper[4802]: E1206 03:40:26.380606 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.380655 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:40:26 crc kubenswrapper[4802]: E1206 03:40:26.380680 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.382132 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.382472 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.383075 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.383307 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.383641 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.383806 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.383868 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.383876 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.384208 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.384500 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.385059 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.386186 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.406675 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.426278 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.441721 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-b2knv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a3ae46f-03b5-4999-9d58-c38ee222f68a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk4hj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-b2knv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.452146 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.468687 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.477482 4802 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.478182 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.485661 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-b2knv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a3ae46f-03b5-4999-9d58-c38ee222f68a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk4hj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-b2knv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.494889 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.508872 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.518675 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.553302 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.553378 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.553406 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.553431 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.553455 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.553478 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.553505 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.553530 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.553552 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.553575 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.553600 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.553625 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.553650 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.553675 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.553698 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.553722 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.553764 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.553795 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.553819 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.553830 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.553855 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.553879 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.553902 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.553902 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.553925 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.553949 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.553945 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.553932 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.553974 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.554047 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.554034 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.554112 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.554162 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.554141 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.554194 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.554202 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.554236 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.554270 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.554282 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.554292 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.554299 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.554301 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.554353 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.554382 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.554379 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.554397 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.554407 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.554430 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.554456 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.554480 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.554501 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.554523 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.554549 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.554583 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.554649 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.554683 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.554711 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.554736 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.554775 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.554803 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.554825 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.554848 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.554868 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.554888 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.554910 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.554937 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.554960 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.554983 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.555006 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.555027 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.555051 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.555107 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.555134 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.555156 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.555176 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.555234 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.555305 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.555330 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.555356 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.555380 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.555406 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.554523 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.554557 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.554552 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.554586 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.554681 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.554773 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.554962 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.554980 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.554991 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.555012 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.555175 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.555215 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.555252 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.555323 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.555346 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.555421 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.555574 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.555432 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.555621 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.555662 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.555691 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.555715 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.555738 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.555789 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.555812 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.555834 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.555858 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.555881 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.555931 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.555954 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.555975 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.555997 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.556020 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.556043 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.556072 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.556096 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.556120 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.556141 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.556163 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.556187 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.556229 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.556251 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.556272 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.556293 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.556314 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.556334 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.556355 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.556377 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.556398 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.556424 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.556448 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.556474 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.556497 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.556521 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.556544 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.556565 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.556587 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.556606 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.556627 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.556648 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.556670 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.556693 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.556715 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.556738 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.556799 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.556823 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.556845 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.556869 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.556896 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.556980 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.557001 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.557025 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.557050 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.557073 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.557095 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.557121 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.557142 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.557198 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.557219 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.557240 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.557263 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.557310 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.557332 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.557354 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.557376 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.557396 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.557417 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.557440 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.557462 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.557482 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.557504 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.557527 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.557552 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.557575 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.557596 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.557620 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.557642 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.557664 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.557686 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.557708 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.557732 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.557792 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.557815 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.557837 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.557861 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.557882 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.557902 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.557924 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.557948 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.557970 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.557991 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.558013 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.558037 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.558060 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.558082 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.558105 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.558129 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.558157 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.558182 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.558204 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.558230 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.558254 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.558276 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.558300 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.558323 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.558345 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.558370 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.558392 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.558414 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.558437 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.558589 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.558614 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.558639 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.558664 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.558689 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.558712 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.558738 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.562518 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.562544 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.562563 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.562583 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.562623 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.562643 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.562660 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.562682 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.562729 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.562766 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.562789 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.562809 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.562869 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.562892 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vk4hj\" (UniqueName: \"kubernetes.io/projected/2a3ae46f-03b5-4999-9d58-c38ee222f68a-kube-api-access-vk4hj\") pod \"node-resolver-b2knv\" (UID: \"2a3ae46f-03b5-4999-9d58-c38ee222f68a\") " pod="openshift-dns/node-resolver-b2knv" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.562915 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.562933 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.562953 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.562973 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/2a3ae46f-03b5-4999-9d58-c38ee222f68a-hosts-file\") pod \"node-resolver-b2knv\" (UID: \"2a3ae46f-03b5-4999-9d58-c38ee222f68a\") " pod="openshift-dns/node-resolver-b2knv" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.562993 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.563012 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.563029 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.563048 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.563066 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.563087 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.563136 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.563147 4802 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.563158 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.563169 4802 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.563179 4802 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.563189 4802 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.563198 4802 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.563208 4802 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.563217 4802 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.563227 4802 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.563237 4802 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.563247 4802 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.563257 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.563268 4802 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.563277 4802 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.563287 4802 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.563297 4802 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.563306 4802 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.563317 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.563326 4802 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.563336 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.563346 4802 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.563360 4802 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.563372 4802 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.563384 4802 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.563397 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.563411 4802 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.563423 4802 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.563437 4802 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.563449 4802 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.563460 4802 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.564177 4802 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.555688 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.555738 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.555817 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.555853 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.555871 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.555892 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.556054 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.556141 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.556401 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.556427 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.556446 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.556706 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.556812 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.556843 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.557006 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.557188 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.557198 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.557403 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.557483 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.557541 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.557580 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.557826 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.558031 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.558081 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.558118 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.558339 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.558416 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.558428 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.558507 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.558821 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.558836 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.559217 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.559398 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.559544 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.559626 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.561790 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.562256 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.562276 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.562481 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.562581 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.562651 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.562776 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.562865 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.563045 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.563078 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.563245 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.563340 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.563862 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.573931 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.575123 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.575378 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.575570 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.575774 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.576078 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.576379 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.576423 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.576516 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.576530 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: E1206 03:40:26.576787 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 03:40:26 crc kubenswrapper[4802]: E1206 03:40:26.576802 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 03:40:26 crc kubenswrapper[4802]: E1206 03:40:26.576815 4802 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:40:26 crc kubenswrapper[4802]: E1206 03:40:26.576871 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 03:40:27.076849819 +0000 UTC m=+19.948758961 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.577089 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.577272 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.580969 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.581595 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.583652 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.583810 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.583918 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.584275 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.584291 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.585786 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.586996 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.587455 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.587083 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.587348 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.588353 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.589633 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.590216 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.590506 4802 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-12-06 03:35:25 +0000 UTC, rotation deadline is 2026-08-28 17:12:30.747917286 +0000 UTC Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.590538 4802 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6373h32m4.157381958s for next certificate rotation Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.591049 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.591104 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.591324 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.591527 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.592021 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: E1206 03:40:26.592076 4802 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 03:40:26 crc kubenswrapper[4802]: E1206 03:40:26.592137 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 03:40:27.092118462 +0000 UTC m=+19.964027614 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.592142 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.592290 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.592284 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.592485 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.592548 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.592646 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.592793 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.592816 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.592955 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.593025 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.593141 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.593330 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.593368 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: E1206 03:40:26.593427 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:40:27.093419048 +0000 UTC m=+19.965328200 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.593577 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.593728 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.593916 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.594324 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.594419 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.594545 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.594676 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.594957 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.595264 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.595493 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.595208 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.599339 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.599495 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.599661 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.599926 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.600107 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.600265 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.600596 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.600621 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.603899 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 03:40:26 crc kubenswrapper[4802]: E1206 03:40:26.606301 4802 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 03:40:26 crc kubenswrapper[4802]: E1206 03:40:26.606395 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 03:40:27.106373887 +0000 UTC m=+19.978283059 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.607706 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.609816 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.610256 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.610329 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.610635 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.611015 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.611031 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.612270 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.612600 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.612603 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.612680 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.612693 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.613458 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.616134 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.618205 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.618291 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.618622 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.618693 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.619483 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: E1206 03:40:26.619911 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 03:40:26 crc kubenswrapper[4802]: E1206 03:40:26.619938 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 03:40:26 crc kubenswrapper[4802]: E1206 03:40:26.619960 4802 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.620083 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.612812 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.621091 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.621220 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.621362 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.624079 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.624141 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.625862 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.628365 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 03:40:26 crc kubenswrapper[4802]: E1206 03:40:26.629930 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 03:40:27.120011725 +0000 UTC m=+19.991921087 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.629983 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.633018 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.633855 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.634356 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.635066 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.635225 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.635740 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.636175 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.636190 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.637111 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.637423 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.637945 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.638484 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.639706 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.640844 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.647073 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.647186 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.649970 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.651188 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.651380 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.651542 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.651903 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.653269 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.660449 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.661296 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.662097 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.664505 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.664568 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/2a3ae46f-03b5-4999-9d58-c38ee222f68a-hosts-file\") pod \"node-resolver-b2knv\" (UID: \"2a3ae46f-03b5-4999-9d58-c38ee222f68a\") " pod="openshift-dns/node-resolver-b2knv" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.664613 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.664702 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vk4hj\" (UniqueName: \"kubernetes.io/projected/2a3ae46f-03b5-4999-9d58-c38ee222f68a-kube-api-access-vk4hj\") pod \"node-resolver-b2knv\" (UID: \"2a3ae46f-03b5-4999-9d58-c38ee222f68a\") " pod="openshift-dns/node-resolver-b2knv" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.664836 4802 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.664855 4802 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.664864 4802 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.664875 4802 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.664889 4802 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.664908 4802 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.664921 4802 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.664934 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.664947 4802 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.664958 4802 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.664968 4802 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.664987 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.664997 4802 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665007 4802 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665016 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665026 4802 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665034 4802 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665043 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665054 4802 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665065 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665085 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665099 4802 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665113 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665125 4802 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665134 4802 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665145 4802 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665154 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665164 4802 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665173 4802 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665183 4802 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665194 4802 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665210 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665221 4802 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665231 4802 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665241 4802 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665281 4802 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665310 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665323 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665383 4802 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665398 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665409 4802 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665419 4802 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665430 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665443 4802 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665443 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665453 4802 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665496 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665510 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665521 4802 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665532 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665542 4802 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665552 4802 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665562 4802 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665572 4802 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665582 4802 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665592 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665602 4802 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665452 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/2a3ae46f-03b5-4999-9d58-c38ee222f68a-hosts-file\") pod \"node-resolver-b2knv\" (UID: \"2a3ae46f-03b5-4999-9d58-c38ee222f68a\") " pod="openshift-dns/node-resolver-b2knv" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665611 4802 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665622 4802 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665632 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665641 4802 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665653 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665662 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665671 4802 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665681 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665691 4802 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665700 4802 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665710 4802 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665720 4802 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665731 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665741 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665772 4802 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665782 4802 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665791 4802 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665801 4802 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665810 4802 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665819 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665828 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665837 4802 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665852 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665861 4802 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665869 4802 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665878 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665888 4802 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665896 4802 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665904 4802 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665912 4802 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665921 4802 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665930 4802 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665939 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665947 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665956 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665964 4802 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665972 4802 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665981 4802 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.665990 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.666000 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.666008 4802 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.666018 4802 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.666027 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.666035 4802 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.666044 4802 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.666053 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.666061 4802 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.666069 4802 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.666077 4802 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.666086 4802 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.666095 4802 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.666103 4802 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.666112 4802 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.666121 4802 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.666130 4802 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.666138 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.666146 4802 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.666155 4802 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.666164 4802 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.666172 4802 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.666181 4802 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.666190 4802 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.666199 4802 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.666208 4802 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.666219 4802 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.666229 4802 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.666237 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.666247 4802 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.666257 4802 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.666265 4802 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.666274 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.666282 4802 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.666291 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.666299 4802 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.666309 4802 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.666318 4802 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.666330 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.666339 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.666349 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.666359 4802 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.666368 4802 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.666377 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.666386 4802 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.666395 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.666404 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.666413 4802 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.666423 4802 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.666432 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.667187 4802 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.667202 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.667215 4802 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.667236 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.667246 4802 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.667256 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.667275 4802 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.667292 4802 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.667301 4802 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.667311 4802 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.667321 4802 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.667334 4802 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.667344 4802 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.667355 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.667366 4802 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.667380 4802 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.686187 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.689264 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.697101 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.698516 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vk4hj\" (UniqueName: \"kubernetes.io/projected/2a3ae46f-03b5-4999-9d58-c38ee222f68a-kube-api-access-vk4hj\") pod \"node-resolver-b2knv\" (UID: \"2a3ae46f-03b5-4999-9d58-c38ee222f68a\") " pod="openshift-dns/node-resolver-b2knv" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.699401 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.705196 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.706833 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.710346 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.714891 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 03:40:26 crc kubenswrapper[4802]: W1206 03:40:26.721174 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-262ad963ca3078a8e518e1a984176e82c139a1db334005adf60da25e1a1956fc WatchSource:0}: Error finding container 262ad963ca3078a8e518e1a984176e82c139a1db334005adf60da25e1a1956fc: Status 404 returned error can't find the container with id 262ad963ca3078a8e518e1a984176e82c139a1db334005adf60da25e1a1956fc Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.721435 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.754729 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.776155 4802 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.776176 4802 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.776185 4802 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.776194 4802 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.793822 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.812222 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.836157 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-b2knv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a3ae46f-03b5-4999-9d58-c38ee222f68a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk4hj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-b2knv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.847792 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.858265 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.868254 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 03:40:26 crc kubenswrapper[4802]: I1206 03:40:26.992994 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-b2knv" Dec 06 03:40:27 crc kubenswrapper[4802]: W1206 03:40:27.004716 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2a3ae46f_03b5_4999_9d58_c38ee222f68a.slice/crio-7873eae391543802a44c57cd1179cc6c87be0c8fe334f7e1a92f46a4c88360d2 WatchSource:0}: Error finding container 7873eae391543802a44c57cd1179cc6c87be0c8fe334f7e1a92f46a4c88360d2: Status 404 returned error can't find the container with id 7873eae391543802a44c57cd1179cc6c87be0c8fe334f7e1a92f46a4c88360d2 Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.079337 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:40:27 crc kubenswrapper[4802]: E1206 03:40:27.079476 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 03:40:27 crc kubenswrapper[4802]: E1206 03:40:27.079493 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 03:40:27 crc kubenswrapper[4802]: E1206 03:40:27.079505 4802 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:40:27 crc kubenswrapper[4802]: E1206 03:40:27.079578 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 03:40:28.079558168 +0000 UTC m=+20.951467320 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.179895 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.180013 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:40:27 crc kubenswrapper[4802]: E1206 03:40:27.180085 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:40:28.180056353 +0000 UTC m=+21.051965505 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:40:27 crc kubenswrapper[4802]: E1206 03:40:27.180140 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 03:40:27 crc kubenswrapper[4802]: E1206 03:40:27.180160 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.180170 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.180200 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:40:27 crc kubenswrapper[4802]: E1206 03:40:27.180176 4802 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:40:27 crc kubenswrapper[4802]: E1206 03:40:27.180277 4802 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 03:40:27 crc kubenswrapper[4802]: E1206 03:40:27.180301 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 03:40:28.180281509 +0000 UTC m=+21.052190741 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:40:27 crc kubenswrapper[4802]: E1206 03:40:27.180227 4802 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 03:40:27 crc kubenswrapper[4802]: E1206 03:40:27.180323 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 03:40:28.18031485 +0000 UTC m=+21.052224002 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 03:40:27 crc kubenswrapper[4802]: E1206 03:40:27.180337 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 03:40:28.18033086 +0000 UTC m=+21.052240012 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.294946 4802 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Dec 06 03:40:27 crc kubenswrapper[4802]: W1206 03:40:27.295175 4802 reflector.go:484] object-"openshift-network-operator"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-operator"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 06 03:40:27 crc kubenswrapper[4802]: W1206 03:40:27.295375 4802 reflector.go:484] object-"openshift-network-operator"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-operator"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 06 03:40:27 crc kubenswrapper[4802]: W1206 03:40:27.295406 4802 reflector.go:484] object-"openshift-network-operator"/"metrics-tls": watch of *v1.Secret ended with: very short watch: object-"openshift-network-operator"/"metrics-tls": Unexpected watch close - watch lasted less than a second and no items received Dec 06 03:40:27 crc kubenswrapper[4802]: W1206 03:40:27.295438 4802 reflector.go:484] object-"openshift-dns"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-dns"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 06 03:40:27 crc kubenswrapper[4802]: W1206 03:40:27.295472 4802 reflector.go:484] object-"openshift-network-node-identity"/"ovnkube-identity-cm": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-node-identity"/"ovnkube-identity-cm": Unexpected watch close - watch lasted less than a second and no items received Dec 06 03:40:27 crc kubenswrapper[4802]: W1206 03:40:27.295563 4802 reflector.go:484] object-"openshift-dns"/"node-resolver-dockercfg-kz9s7": watch of *v1.Secret ended with: very short watch: object-"openshift-dns"/"node-resolver-dockercfg-kz9s7": Unexpected watch close - watch lasted less than a second and no items received Dec 06 03:40:27 crc kubenswrapper[4802]: W1206 03:40:27.295586 4802 reflector.go:484] object-"openshift-network-node-identity"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-node-identity"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 06 03:40:27 crc kubenswrapper[4802]: W1206 03:40:27.295406 4802 reflector.go:484] object-"openshift-network-node-identity"/"network-node-identity-cert": watch of *v1.Secret ended with: very short watch: object-"openshift-network-node-identity"/"network-node-identity-cert": Unexpected watch close - watch lasted less than a second and no items received Dec 06 03:40:27 crc kubenswrapper[4802]: W1206 03:40:27.295601 4802 reflector.go:484] object-"openshift-network-operator"/"iptables-alerter-script": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-operator"/"iptables-alerter-script": Unexpected watch close - watch lasted less than a second and no items received Dec 06 03:40:27 crc kubenswrapper[4802]: W1206 03:40:27.295630 4802 reflector.go:484] object-"openshift-dns"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-dns"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 06 03:40:27 crc kubenswrapper[4802]: W1206 03:40:27.295704 4802 reflector.go:484] object-"openshift-network-node-identity"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-node-identity"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 06 03:40:27 crc kubenswrapper[4802]: W1206 03:40:27.295728 4802 reflector.go:484] object-"openshift-network-node-identity"/"env-overrides": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-node-identity"/"env-overrides": Unexpected watch close - watch lasted less than a second and no items received Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.467049 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:27Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.478988 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:27Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.491367 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2093b6-c4b5-4db8-be16-fbc859885e8d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0d6e582161f0d9126126e03b27211be22f6b5f797154928c92fa13a8118f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03004a29aefe7fe2452fb9f6663c0807afccb055d7ddbf6b907f305e6a19c81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1801ac64db4ba91c2e8a249e3651d3e50ab62e9f2c87312ce9bf7f38518a82cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6e454507cadbb982a566a3692995aaa3c733ff1cd7fb4870476602190e1eeba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:27Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.493909 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.494591 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.495463 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.496185 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.496724 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.497233 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.497906 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.498463 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.499367 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.499934 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.500481 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.501197 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.501739 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.502319 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.502925 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.503489 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.503499 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-b2knv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a3ae46f-03b5-4999-9d58-c38ee222f68a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk4hj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-b2knv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:27Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.505208 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.505712 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.506579 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.507278 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.507847 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.508464 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.508990 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.509868 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.510357 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.511037 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.511704 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.512275 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.515472 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.516039 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.516575 4802 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.517073 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:27Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.517616 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.520281 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.521308 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.522701 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.524537 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.525567 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.526671 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.527366 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.528609 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.529209 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.530378 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.531155 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.532275 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.532916 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.533907 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.533890 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:27Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.534496 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.535604 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.536170 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.537088 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.537654 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.538422 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.539807 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.540406 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.544461 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-zpxxw"] Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.545167 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-xtw2k"] Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.545606 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.546812 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-22rrq"] Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.547983 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-xtw2k" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.547991 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-22rrq" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.549645 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:27Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.550165 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.550245 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.550376 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.550401 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.550419 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.553059 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.553995 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.554236 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.554424 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.555013 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.557434 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.557585 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.572744 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:27Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.586590 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:27Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.597347 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"cb3f278fff47c900c56d8a563d5ed75a2ae84e2034b0f45aad4083790b9e7946"} Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.597394 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"3b10ca16ac26904ec5e6245344b71d836888cae9b069cae6a7a4831d609752db"} Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.597406 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"e63aac5887ebfed649c441e7ada4c6cf55c67da6ac3f28cc759b195275644f01"} Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.599585 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.600112 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.602526 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtw2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3a3bac2-3923-4507-a4de-1556b3ad473e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtw2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:27Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.603287 4802 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc" exitCode=255 Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.603319 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc"} Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.603553 4802 scope.go:117] "RemoveContainer" containerID="b602a9bcbfd0f984f1778c0fcdf44058ce528b0d21af6563e76be589971384b5" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.605443 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"0fad9b63d84f41fccf696cd313d7ca7f72b3d3251d6c1b7341573a3fadf1e378"} Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.605468 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"60e399c9c46db57f08d96a0a6d772cbeb2a6c861056424a866d50f8b6bbcb6ae"} Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.607279 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"262ad963ca3078a8e518e1a984176e82c139a1db334005adf60da25e1a1956fc"} Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.610979 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-b2knv" event={"ID":"2a3ae46f-03b5-4999-9d58-c38ee222f68a","Type":"ContainerStarted","Data":"76665d4c127445009fd66078827b54607c9ef812c8011d1c0b4cde1c0516853a"} Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.611059 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-b2knv" event={"ID":"2a3ae46f-03b5-4999-9d58-c38ee222f68a","Type":"ContainerStarted","Data":"7873eae391543802a44c57cd1179cc6c87be0c8fe334f7e1a92f46a4c88360d2"} Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.614683 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.615003 4802 scope.go:117] "RemoveContainer" containerID="2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc" Dec 06 03:40:27 crc kubenswrapper[4802]: E1206 03:40:27.615128 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.619124 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22rrq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03112a5d-5882-4fac-9ae5-13525ff82fe0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trlt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22rrq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:27Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.633854 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-b2knv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a3ae46f-03b5-4999-9d58-c38ee222f68a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk4hj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-b2knv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:27Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.649881 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:27Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.665908 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:27Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.677786 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6df38316-e0d3-4018-8d27-3620eba3a68d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpxxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:27Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.684131 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/03112a5d-5882-4fac-9ae5-13525ff82fe0-host-run-netns\") pod \"multus-22rrq\" (UID: \"03112a5d-5882-4fac-9ae5-13525ff82fe0\") " pod="openshift-multus/multus-22rrq" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.684182 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/03112a5d-5882-4fac-9ae5-13525ff82fe0-host-var-lib-kubelet\") pod \"multus-22rrq\" (UID: \"03112a5d-5882-4fac-9ae5-13525ff82fe0\") " pod="openshift-multus/multus-22rrq" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.684282 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/03112a5d-5882-4fac-9ae5-13525ff82fe0-hostroot\") pod \"multus-22rrq\" (UID: \"03112a5d-5882-4fac-9ae5-13525ff82fe0\") " pod="openshift-multus/multus-22rrq" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.684315 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/03112a5d-5882-4fac-9ae5-13525ff82fe0-multus-cni-dir\") pod \"multus-22rrq\" (UID: \"03112a5d-5882-4fac-9ae5-13525ff82fe0\") " pod="openshift-multus/multus-22rrq" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.684410 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/03112a5d-5882-4fac-9ae5-13525ff82fe0-host-var-lib-cni-multus\") pod \"multus-22rrq\" (UID: \"03112a5d-5882-4fac-9ae5-13525ff82fe0\") " pod="openshift-multus/multus-22rrq" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.684500 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfj4n\" (UniqueName: \"kubernetes.io/projected/6df38316-e0d3-4018-8d27-3620eba3a68d-kube-api-access-mfj4n\") pod \"machine-config-daemon-zpxxw\" (UID: \"6df38316-e0d3-4018-8d27-3620eba3a68d\") " pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.684578 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/03112a5d-5882-4fac-9ae5-13525ff82fe0-multus-socket-dir-parent\") pod \"multus-22rrq\" (UID: \"03112a5d-5882-4fac-9ae5-13525ff82fe0\") " pod="openshift-multus/multus-22rrq" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.684651 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/03112a5d-5882-4fac-9ae5-13525ff82fe0-etc-kubernetes\") pod \"multus-22rrq\" (UID: \"03112a5d-5882-4fac-9ae5-13525ff82fe0\") " pod="openshift-multus/multus-22rrq" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.684722 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/a3a3bac2-3923-4507-a4de-1556b3ad473e-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-xtw2k\" (UID: \"a3a3bac2-3923-4507-a4de-1556b3ad473e\") " pod="openshift-multus/multus-additional-cni-plugins-xtw2k" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.684802 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6df38316-e0d3-4018-8d27-3620eba3a68d-mcd-auth-proxy-config\") pod \"machine-config-daemon-zpxxw\" (UID: \"6df38316-e0d3-4018-8d27-3620eba3a68d\") " pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.684869 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/03112a5d-5882-4fac-9ae5-13525ff82fe0-os-release\") pod \"multus-22rrq\" (UID: \"03112a5d-5882-4fac-9ae5-13525ff82fe0\") " pod="openshift-multus/multus-22rrq" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.684935 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/6df38316-e0d3-4018-8d27-3620eba3a68d-rootfs\") pod \"machine-config-daemon-zpxxw\" (UID: \"6df38316-e0d3-4018-8d27-3620eba3a68d\") " pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.684999 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/03112a5d-5882-4fac-9ae5-13525ff82fe0-cni-binary-copy\") pod \"multus-22rrq\" (UID: \"03112a5d-5882-4fac-9ae5-13525ff82fe0\") " pod="openshift-multus/multus-22rrq" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.685110 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/a3a3bac2-3923-4507-a4de-1556b3ad473e-cnibin\") pod \"multus-additional-cni-plugins-xtw2k\" (UID: \"a3a3bac2-3923-4507-a4de-1556b3ad473e\") " pod="openshift-multus/multus-additional-cni-plugins-xtw2k" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.685217 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/03112a5d-5882-4fac-9ae5-13525ff82fe0-multus-conf-dir\") pod \"multus-22rrq\" (UID: \"03112a5d-5882-4fac-9ae5-13525ff82fe0\") " pod="openshift-multus/multus-22rrq" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.686582 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/03112a5d-5882-4fac-9ae5-13525ff82fe0-host-var-lib-cni-bin\") pod \"multus-22rrq\" (UID: \"03112a5d-5882-4fac-9ae5-13525ff82fe0\") " pod="openshift-multus/multus-22rrq" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.686666 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gr56s\" (UniqueName: \"kubernetes.io/projected/a3a3bac2-3923-4507-a4de-1556b3ad473e-kube-api-access-gr56s\") pod \"multus-additional-cni-plugins-xtw2k\" (UID: \"a3a3bac2-3923-4507-a4de-1556b3ad473e\") " pod="openshift-multus/multus-additional-cni-plugins-xtw2k" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.686807 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/03112a5d-5882-4fac-9ae5-13525ff82fe0-system-cni-dir\") pod \"multus-22rrq\" (UID: \"03112a5d-5882-4fac-9ae5-13525ff82fe0\") " pod="openshift-multus/multus-22rrq" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.686929 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/03112a5d-5882-4fac-9ae5-13525ff82fe0-host-run-multus-certs\") pod \"multus-22rrq\" (UID: \"03112a5d-5882-4fac-9ae5-13525ff82fe0\") " pod="openshift-multus/multus-22rrq" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.687042 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trlt4\" (UniqueName: \"kubernetes.io/projected/03112a5d-5882-4fac-9ae5-13525ff82fe0-kube-api-access-trlt4\") pod \"multus-22rrq\" (UID: \"03112a5d-5882-4fac-9ae5-13525ff82fe0\") " pod="openshift-multus/multus-22rrq" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.687158 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a3a3bac2-3923-4507-a4de-1556b3ad473e-system-cni-dir\") pod \"multus-additional-cni-plugins-xtw2k\" (UID: \"a3a3bac2-3923-4507-a4de-1556b3ad473e\") " pod="openshift-multus/multus-additional-cni-plugins-xtw2k" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.687195 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/03112a5d-5882-4fac-9ae5-13525ff82fe0-host-run-k8s-cni-cncf-io\") pod \"multus-22rrq\" (UID: \"03112a5d-5882-4fac-9ae5-13525ff82fe0\") " pod="openshift-multus/multus-22rrq" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.687224 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/03112a5d-5882-4fac-9ae5-13525ff82fe0-cnibin\") pod \"multus-22rrq\" (UID: \"03112a5d-5882-4fac-9ae5-13525ff82fe0\") " pod="openshift-multus/multus-22rrq" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.687305 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6df38316-e0d3-4018-8d27-3620eba3a68d-proxy-tls\") pod \"machine-config-daemon-zpxxw\" (UID: \"6df38316-e0d3-4018-8d27-3620eba3a68d\") " pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.688043 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/a3a3bac2-3923-4507-a4de-1556b3ad473e-cni-binary-copy\") pod \"multus-additional-cni-plugins-xtw2k\" (UID: \"a3a3bac2-3923-4507-a4de-1556b3ad473e\") " pod="openshift-multus/multus-additional-cni-plugins-xtw2k" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.688120 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a3a3bac2-3923-4507-a4de-1556b3ad473e-tuning-conf-dir\") pod \"multus-additional-cni-plugins-xtw2k\" (UID: \"a3a3bac2-3923-4507-a4de-1556b3ad473e\") " pod="openshift-multus/multus-additional-cni-plugins-xtw2k" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.688170 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/a3a3bac2-3923-4507-a4de-1556b3ad473e-os-release\") pod \"multus-additional-cni-plugins-xtw2k\" (UID: \"a3a3bac2-3923-4507-a4de-1556b3ad473e\") " pod="openshift-multus/multus-additional-cni-plugins-xtw2k" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.688206 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/03112a5d-5882-4fac-9ae5-13525ff82fe0-multus-daemon-config\") pod \"multus-22rrq\" (UID: \"03112a5d-5882-4fac-9ae5-13525ff82fe0\") " pod="openshift-multus/multus-22rrq" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.691057 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:27Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.702077 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2093b6-c4b5-4db8-be16-fbc859885e8d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0d6e582161f0d9126126e03b27211be22f6b5f797154928c92fa13a8118f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03004a29aefe7fe2452fb9f6663c0807afccb055d7ddbf6b907f305e6a19c81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1801ac64db4ba91c2e8a249e3651d3e50ab62e9f2c87312ce9bf7f38518a82cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6e454507cadbb982a566a3692995aaa3c733ff1cd7fb4870476602190e1eeba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:27Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.713864 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:27Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.738919 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:27Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.759768 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22rrq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03112a5d-5882-4fac-9ae5-13525ff82fe0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trlt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22rrq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:27Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.781787 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1721e985-40bd-4b2d-a0e8-d4e365094ba6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a895e0cc04880c3287ba725ac5484835235f1e066a94ce98fcf4b6133c5cbd2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a21a97385f856ec996105b597d0ae81c56ed28b625ef3f2905e1af6ec8f6a45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://646e01a58908ba2e703aae5eef53e1b981b0f0c3d1d3202487f2a8b94ee3f55d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b602a9bcbfd0f984f1778c0fcdf44058ce528b0d21af6563e76be589971384b5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:40:20Z\\\",\\\"message\\\":\\\"W1206 03:40:09.599454 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1206 03:40:09.599705 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764992409 cert, and key in /tmp/serving-cert-3362951094/serving-signer.crt, /tmp/serving-cert-3362951094/serving-signer.key\\\\nI1206 03:40:09.788947 1 observer_polling.go:159] Starting file observer\\\\nW1206 03:40:09.790951 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1206 03:40:09.791214 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 03:40:09.792215 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3362951094/tls.crt::/tmp/serving-cert-3362951094/tls.key\\\\\\\"\\\\nF1206 03:40:20.204085 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"ver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:40:26.813730 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814130 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:40:26.814139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:40:26.814143 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:40:26.814145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 03:40:26.821403 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821442 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821482 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821496 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 03:40:26.821509 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821514 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 03:40:26.821612 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 03:40:26.821619 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1206 03:40:26.822049 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16a92d501c9e626fe91e4a6793e426d10eb292cfbab766db8b8cc72e5dae94ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:27Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.789512 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/03112a5d-5882-4fac-9ae5-13525ff82fe0-host-var-lib-cni-bin\") pod \"multus-22rrq\" (UID: \"03112a5d-5882-4fac-9ae5-13525ff82fe0\") " pod="openshift-multus/multus-22rrq" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.789551 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/03112a5d-5882-4fac-9ae5-13525ff82fe0-multus-conf-dir\") pod \"multus-22rrq\" (UID: \"03112a5d-5882-4fac-9ae5-13525ff82fe0\") " pod="openshift-multus/multus-22rrq" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.789595 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trlt4\" (UniqueName: \"kubernetes.io/projected/03112a5d-5882-4fac-9ae5-13525ff82fe0-kube-api-access-trlt4\") pod \"multus-22rrq\" (UID: \"03112a5d-5882-4fac-9ae5-13525ff82fe0\") " pod="openshift-multus/multus-22rrq" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.789610 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/03112a5d-5882-4fac-9ae5-13525ff82fe0-host-var-lib-cni-bin\") pod \"multus-22rrq\" (UID: \"03112a5d-5882-4fac-9ae5-13525ff82fe0\") " pod="openshift-multus/multus-22rrq" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.789642 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/03112a5d-5882-4fac-9ae5-13525ff82fe0-multus-conf-dir\") pod \"multus-22rrq\" (UID: \"03112a5d-5882-4fac-9ae5-13525ff82fe0\") " pod="openshift-multus/multus-22rrq" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.789619 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gr56s\" (UniqueName: \"kubernetes.io/projected/a3a3bac2-3923-4507-a4de-1556b3ad473e-kube-api-access-gr56s\") pod \"multus-additional-cni-plugins-xtw2k\" (UID: \"a3a3bac2-3923-4507-a4de-1556b3ad473e\") " pod="openshift-multus/multus-additional-cni-plugins-xtw2k" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.789723 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/03112a5d-5882-4fac-9ae5-13525ff82fe0-system-cni-dir\") pod \"multus-22rrq\" (UID: \"03112a5d-5882-4fac-9ae5-13525ff82fe0\") " pod="openshift-multus/multus-22rrq" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.789743 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/03112a5d-5882-4fac-9ae5-13525ff82fe0-host-run-multus-certs\") pod \"multus-22rrq\" (UID: \"03112a5d-5882-4fac-9ae5-13525ff82fe0\") " pod="openshift-multus/multus-22rrq" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.789809 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a3a3bac2-3923-4507-a4de-1556b3ad473e-system-cni-dir\") pod \"multus-additional-cni-plugins-xtw2k\" (UID: \"a3a3bac2-3923-4507-a4de-1556b3ad473e\") " pod="openshift-multus/multus-additional-cni-plugins-xtw2k" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.789825 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/03112a5d-5882-4fac-9ae5-13525ff82fe0-host-run-k8s-cni-cncf-io\") pod \"multus-22rrq\" (UID: \"03112a5d-5882-4fac-9ae5-13525ff82fe0\") " pod="openshift-multus/multus-22rrq" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.789848 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6df38316-e0d3-4018-8d27-3620eba3a68d-proxy-tls\") pod \"machine-config-daemon-zpxxw\" (UID: \"6df38316-e0d3-4018-8d27-3620eba3a68d\") " pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.789870 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/03112a5d-5882-4fac-9ae5-13525ff82fe0-cnibin\") pod \"multus-22rrq\" (UID: \"03112a5d-5882-4fac-9ae5-13525ff82fe0\") " pod="openshift-multus/multus-22rrq" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.789888 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/a3a3bac2-3923-4507-a4de-1556b3ad473e-cni-binary-copy\") pod \"multus-additional-cni-plugins-xtw2k\" (UID: \"a3a3bac2-3923-4507-a4de-1556b3ad473e\") " pod="openshift-multus/multus-additional-cni-plugins-xtw2k" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.789906 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a3a3bac2-3923-4507-a4de-1556b3ad473e-tuning-conf-dir\") pod \"multus-additional-cni-plugins-xtw2k\" (UID: \"a3a3bac2-3923-4507-a4de-1556b3ad473e\") " pod="openshift-multus/multus-additional-cni-plugins-xtw2k" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.789922 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/a3a3bac2-3923-4507-a4de-1556b3ad473e-os-release\") pod \"multus-additional-cni-plugins-xtw2k\" (UID: \"a3a3bac2-3923-4507-a4de-1556b3ad473e\") " pod="openshift-multus/multus-additional-cni-plugins-xtw2k" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.789938 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/03112a5d-5882-4fac-9ae5-13525ff82fe0-multus-daemon-config\") pod \"multus-22rrq\" (UID: \"03112a5d-5882-4fac-9ae5-13525ff82fe0\") " pod="openshift-multus/multus-22rrq" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.789976 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/03112a5d-5882-4fac-9ae5-13525ff82fe0-host-run-netns\") pod \"multus-22rrq\" (UID: \"03112a5d-5882-4fac-9ae5-13525ff82fe0\") " pod="openshift-multus/multus-22rrq" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.789981 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/03112a5d-5882-4fac-9ae5-13525ff82fe0-system-cni-dir\") pod \"multus-22rrq\" (UID: \"03112a5d-5882-4fac-9ae5-13525ff82fe0\") " pod="openshift-multus/multus-22rrq" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.789993 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/03112a5d-5882-4fac-9ae5-13525ff82fe0-host-var-lib-kubelet\") pod \"multus-22rrq\" (UID: \"03112a5d-5882-4fac-9ae5-13525ff82fe0\") " pod="openshift-multus/multus-22rrq" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.790009 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/03112a5d-5882-4fac-9ae5-13525ff82fe0-host-var-lib-cni-multus\") pod \"multus-22rrq\" (UID: \"03112a5d-5882-4fac-9ae5-13525ff82fe0\") " pod="openshift-multus/multus-22rrq" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.790024 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/03112a5d-5882-4fac-9ae5-13525ff82fe0-hostroot\") pod \"multus-22rrq\" (UID: \"03112a5d-5882-4fac-9ae5-13525ff82fe0\") " pod="openshift-multus/multus-22rrq" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.790040 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/03112a5d-5882-4fac-9ae5-13525ff82fe0-multus-cni-dir\") pod \"multus-22rrq\" (UID: \"03112a5d-5882-4fac-9ae5-13525ff82fe0\") " pod="openshift-multus/multus-22rrq" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.790058 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6df38316-e0d3-4018-8d27-3620eba3a68d-mcd-auth-proxy-config\") pod \"machine-config-daemon-zpxxw\" (UID: \"6df38316-e0d3-4018-8d27-3620eba3a68d\") " pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.790073 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfj4n\" (UniqueName: \"kubernetes.io/projected/6df38316-e0d3-4018-8d27-3620eba3a68d-kube-api-access-mfj4n\") pod \"machine-config-daemon-zpxxw\" (UID: \"6df38316-e0d3-4018-8d27-3620eba3a68d\") " pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.790088 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/03112a5d-5882-4fac-9ae5-13525ff82fe0-multus-socket-dir-parent\") pod \"multus-22rrq\" (UID: \"03112a5d-5882-4fac-9ae5-13525ff82fe0\") " pod="openshift-multus/multus-22rrq" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.790124 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/03112a5d-5882-4fac-9ae5-13525ff82fe0-etc-kubernetes\") pod \"multus-22rrq\" (UID: \"03112a5d-5882-4fac-9ae5-13525ff82fe0\") " pod="openshift-multus/multus-22rrq" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.790145 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/a3a3bac2-3923-4507-a4de-1556b3ad473e-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-xtw2k\" (UID: \"a3a3bac2-3923-4507-a4de-1556b3ad473e\") " pod="openshift-multus/multus-additional-cni-plugins-xtw2k" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.790163 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/03112a5d-5882-4fac-9ae5-13525ff82fe0-os-release\") pod \"multus-22rrq\" (UID: \"03112a5d-5882-4fac-9ae5-13525ff82fe0\") " pod="openshift-multus/multus-22rrq" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.790179 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/03112a5d-5882-4fac-9ae5-13525ff82fe0-cni-binary-copy\") pod \"multus-22rrq\" (UID: \"03112a5d-5882-4fac-9ae5-13525ff82fe0\") " pod="openshift-multus/multus-22rrq" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.790177 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/03112a5d-5882-4fac-9ae5-13525ff82fe0-host-run-multus-certs\") pod \"multus-22rrq\" (UID: \"03112a5d-5882-4fac-9ae5-13525ff82fe0\") " pod="openshift-multus/multus-22rrq" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.790219 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/6df38316-e0d3-4018-8d27-3620eba3a68d-rootfs\") pod \"machine-config-daemon-zpxxw\" (UID: \"6df38316-e0d3-4018-8d27-3620eba3a68d\") " pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.790238 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/a3a3bac2-3923-4507-a4de-1556b3ad473e-cnibin\") pod \"multus-additional-cni-plugins-xtw2k\" (UID: \"a3a3bac2-3923-4507-a4de-1556b3ad473e\") " pod="openshift-multus/multus-additional-cni-plugins-xtw2k" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.790481 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a3a3bac2-3923-4507-a4de-1556b3ad473e-tuning-conf-dir\") pod \"multus-additional-cni-plugins-xtw2k\" (UID: \"a3a3bac2-3923-4507-a4de-1556b3ad473e\") " pod="openshift-multus/multus-additional-cni-plugins-xtw2k" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.790515 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/03112a5d-5882-4fac-9ae5-13525ff82fe0-cnibin\") pod \"multus-22rrq\" (UID: \"03112a5d-5882-4fac-9ae5-13525ff82fe0\") " pod="openshift-multus/multus-22rrq" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.790692 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a3a3bac2-3923-4507-a4de-1556b3ad473e-system-cni-dir\") pod \"multus-additional-cni-plugins-xtw2k\" (UID: \"a3a3bac2-3923-4507-a4de-1556b3ad473e\") " pod="openshift-multus/multus-additional-cni-plugins-xtw2k" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.790731 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/03112a5d-5882-4fac-9ae5-13525ff82fe0-host-run-netns\") pod \"multus-22rrq\" (UID: \"03112a5d-5882-4fac-9ae5-13525ff82fe0\") " pod="openshift-multus/multus-22rrq" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.790768 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/03112a5d-5882-4fac-9ae5-13525ff82fe0-host-run-k8s-cni-cncf-io\") pod \"multus-22rrq\" (UID: \"03112a5d-5882-4fac-9ae5-13525ff82fe0\") " pod="openshift-multus/multus-22rrq" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.790842 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/a3a3bac2-3923-4507-a4de-1556b3ad473e-os-release\") pod \"multus-additional-cni-plugins-xtw2k\" (UID: \"a3a3bac2-3923-4507-a4de-1556b3ad473e\") " pod="openshift-multus/multus-additional-cni-plugins-xtw2k" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.790865 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/03112a5d-5882-4fac-9ae5-13525ff82fe0-multus-socket-dir-parent\") pod \"multus-22rrq\" (UID: \"03112a5d-5882-4fac-9ae5-13525ff82fe0\") " pod="openshift-multus/multus-22rrq" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.790898 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/03112a5d-5882-4fac-9ae5-13525ff82fe0-etc-kubernetes\") pod \"multus-22rrq\" (UID: \"03112a5d-5882-4fac-9ae5-13525ff82fe0\") " pod="openshift-multus/multus-22rrq" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.791135 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/a3a3bac2-3923-4507-a4de-1556b3ad473e-cni-binary-copy\") pod \"multus-additional-cni-plugins-xtw2k\" (UID: \"a3a3bac2-3923-4507-a4de-1556b3ad473e\") " pod="openshift-multus/multus-additional-cni-plugins-xtw2k" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.791177 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/03112a5d-5882-4fac-9ae5-13525ff82fe0-host-var-lib-cni-multus\") pod \"multus-22rrq\" (UID: \"03112a5d-5882-4fac-9ae5-13525ff82fe0\") " pod="openshift-multus/multus-22rrq" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.791200 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/03112a5d-5882-4fac-9ae5-13525ff82fe0-host-var-lib-kubelet\") pod \"multus-22rrq\" (UID: \"03112a5d-5882-4fac-9ae5-13525ff82fe0\") " pod="openshift-multus/multus-22rrq" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.791221 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/03112a5d-5882-4fac-9ae5-13525ff82fe0-hostroot\") pod \"multus-22rrq\" (UID: \"03112a5d-5882-4fac-9ae5-13525ff82fe0\") " pod="openshift-multus/multus-22rrq" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.791302 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/03112a5d-5882-4fac-9ae5-13525ff82fe0-multus-cni-dir\") pod \"multus-22rrq\" (UID: \"03112a5d-5882-4fac-9ae5-13525ff82fe0\") " pod="openshift-multus/multus-22rrq" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.791338 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/03112a5d-5882-4fac-9ae5-13525ff82fe0-os-release\") pod \"multus-22rrq\" (UID: \"03112a5d-5882-4fac-9ae5-13525ff82fe0\") " pod="openshift-multus/multus-22rrq" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.791362 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/6df38316-e0d3-4018-8d27-3620eba3a68d-rootfs\") pod \"machine-config-daemon-zpxxw\" (UID: \"6df38316-e0d3-4018-8d27-3620eba3a68d\") " pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.791385 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/a3a3bac2-3923-4507-a4de-1556b3ad473e-cnibin\") pod \"multus-additional-cni-plugins-xtw2k\" (UID: \"a3a3bac2-3923-4507-a4de-1556b3ad473e\") " pod="openshift-multus/multus-additional-cni-plugins-xtw2k" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.791410 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/a3a3bac2-3923-4507-a4de-1556b3ad473e-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-xtw2k\" (UID: \"a3a3bac2-3923-4507-a4de-1556b3ad473e\") " pod="openshift-multus/multus-additional-cni-plugins-xtw2k" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.791425 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6df38316-e0d3-4018-8d27-3620eba3a68d-mcd-auth-proxy-config\") pod \"machine-config-daemon-zpxxw\" (UID: \"6df38316-e0d3-4018-8d27-3620eba3a68d\") " pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.791576 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/03112a5d-5882-4fac-9ae5-13525ff82fe0-multus-daemon-config\") pod \"multus-22rrq\" (UID: \"03112a5d-5882-4fac-9ae5-13525ff82fe0\") " pod="openshift-multus/multus-22rrq" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.791714 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/03112a5d-5882-4fac-9ae5-13525ff82fe0-cni-binary-copy\") pod \"multus-22rrq\" (UID: \"03112a5d-5882-4fac-9ae5-13525ff82fe0\") " pod="openshift-multus/multus-22rrq" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.797955 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-b2knv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a3ae46f-03b5-4999-9d58-c38ee222f68a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76665d4c127445009fd66078827b54607c9ef812c8011d1c0b4cde1c0516853a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk4hj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-b2knv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:27Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.798101 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6df38316-e0d3-4018-8d27-3620eba3a68d-proxy-tls\") pod \"machine-config-daemon-zpxxw\" (UID: \"6df38316-e0d3-4018-8d27-3620eba3a68d\") " pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.807885 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfj4n\" (UniqueName: \"kubernetes.io/projected/6df38316-e0d3-4018-8d27-3620eba3a68d-kube-api-access-mfj4n\") pod \"machine-config-daemon-zpxxw\" (UID: \"6df38316-e0d3-4018-8d27-3620eba3a68d\") " pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.808207 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gr56s\" (UniqueName: \"kubernetes.io/projected/a3a3bac2-3923-4507-a4de-1556b3ad473e-kube-api-access-gr56s\") pod \"multus-additional-cni-plugins-xtw2k\" (UID: \"a3a3bac2-3923-4507-a4de-1556b3ad473e\") " pod="openshift-multus/multus-additional-cni-plugins-xtw2k" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.808972 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trlt4\" (UniqueName: \"kubernetes.io/projected/03112a5d-5882-4fac-9ae5-13525ff82fe0-kube-api-access-trlt4\") pod \"multus-22rrq\" (UID: \"03112a5d-5882-4fac-9ae5-13525ff82fe0\") " pod="openshift-multus/multus-22rrq" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.810236 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:27Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.822143 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:27Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.824591 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-pvm6q"] Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.825722 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.827118 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.827339 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.828529 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.828602 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.828728 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.828741 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.828793 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.833362 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6df38316-e0d3-4018-8d27-3620eba3a68d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpxxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:27Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.844397 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb3f278fff47c900c56d8a563d5ed75a2ae84e2034b0f45aad4083790b9e7946\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b10ca16ac26904ec5e6245344b71d836888cae9b069cae6a7a4831d609752db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:27Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.856668 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2093b6-c4b5-4db8-be16-fbc859885e8d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0d6e582161f0d9126126e03b27211be22f6b5f797154928c92fa13a8118f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03004a29aefe7fe2452fb9f6663c0807afccb055d7ddbf6b907f305e6a19c81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1801ac64db4ba91c2e8a249e3651d3e50ab62e9f2c87312ce9bf7f38518a82cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6e454507cadbb982a566a3692995aaa3c733ff1cd7fb4870476602190e1eeba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:27Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.863910 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.871296 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:27Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.874665 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-xtw2k" Dec 06 03:40:27 crc kubenswrapper[4802]: W1206 03:40:27.875166 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6df38316_e0d3_4018_8d27_3620eba3a68d.slice/crio-99c5a9de0960265c4986137b196c8e15dcf6bb70c365936915181e63a54dffd0 WatchSource:0}: Error finding container 99c5a9de0960265c4986137b196c8e15dcf6bb70c365936915181e63a54dffd0: Status 404 returned error can't find the container with id 99c5a9de0960265c4986137b196c8e15dcf6bb70c365936915181e63a54dffd0 Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.881082 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-22rrq" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.890809 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fad9b63d84f41fccf696cd313d7ca7f72b3d3251d6c1b7341573a3fadf1e378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:27Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.904764 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:27Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.923029 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtw2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3a3bac2-3923-4507-a4de-1556b3ad473e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtw2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:27Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.940368 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2093b6-c4b5-4db8-be16-fbc859885e8d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0d6e582161f0d9126126e03b27211be22f6b5f797154928c92fa13a8118f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03004a29aefe7fe2452fb9f6663c0807afccb055d7ddbf6b907f305e6a19c81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1801ac64db4ba91c2e8a249e3651d3e50ab62e9f2c87312ce9bf7f38518a82cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6e454507cadbb982a566a3692995aaa3c733ff1cd7fb4870476602190e1eeba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:27Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.955565 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:27Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.973283 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fad9b63d84f41fccf696cd313d7ca7f72b3d3251d6c1b7341573a3fadf1e378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:27Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.991526 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb3f278fff47c900c56d8a563d5ed75a2ae84e2034b0f45aad4083790b9e7946\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b10ca16ac26904ec5e6245344b71d836888cae9b069cae6a7a4831d609752db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:27Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.991607 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62rc4\" (UniqueName: \"kubernetes.io/projected/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-kube-api-access-62rc4\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.991636 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-host-cni-netd\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.991652 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-ovn-node-metrics-cert\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.991666 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-host-cni-bin\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.991680 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-host-run-ovn-kubernetes\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.991705 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-run-openvswitch\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.991721 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-etc-openvswitch\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.991780 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-var-lib-openvswitch\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.991802 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-host-kubelet\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.991819 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-systemd-units\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.991832 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-host-run-netns\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.991918 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-run-ovn\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.991995 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-log-socket\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.992020 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-ovnkube-script-lib\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.992048 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-host-slash\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.992075 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-node-log\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.992136 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-env-overrides\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.992158 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-run-systemd\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.992182 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:27 crc kubenswrapper[4802]: I1206 03:40:27.992205 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-ovnkube-config\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.003942 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.016043 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtw2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3a3bac2-3923-4507-a4de-1556b3ad473e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtw2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.032603 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pvm6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.044474 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1721e985-40bd-4b2d-a0e8-d4e365094ba6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a895e0cc04880c3287ba725ac5484835235f1e066a94ce98fcf4b6133c5cbd2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a21a97385f856ec996105b597d0ae81c56ed28b625ef3f2905e1af6ec8f6a45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://646e01a58908ba2e703aae5eef53e1b981b0f0c3d1d3202487f2a8b94ee3f55d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b602a9bcbfd0f984f1778c0fcdf44058ce528b0d21af6563e76be589971384b5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:40:20Z\\\",\\\"message\\\":\\\"W1206 03:40:09.599454 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1206 03:40:09.599705 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764992409 cert, and key in /tmp/serving-cert-3362951094/serving-signer.crt, /tmp/serving-cert-3362951094/serving-signer.key\\\\nI1206 03:40:09.788947 1 observer_polling.go:159] Starting file observer\\\\nW1206 03:40:09.790951 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1206 03:40:09.791214 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 03:40:09.792215 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3362951094/tls.crt::/tmp/serving-cert-3362951094/tls.key\\\\\\\"\\\\nF1206 03:40:20.204085 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"ver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:40:26.813730 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814130 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:40:26.814139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:40:26.814143 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:40:26.814145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 03:40:26.821403 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821442 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821482 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821496 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 03:40:26.821509 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821514 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 03:40:26.821612 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 03:40:26.821619 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1206 03:40:26.822049 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16a92d501c9e626fe91e4a6793e426d10eb292cfbab766db8b8cc72e5dae94ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.056233 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-b2knv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a3ae46f-03b5-4999-9d58-c38ee222f68a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76665d4c127445009fd66078827b54607c9ef812c8011d1c0b4cde1c0516853a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk4hj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-b2knv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.072656 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22rrq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03112a5d-5882-4fac-9ae5-13525ff82fe0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trlt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22rrq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.083352 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.093227 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-host-cni-bin\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.093271 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-host-run-ovn-kubernetes\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.093298 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-etc-openvswitch\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.093313 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-run-openvswitch\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.093334 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-var-lib-openvswitch\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.093349 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-host-kubelet\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.093364 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-systemd-units\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.093377 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-host-run-netns\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.093394 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-run-ovn\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.093411 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-host-slash\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.093412 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-etc-openvswitch\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.093428 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-log-socket\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.093443 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-ovnkube-script-lib\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.093460 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-node-log\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.093463 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-host-cni-bin\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.093476 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.093486 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-host-run-ovn-kubernetes\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.093497 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-var-lib-openvswitch\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.093518 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-systemd-units\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.093512 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-run-openvswitch\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.093550 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-run-ovn\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.093536 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-host-run-netns\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.093491 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-host-kubelet\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.093590 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-log-socket\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.093659 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-env-overrides\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.093662 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-node-log\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.093727 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-host-slash\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.093800 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-run-systemd\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.093832 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.093858 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-ovnkube-config\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.093896 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.093914 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-run-systemd\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.093998 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62rc4\" (UniqueName: \"kubernetes.io/projected/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-kube-api-access-62rc4\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.094029 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-host-cni-netd\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.094048 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-ovn-node-metrics-cert\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.094118 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-host-cni-netd\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.094300 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-env-overrides\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.094341 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-ovnkube-script-lib\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:28 crc kubenswrapper[4802]: E1206 03:40:28.094411 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 03:40:28 crc kubenswrapper[4802]: E1206 03:40:28.094426 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 03:40:28 crc kubenswrapper[4802]: E1206 03:40:28.094437 4802 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:40:28 crc kubenswrapper[4802]: E1206 03:40:28.094484 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 03:40:30.09447297 +0000 UTC m=+22.966382122 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.094644 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-ovnkube-config\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.095652 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.098363 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-ovn-node-metrics-cert\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.106593 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6df38316-e0d3-4018-8d27-3620eba3a68d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpxxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.114106 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62rc4\" (UniqueName: \"kubernetes.io/projected/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-kube-api-access-62rc4\") pod \"ovnkube-node-pvm6q\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.138231 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:28 crc kubenswrapper[4802]: W1206 03:40:28.151614 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeafce4e0_e7fb_4877_b0ab_3283829ba1ef.slice/crio-5d23efa094b6dc407ad7c00fe428e299a5f32feddee8d4369230e195dbcaeb2f WatchSource:0}: Error finding container 5d23efa094b6dc407ad7c00fe428e299a5f32feddee8d4369230e195dbcaeb2f: Status 404 returned error can't find the container with id 5d23efa094b6dc407ad7c00fe428e299a5f32feddee8d4369230e195dbcaeb2f Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.153930 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.194462 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:40:28 crc kubenswrapper[4802]: E1206 03:40:28.194586 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:40:30.194565703 +0000 UTC m=+23.066474855 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.194614 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.194654 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.194681 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:40:28 crc kubenswrapper[4802]: E1206 03:40:28.194784 4802 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 03:40:28 crc kubenswrapper[4802]: E1206 03:40:28.194814 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 03:40:30.194807669 +0000 UTC m=+23.066716821 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 03:40:28 crc kubenswrapper[4802]: E1206 03:40:28.194846 4802 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 03:40:28 crc kubenswrapper[4802]: E1206 03:40:28.194864 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 03:40:30.194859001 +0000 UTC m=+23.066768153 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 03:40:28 crc kubenswrapper[4802]: E1206 03:40:28.194913 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 03:40:28 crc kubenswrapper[4802]: E1206 03:40:28.194923 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 03:40:28 crc kubenswrapper[4802]: E1206 03:40:28.194933 4802 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:40:28 crc kubenswrapper[4802]: E1206 03:40:28.194954 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 03:40:30.194947004 +0000 UTC m=+23.066856156 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.212524 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.257634 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.258627 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.449417 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:40:28 crc kubenswrapper[4802]: E1206 03:40:28.449817 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.449573 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:40:28 crc kubenswrapper[4802]: E1206 03:40:28.449919 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.449529 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:40:28 crc kubenswrapper[4802]: E1206 03:40:28.450005 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.503550 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.602872 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.615727 4802 generic.go:334] "Generic (PLEG): container finished" podID="a3a3bac2-3923-4507-a4de-1556b3ad473e" containerID="ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef" exitCode=0 Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.615851 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xtw2k" event={"ID":"a3a3bac2-3923-4507-a4de-1556b3ad473e","Type":"ContainerDied","Data":"ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef"} Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.615908 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xtw2k" event={"ID":"a3a3bac2-3923-4507-a4de-1556b3ad473e","Type":"ContainerStarted","Data":"bc7693d8229c7be6ca1844217c3752de1a91ad7efcb435c0f889a967dcaa0443"} Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.618049 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" event={"ID":"6df38316-e0d3-4018-8d27-3620eba3a68d","Type":"ContainerStarted","Data":"57381e9589a2dc04a4d2733c37985f42594182128a374b7be0f471ea69138b1a"} Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.618103 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" event={"ID":"6df38316-e0d3-4018-8d27-3620eba3a68d","Type":"ContainerStarted","Data":"3bf5e41701c7a2d45adc4740ceb4ac1785db17bc7cb15daed439f7556812d1e9"} Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.618115 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" event={"ID":"6df38316-e0d3-4018-8d27-3620eba3a68d","Type":"ContainerStarted","Data":"99c5a9de0960265c4986137b196c8e15dcf6bb70c365936915181e63a54dffd0"} Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.625812 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.634817 4802 scope.go:117] "RemoveContainer" containerID="2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc" Dec 06 03:40:28 crc kubenswrapper[4802]: E1206 03:40:28.635042 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.636585 4802 generic.go:334] "Generic (PLEG): container finished" podID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerID="a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7" exitCode=0 Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.636663 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" event={"ID":"eafce4e0-e7fb-4877-b0ab-3283829ba1ef","Type":"ContainerDied","Data":"a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7"} Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.636693 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" event={"ID":"eafce4e0-e7fb-4877-b0ab-3283829ba1ef","Type":"ContainerStarted","Data":"5d23efa094b6dc407ad7c00fe428e299a5f32feddee8d4369230e195dbcaeb2f"} Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.637954 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-22rrq" event={"ID":"03112a5d-5882-4fac-9ae5-13525ff82fe0","Type":"ContainerStarted","Data":"c0205b089b12c42688203bb7ec0f27dc5f9a577dabdb446b811f184046f85dc3"} Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.637999 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-22rrq" event={"ID":"03112a5d-5882-4fac-9ae5-13525ff82fe0","Type":"ContainerStarted","Data":"1959399a50fe6b57274fd584ef680ef2b9a0b207a8b301aa1f483fc5d7b54423"} Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.640041 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.697326 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtw2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3a3bac2-3923-4507-a4de-1556b3ad473e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtw2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.701314 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.716928 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pvm6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.722184 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.726087 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.732335 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1721e985-40bd-4b2d-a0e8-d4e365094ba6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a895e0cc04880c3287ba725ac5484835235f1e066a94ce98fcf4b6133c5cbd2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a21a97385f856ec996105b597d0ae81c56ed28b625ef3f2905e1af6ec8f6a45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://646e01a58908ba2e703aae5eef53e1b981b0f0c3d1d3202487f2a8b94ee3f55d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b602a9bcbfd0f984f1778c0fcdf44058ce528b0d21af6563e76be589971384b5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:40:20Z\\\",\\\"message\\\":\\\"W1206 03:40:09.599454 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1206 03:40:09.599705 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764992409 cert, and key in /tmp/serving-cert-3362951094/serving-signer.crt, /tmp/serving-cert-3362951094/serving-signer.key\\\\nI1206 03:40:09.788947 1 observer_polling.go:159] Starting file observer\\\\nW1206 03:40:09.790951 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1206 03:40:09.791214 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 03:40:09.792215 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3362951094/tls.crt::/tmp/serving-cert-3362951094/tls.key\\\\\\\"\\\\nF1206 03:40:20.204085 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"ver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:40:26.813730 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814130 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:40:26.814139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:40:26.814143 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:40:26.814145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 03:40:26.821403 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821442 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821482 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821496 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 03:40:26.821509 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821514 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 03:40:26.821612 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 03:40:26.821619 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1206 03:40:26.822049 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16a92d501c9e626fe91e4a6793e426d10eb292cfbab766db8b8cc72e5dae94ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.739156 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.741267 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.741384 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.741406 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.741538 4802 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.741606 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-b2knv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a3ae46f-03b5-4999-9d58-c38ee222f68a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76665d4c127445009fd66078827b54607c9ef812c8011d1c0b4cde1c0516853a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk4hj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-b2knv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.753299 4802 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.753783 4802 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.754944 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.754991 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.755003 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.755025 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.755038 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:28Z","lastTransitionTime":"2025-12-06T03:40:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.758617 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22rrq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03112a5d-5882-4fac-9ae5-13525ff82fe0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trlt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22rrq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.759067 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.782047 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:28 crc kubenswrapper[4802]: E1206 03:40:28.809436 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0e01ca3a-758f-44bc-b4a2-30c1bc04fda0\\\",\\\"systemUUID\\\":\\\"c7c815e4-d2a4-4244-b035-f988fc95e215\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.811642 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.815162 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.815205 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.815214 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.815227 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.815236 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:28Z","lastTransitionTime":"2025-12-06T03:40:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.836245 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.837604 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6df38316-e0d3-4018-8d27-3620eba3a68d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpxxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:28 crc kubenswrapper[4802]: E1206 03:40:28.838302 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0e01ca3a-758f-44bc-b4a2-30c1bc04fda0\\\",\\\"systemUUID\\\":\\\"c7c815e4-d2a4-4244-b035-f988fc95e215\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.842072 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.842105 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.842115 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.842128 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.842137 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:28Z","lastTransitionTime":"2025-12-06T03:40:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.851512 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2093b6-c4b5-4db8-be16-fbc859885e8d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0d6e582161f0d9126126e03b27211be22f6b5f797154928c92fa13a8118f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03004a29aefe7fe2452fb9f6663c0807afccb055d7ddbf6b907f305e6a19c81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1801ac64db4ba91c2e8a249e3651d3e50ab62e9f2c87312ce9bf7f38518a82cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6e454507cadbb982a566a3692995aaa3c733ff1cd7fb4870476602190e1eeba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:28 crc kubenswrapper[4802]: E1206 03:40:28.857054 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0e01ca3a-758f-44bc-b4a2-30c1bc04fda0\\\",\\\"systemUUID\\\":\\\"c7c815e4-d2a4-4244-b035-f988fc95e215\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.859934 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.859959 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.859974 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.859988 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.859997 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:28Z","lastTransitionTime":"2025-12-06T03:40:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.863735 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:28 crc kubenswrapper[4802]: E1206 03:40:28.871956 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0e01ca3a-758f-44bc-b4a2-30c1bc04fda0\\\",\\\"systemUUID\\\":\\\"c7c815e4-d2a4-4244-b035-f988fc95e215\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.875104 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.875129 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.875138 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.875152 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.875161 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:28Z","lastTransitionTime":"2025-12-06T03:40:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.877860 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fad9b63d84f41fccf696cd313d7ca7f72b3d3251d6c1b7341573a3fadf1e378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.883383 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 06 03:40:28 crc kubenswrapper[4802]: E1206 03:40:28.889223 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0e01ca3a-758f-44bc-b4a2-30c1bc04fda0\\\",\\\"systemUUID\\\":\\\"c7c815e4-d2a4-4244-b035-f988fc95e215\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:28 crc kubenswrapper[4802]: E1206 03:40:28.889374 4802 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.891030 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.891066 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.891076 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.891117 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.891129 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:28Z","lastTransitionTime":"2025-12-06T03:40:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.895478 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb3f278fff47c900c56d8a563d5ed75a2ae84e2034b0f45aad4083790b9e7946\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b10ca16ac26904ec5e6245344b71d836888cae9b069cae6a7a4831d609752db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.909072 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.921447 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.927940 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-kj7hf"] Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.928279 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-kj7hf" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.930068 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.930274 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.930524 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.933059 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.937571 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6df38316-e0d3-4018-8d27-3620eba3a68d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57381e9589a2dc04a4d2733c37985f42594182128a374b7be0f471ea69138b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bf5e41701c7a2d45adc4740ceb4ac1785db17bc7cb15daed439f7556812d1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpxxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.949004 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2093b6-c4b5-4db8-be16-fbc859885e8d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0d6e582161f0d9126126e03b27211be22f6b5f797154928c92fa13a8118f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03004a29aefe7fe2452fb9f6663c0807afccb055d7ddbf6b907f305e6a19c81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1801ac64db4ba91c2e8a249e3651d3e50ab62e9f2c87312ce9bf7f38518a82cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6e454507cadbb982a566a3692995aaa3c733ff1cd7fb4870476602190e1eeba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.959872 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.974644 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fad9b63d84f41fccf696cd313d7ca7f72b3d3251d6c1b7341573a3fadf1e378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.987510 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb3f278fff47c900c56d8a563d5ed75a2ae84e2034b0f45aad4083790b9e7946\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b10ca16ac26904ec5e6245344b71d836888cae9b069cae6a7a4831d609752db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.993334 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.993370 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.993390 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.993407 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:28 crc kubenswrapper[4802]: I1206 03:40:28.993419 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:28Z","lastTransitionTime":"2025-12-06T03:40:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.000183 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.014186 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtw2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3a3bac2-3923-4507-a4de-1556b3ad473e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtw2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:29Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.033826 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pvm6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:29Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.065465 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1721e985-40bd-4b2d-a0e8-d4e365094ba6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a895e0cc04880c3287ba725ac5484835235f1e066a94ce98fcf4b6133c5cbd2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a21a97385f856ec996105b597d0ae81c56ed28b625ef3f2905e1af6ec8f6a45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://646e01a58908ba2e703aae5eef53e1b981b0f0c3d1d3202487f2a8b94ee3f55d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"ver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:40:26.813730 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814130 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:40:26.814139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:40:26.814143 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:40:26.814145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 03:40:26.821403 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821442 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821482 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821496 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 03:40:26.821509 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821514 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 03:40:26.821612 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 03:40:26.821619 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1206 03:40:26.822049 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16a92d501c9e626fe91e4a6793e426d10eb292cfbab766db8b8cc72e5dae94ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:29Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.096172 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.096516 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.096614 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.096702 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.096807 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:29Z","lastTransitionTime":"2025-12-06T03:40:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.102889 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fd46dfa5-7d4a-49ed-8199-523f10cebdf1-host\") pod \"node-ca-kj7hf\" (UID: \"fd46dfa5-7d4a-49ed-8199-523f10cebdf1\") " pod="openshift-image-registry/node-ca-kj7hf" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.102930 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmkcz\" (UniqueName: \"kubernetes.io/projected/fd46dfa5-7d4a-49ed-8199-523f10cebdf1-kube-api-access-cmkcz\") pod \"node-ca-kj7hf\" (UID: \"fd46dfa5-7d4a-49ed-8199-523f10cebdf1\") " pod="openshift-image-registry/node-ca-kj7hf" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.102963 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/fd46dfa5-7d4a-49ed-8199-523f10cebdf1-serviceca\") pod \"node-ca-kj7hf\" (UID: \"fd46dfa5-7d4a-49ed-8199-523f10cebdf1\") " pod="openshift-image-registry/node-ca-kj7hf" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.104127 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-b2knv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a3ae46f-03b5-4999-9d58-c38ee222f68a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76665d4c127445009fd66078827b54607c9ef812c8011d1c0b4cde1c0516853a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk4hj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-b2knv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:29Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.147827 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22rrq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03112a5d-5882-4fac-9ae5-13525ff82fe0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0205b089b12c42688203bb7ec0f27dc5f9a577dabdb446b811f184046f85dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trlt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22rrq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:29Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.184570 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2093b6-c4b5-4db8-be16-fbc859885e8d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0d6e582161f0d9126126e03b27211be22f6b5f797154928c92fa13a8118f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03004a29aefe7fe2452fb9f6663c0807afccb055d7ddbf6b907f305e6a19c81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1801ac64db4ba91c2e8a249e3651d3e50ab62e9f2c87312ce9bf7f38518a82cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6e454507cadbb982a566a3692995aaa3c733ff1cd7fb4870476602190e1eeba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:29Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.198502 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.198537 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.198546 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.198562 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.198576 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:29Z","lastTransitionTime":"2025-12-06T03:40:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.203912 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fd46dfa5-7d4a-49ed-8199-523f10cebdf1-host\") pod \"node-ca-kj7hf\" (UID: \"fd46dfa5-7d4a-49ed-8199-523f10cebdf1\") " pod="openshift-image-registry/node-ca-kj7hf" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.203949 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmkcz\" (UniqueName: \"kubernetes.io/projected/fd46dfa5-7d4a-49ed-8199-523f10cebdf1-kube-api-access-cmkcz\") pod \"node-ca-kj7hf\" (UID: \"fd46dfa5-7d4a-49ed-8199-523f10cebdf1\") " pod="openshift-image-registry/node-ca-kj7hf" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.203971 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/fd46dfa5-7d4a-49ed-8199-523f10cebdf1-serviceca\") pod \"node-ca-kj7hf\" (UID: \"fd46dfa5-7d4a-49ed-8199-523f10cebdf1\") " pod="openshift-image-registry/node-ca-kj7hf" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.204029 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fd46dfa5-7d4a-49ed-8199-523f10cebdf1-host\") pod \"node-ca-kj7hf\" (UID: \"fd46dfa5-7d4a-49ed-8199-523f10cebdf1\") " pod="openshift-image-registry/node-ca-kj7hf" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.205146 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/fd46dfa5-7d4a-49ed-8199-523f10cebdf1-serviceca\") pod \"node-ca-kj7hf\" (UID: \"fd46dfa5-7d4a-49ed-8199-523f10cebdf1\") " pod="openshift-image-registry/node-ca-kj7hf" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.223140 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:29Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.251156 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmkcz\" (UniqueName: \"kubernetes.io/projected/fd46dfa5-7d4a-49ed-8199-523f10cebdf1-kube-api-access-cmkcz\") pod \"node-ca-kj7hf\" (UID: \"fd46dfa5-7d4a-49ed-8199-523f10cebdf1\") " pod="openshift-image-registry/node-ca-kj7hf" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.287108 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fad9b63d84f41fccf696cd313d7ca7f72b3d3251d6c1b7341573a3fadf1e378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:29Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.300901 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.300929 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.300937 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.300951 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.300961 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:29Z","lastTransitionTime":"2025-12-06T03:40:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.326189 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb3f278fff47c900c56d8a563d5ed75a2ae84e2034b0f45aad4083790b9e7946\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b10ca16ac26904ec5e6245344b71d836888cae9b069cae6a7a4831d609752db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:29Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.363154 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:29Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.403177 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.403214 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.403226 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.403241 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.403252 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:29Z","lastTransitionTime":"2025-12-06T03:40:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.405942 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtw2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3a3bac2-3923-4507-a4de-1556b3ad473e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtw2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:29Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.450950 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pvm6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:29Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.486854 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1721e985-40bd-4b2d-a0e8-d4e365094ba6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a895e0cc04880c3287ba725ac5484835235f1e066a94ce98fcf4b6133c5cbd2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a21a97385f856ec996105b597d0ae81c56ed28b625ef3f2905e1af6ec8f6a45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://646e01a58908ba2e703aae5eef53e1b981b0f0c3d1d3202487f2a8b94ee3f55d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"ver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:40:26.813730 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814130 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:40:26.814139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:40:26.814143 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:40:26.814145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 03:40:26.821403 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821442 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821482 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821496 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 03:40:26.821509 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821514 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 03:40:26.821612 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 03:40:26.821619 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1206 03:40:26.822049 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16a92d501c9e626fe91e4a6793e426d10eb292cfbab766db8b8cc72e5dae94ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:29Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.505563 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.505621 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.505638 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.505662 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.505679 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:29Z","lastTransitionTime":"2025-12-06T03:40:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.522233 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-b2knv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a3ae46f-03b5-4999-9d58-c38ee222f68a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76665d4c127445009fd66078827b54607c9ef812c8011d1c0b4cde1c0516853a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk4hj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-b2knv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:29Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.541203 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-kj7hf" Dec 06 03:40:29 crc kubenswrapper[4802]: W1206 03:40:29.553282 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfd46dfa5_7d4a_49ed_8199_523f10cebdf1.slice/crio-aaafeb6d70ed3d98f6efba970c0e490b681a819a7886aad167229727d18a1856 WatchSource:0}: Error finding container aaafeb6d70ed3d98f6efba970c0e490b681a819a7886aad167229727d18a1856: Status 404 returned error can't find the container with id aaafeb6d70ed3d98f6efba970c0e490b681a819a7886aad167229727d18a1856 Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.566482 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22rrq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03112a5d-5882-4fac-9ae5-13525ff82fe0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0205b089b12c42688203bb7ec0f27dc5f9a577dabdb446b811f184046f85dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trlt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22rrq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:29Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.604458 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:29Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.608577 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.608633 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.608651 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.608672 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.608688 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:29Z","lastTransitionTime":"2025-12-06T03:40:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.647247 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" event={"ID":"eafce4e0-e7fb-4877-b0ab-3283829ba1ef","Type":"ContainerStarted","Data":"0a2357f37497fde64937eca7b38693893af745ab278653ebc79406e122a6d4ff"} Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.647298 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" event={"ID":"eafce4e0-e7fb-4877-b0ab-3283829ba1ef","Type":"ContainerStarted","Data":"8941bd25107333b264d3f081a811fdfeea23ad92193fc61367665fd2cd1209c5"} Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.647312 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" event={"ID":"eafce4e0-e7fb-4877-b0ab-3283829ba1ef","Type":"ContainerStarted","Data":"a158aebd701ef46225d3e37109ae28ab58d107087bf06650c4b6d0055a3bbba5"} Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.647325 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" event={"ID":"eafce4e0-e7fb-4877-b0ab-3283829ba1ef","Type":"ContainerStarted","Data":"7730de4f033561ad77046e028cd1a3f0bfe2775387cbdfa35f47ea79909ec4cd"} Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.647361 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" event={"ID":"eafce4e0-e7fb-4877-b0ab-3283829ba1ef","Type":"ContainerStarted","Data":"38f9cd4dc75e63f09f60b48b6da284643910bb32ba51f2a808cb4d9803f30042"} Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.647374 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" event={"ID":"eafce4e0-e7fb-4877-b0ab-3283829ba1ef","Type":"ContainerStarted","Data":"5cf6796f65ee9c5bd0c18f205697135d17677c5daf3ba378cc992cba795e0188"} Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.648415 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:29Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.649912 4802 generic.go:334] "Generic (PLEG): container finished" podID="a3a3bac2-3923-4507-a4de-1556b3ad473e" containerID="e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce" exitCode=0 Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.649989 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xtw2k" event={"ID":"a3a3bac2-3923-4507-a4de-1556b3ad473e","Type":"ContainerDied","Data":"e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce"} Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.651971 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-kj7hf" event={"ID":"fd46dfa5-7d4a-49ed-8199-523f10cebdf1","Type":"ContainerStarted","Data":"aaafeb6d70ed3d98f6efba970c0e490b681a819a7886aad167229727d18a1856"} Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.653919 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"d2466610962332c6ca120fe055e4e8a3fc623088119dc1a412e06b745176f89e"} Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.685569 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6df38316-e0d3-4018-8d27-3620eba3a68d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57381e9589a2dc04a4d2733c37985f42594182128a374b7be0f471ea69138b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bf5e41701c7a2d45adc4740ceb4ac1785db17bc7cb15daed439f7556812d1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpxxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:29Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.711612 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.711696 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.711709 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.711771 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.711789 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:29Z","lastTransitionTime":"2025-12-06T03:40:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.722716 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kj7hf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd46dfa5-7d4a-49ed-8199-523f10cebdf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmkcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kj7hf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:29Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.764575 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtw2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3a3bac2-3923-4507-a4de-1556b3ad473e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtw2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:29Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.814314 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.814367 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.814383 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.814402 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.814414 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:29Z","lastTransitionTime":"2025-12-06T03:40:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.814341 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pvm6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:29Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.846262 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:29Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.886567 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1721e985-40bd-4b2d-a0e8-d4e365094ba6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a895e0cc04880c3287ba725ac5484835235f1e066a94ce98fcf4b6133c5cbd2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a21a97385f856ec996105b597d0ae81c56ed28b625ef3f2905e1af6ec8f6a45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://646e01a58908ba2e703aae5eef53e1b981b0f0c3d1d3202487f2a8b94ee3f55d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"ver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:40:26.813730 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814130 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:40:26.814139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:40:26.814143 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:40:26.814145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 03:40:26.821403 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821442 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821482 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821496 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 03:40:26.821509 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821514 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 03:40:26.821612 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 03:40:26.821619 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1206 03:40:26.822049 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16a92d501c9e626fe91e4a6793e426d10eb292cfbab766db8b8cc72e5dae94ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:29Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.916677 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.916711 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.916720 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.916734 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.916743 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:29Z","lastTransitionTime":"2025-12-06T03:40:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.922629 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-b2knv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a3ae46f-03b5-4999-9d58-c38ee222f68a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76665d4c127445009fd66078827b54607c9ef812c8011d1c0b4cde1c0516853a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk4hj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-b2knv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:29Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:29 crc kubenswrapper[4802]: I1206 03:40:29.964122 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22rrq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03112a5d-5882-4fac-9ae5-13525ff82fe0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0205b089b12c42688203bb7ec0f27dc5f9a577dabdb446b811f184046f85dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trlt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22rrq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:29Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.008509 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:30Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.018565 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.018620 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.018638 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.018665 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.018687 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:30Z","lastTransitionTime":"2025-12-06T03:40:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.046618 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6df38316-e0d3-4018-8d27-3620eba3a68d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57381e9589a2dc04a4d2733c37985f42594182128a374b7be0f471ea69138b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bf5e41701c7a2d45adc4740ceb4ac1785db17bc7cb15daed439f7556812d1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpxxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:30Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.082046 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kj7hf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd46dfa5-7d4a-49ed-8199-523f10cebdf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmkcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kj7hf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:30Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.112957 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:40:30 crc kubenswrapper[4802]: E1206 03:40:30.113205 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 03:40:30 crc kubenswrapper[4802]: E1206 03:40:30.113241 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 03:40:30 crc kubenswrapper[4802]: E1206 03:40:30.113254 4802 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:40:30 crc kubenswrapper[4802]: E1206 03:40:30.113317 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 03:40:34.113296668 +0000 UTC m=+26.985205820 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.121163 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.121201 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.121213 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.121229 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.121241 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:30Z","lastTransitionTime":"2025-12-06T03:40:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.122798 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2466610962332c6ca120fe055e4e8a3fc623088119dc1a412e06b745176f89e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:30Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.166128 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2093b6-c4b5-4db8-be16-fbc859885e8d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0d6e582161f0d9126126e03b27211be22f6b5f797154928c92fa13a8118f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03004a29aefe7fe2452fb9f6663c0807afccb055d7ddbf6b907f305e6a19c81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1801ac64db4ba91c2e8a249e3651d3e50ab62e9f2c87312ce9bf7f38518a82cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6e454507cadbb982a566a3692995aaa3c733ff1cd7fb4870476602190e1eeba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:30Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.204455 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:30Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.213797 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.213897 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:40:30 crc kubenswrapper[4802]: E1206 03:40:30.213949 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:40:34.213921216 +0000 UTC m=+27.085830368 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.213991 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:40:30 crc kubenswrapper[4802]: E1206 03:40:30.214025 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.214033 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:40:30 crc kubenswrapper[4802]: E1206 03:40:30.214044 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 03:40:30 crc kubenswrapper[4802]: E1206 03:40:30.214058 4802 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:40:30 crc kubenswrapper[4802]: E1206 03:40:30.214110 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 03:40:34.21409147 +0000 UTC m=+27.086000702 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:40:30 crc kubenswrapper[4802]: E1206 03:40:30.214125 4802 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 03:40:30 crc kubenswrapper[4802]: E1206 03:40:30.214136 4802 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 03:40:30 crc kubenswrapper[4802]: E1206 03:40:30.214165 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 03:40:34.214155562 +0000 UTC m=+27.086064714 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 03:40:30 crc kubenswrapper[4802]: E1206 03:40:30.214184 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 03:40:34.214175743 +0000 UTC m=+27.086085005 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.222839 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.222875 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.222884 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.222898 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.222909 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:30Z","lastTransitionTime":"2025-12-06T03:40:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.243606 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fad9b63d84f41fccf696cd313d7ca7f72b3d3251d6c1b7341573a3fadf1e378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:30Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.283661 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb3f278fff47c900c56d8a563d5ed75a2ae84e2034b0f45aad4083790b9e7946\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b10ca16ac26904ec5e6245344b71d836888cae9b069cae6a7a4831d609752db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:30Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.325067 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.325110 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.325124 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.325140 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.325152 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:30Z","lastTransitionTime":"2025-12-06T03:40:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.427710 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.427783 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.427803 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.427828 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.427844 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:30Z","lastTransitionTime":"2025-12-06T03:40:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.450155 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.450169 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:40:30 crc kubenswrapper[4802]: E1206 03:40:30.450280 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.450155 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:40:30 crc kubenswrapper[4802]: E1206 03:40:30.450482 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:40:30 crc kubenswrapper[4802]: E1206 03:40:30.450622 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.530374 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.530442 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.530460 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.530487 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.530509 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:30Z","lastTransitionTime":"2025-12-06T03:40:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.633904 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.634156 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.634182 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.634210 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.634234 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:30Z","lastTransitionTime":"2025-12-06T03:40:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.661068 4802 generic.go:334] "Generic (PLEG): container finished" podID="a3a3bac2-3923-4507-a4de-1556b3ad473e" containerID="eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd" exitCode=0 Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.661161 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xtw2k" event={"ID":"a3a3bac2-3923-4507-a4de-1556b3ad473e","Type":"ContainerDied","Data":"eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd"} Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.664461 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-kj7hf" event={"ID":"fd46dfa5-7d4a-49ed-8199-523f10cebdf1","Type":"ContainerStarted","Data":"d5eee8a0cf96ceb3785dac9df31bf8d31715d1c88fbea17c5e582b0a4554bb63"} Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.702342 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:30Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.737218 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.737255 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.737272 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.737293 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.737305 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:30Z","lastTransitionTime":"2025-12-06T03:40:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.756046 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtw2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3a3bac2-3923-4507-a4de-1556b3ad473e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtw2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:30Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.791668 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pvm6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:30Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.804893 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1721e985-40bd-4b2d-a0e8-d4e365094ba6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a895e0cc04880c3287ba725ac5484835235f1e066a94ce98fcf4b6133c5cbd2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a21a97385f856ec996105b597d0ae81c56ed28b625ef3f2905e1af6ec8f6a45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://646e01a58908ba2e703aae5eef53e1b981b0f0c3d1d3202487f2a8b94ee3f55d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"ver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:40:26.813730 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814130 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:40:26.814139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:40:26.814143 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:40:26.814145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 03:40:26.821403 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821442 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821482 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821496 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 03:40:26.821509 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821514 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 03:40:26.821612 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 03:40:26.821619 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1206 03:40:26.822049 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16a92d501c9e626fe91e4a6793e426d10eb292cfbab766db8b8cc72e5dae94ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:30Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.815304 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-b2knv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a3ae46f-03b5-4999-9d58-c38ee222f68a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76665d4c127445009fd66078827b54607c9ef812c8011d1c0b4cde1c0516853a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk4hj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-b2knv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:30Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.829770 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22rrq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03112a5d-5882-4fac-9ae5-13525ff82fe0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0205b089b12c42688203bb7ec0f27dc5f9a577dabdb446b811f184046f85dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trlt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22rrq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:30Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.839734 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.839793 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.839807 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.839825 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.839836 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:30Z","lastTransitionTime":"2025-12-06T03:40:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.841295 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2466610962332c6ca120fe055e4e8a3fc623088119dc1a412e06b745176f89e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:30Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.855705 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:30Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.867736 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6df38316-e0d3-4018-8d27-3620eba3a68d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57381e9589a2dc04a4d2733c37985f42594182128a374b7be0f471ea69138b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bf5e41701c7a2d45adc4740ceb4ac1785db17bc7cb15daed439f7556812d1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpxxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:30Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.876883 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kj7hf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd46dfa5-7d4a-49ed-8199-523f10cebdf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmkcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kj7hf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:30Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.888699 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2093b6-c4b5-4db8-be16-fbc859885e8d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0d6e582161f0d9126126e03b27211be22f6b5f797154928c92fa13a8118f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03004a29aefe7fe2452fb9f6663c0807afccb055d7ddbf6b907f305e6a19c81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1801ac64db4ba91c2e8a249e3651d3e50ab62e9f2c87312ce9bf7f38518a82cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6e454507cadbb982a566a3692995aaa3c733ff1cd7fb4870476602190e1eeba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:30Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.900495 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:30Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.912790 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fad9b63d84f41fccf696cd313d7ca7f72b3d3251d6c1b7341573a3fadf1e378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:30Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.924140 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb3f278fff47c900c56d8a563d5ed75a2ae84e2034b0f45aad4083790b9e7946\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b10ca16ac26904ec5e6245344b71d836888cae9b069cae6a7a4831d609752db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:30Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.936023 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2093b6-c4b5-4db8-be16-fbc859885e8d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0d6e582161f0d9126126e03b27211be22f6b5f797154928c92fa13a8118f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03004a29aefe7fe2452fb9f6663c0807afccb055d7ddbf6b907f305e6a19c81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1801ac64db4ba91c2e8a249e3651d3e50ab62e9f2c87312ce9bf7f38518a82cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6e454507cadbb982a566a3692995aaa3c733ff1cd7fb4870476602190e1eeba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:30Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.941929 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.941979 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.941994 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.942013 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.942024 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:30Z","lastTransitionTime":"2025-12-06T03:40:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.947924 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:30Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.963989 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fad9b63d84f41fccf696cd313d7ca7f72b3d3251d6c1b7341573a3fadf1e378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:30Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.980937 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 03:40:30 crc kubenswrapper[4802]: I1206 03:40:30.984928 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.007875 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb3f278fff47c900c56d8a563d5ed75a2ae84e2034b0f45aad4083790b9e7946\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b10ca16ac26904ec5e6245344b71d836888cae9b069cae6a7a4831d609752db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.045321 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.045372 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.045389 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.045415 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.045438 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:31Z","lastTransitionTime":"2025-12-06T03:40:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.046491 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.093974 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtw2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3a3bac2-3923-4507-a4de-1556b3ad473e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtw2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.135006 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pvm6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.148810 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.148906 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.148932 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.148964 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.148987 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:31Z","lastTransitionTime":"2025-12-06T03:40:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.172513 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1721e985-40bd-4b2d-a0e8-d4e365094ba6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a895e0cc04880c3287ba725ac5484835235f1e066a94ce98fcf4b6133c5cbd2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a21a97385f856ec996105b597d0ae81c56ed28b625ef3f2905e1af6ec8f6a45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://646e01a58908ba2e703aae5eef53e1b981b0f0c3d1d3202487f2a8b94ee3f55d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"ver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:40:26.813730 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814130 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:40:26.814139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:40:26.814143 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:40:26.814145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 03:40:26.821403 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821442 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821482 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821496 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 03:40:26.821509 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821514 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 03:40:26.821612 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 03:40:26.821619 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1206 03:40:26.822049 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16a92d501c9e626fe91e4a6793e426d10eb292cfbab766db8b8cc72e5dae94ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.207070 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-b2knv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a3ae46f-03b5-4999-9d58-c38ee222f68a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76665d4c127445009fd66078827b54607c9ef812c8011d1c0b4cde1c0516853a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk4hj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-b2knv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.245724 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22rrq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03112a5d-5882-4fac-9ae5-13525ff82fe0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0205b089b12c42688203bb7ec0f27dc5f9a577dabdb446b811f184046f85dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trlt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22rrq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.251866 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.251919 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.251935 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.251980 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.251997 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:31Z","lastTransitionTime":"2025-12-06T03:40:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.285666 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2466610962332c6ca120fe055e4e8a3fc623088119dc1a412e06b745176f89e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.328979 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.354822 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.354865 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.354877 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.354895 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.354908 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:31Z","lastTransitionTime":"2025-12-06T03:40:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.370710 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6df38316-e0d3-4018-8d27-3620eba3a68d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57381e9589a2dc04a4d2733c37985f42594182128a374b7be0f471ea69138b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bf5e41701c7a2d45adc4740ceb4ac1785db17bc7cb15daed439f7556812d1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpxxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.404228 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kj7hf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd46dfa5-7d4a-49ed-8199-523f10cebdf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5eee8a0cf96ceb3785dac9df31bf8d31715d1c88fbea17c5e582b0a4554bb63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmkcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kj7hf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.453372 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2466610962332c6ca120fe055e4e8a3fc623088119dc1a412e06b745176f89e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.458063 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.458120 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.458144 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.458177 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.458203 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:31Z","lastTransitionTime":"2025-12-06T03:40:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.490193 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.523879 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6df38316-e0d3-4018-8d27-3620eba3a68d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57381e9589a2dc04a4d2733c37985f42594182128a374b7be0f471ea69138b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bf5e41701c7a2d45adc4740ceb4ac1785db17bc7cb15daed439f7556812d1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpxxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.562232 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.562282 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.562296 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.562319 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.562335 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:31Z","lastTransitionTime":"2025-12-06T03:40:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.565417 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kj7hf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd46dfa5-7d4a-49ed-8199-523f10cebdf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5eee8a0cf96ceb3785dac9df31bf8d31715d1c88fbea17c5e582b0a4554bb63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmkcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kj7hf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.606209 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2093b6-c4b5-4db8-be16-fbc859885e8d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0d6e582161f0d9126126e03b27211be22f6b5f797154928c92fa13a8118f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03004a29aefe7fe2452fb9f6663c0807afccb055d7ddbf6b907f305e6a19c81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1801ac64db4ba91c2e8a249e3651d3e50ab62e9f2c87312ce9bf7f38518a82cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6e454507cadbb982a566a3692995aaa3c733ff1cd7fb4870476602190e1eeba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.642717 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.664551 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.664591 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.664600 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.664614 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.664623 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:31Z","lastTransitionTime":"2025-12-06T03:40:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.668953 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" event={"ID":"eafce4e0-e7fb-4877-b0ab-3283829ba1ef","Type":"ContainerStarted","Data":"b3cb210bfc97a439a7ff98c02c59f8ad16bc400dc214a616a44b43a8b01e2782"} Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.671083 4802 generic.go:334] "Generic (PLEG): container finished" podID="a3a3bac2-3923-4507-a4de-1556b3ad473e" containerID="9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526" exitCode=0 Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.671111 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xtw2k" event={"ID":"a3a3bac2-3923-4507-a4de-1556b3ad473e","Type":"ContainerDied","Data":"9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526"} Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.684958 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fad9b63d84f41fccf696cd313d7ca7f72b3d3251d6c1b7341573a3fadf1e378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.724246 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb3f278fff47c900c56d8a563d5ed75a2ae84e2034b0f45aad4083790b9e7946\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b10ca16ac26904ec5e6245344b71d836888cae9b069cae6a7a4831d609752db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.766441 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.767190 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.767221 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.767231 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.767246 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.767258 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:31Z","lastTransitionTime":"2025-12-06T03:40:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.805105 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtw2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3a3bac2-3923-4507-a4de-1556b3ad473e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtw2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.848206 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pvm6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.869810 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.869875 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.869894 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.869920 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.869939 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:31Z","lastTransitionTime":"2025-12-06T03:40:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.889079 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1721e985-40bd-4b2d-a0e8-d4e365094ba6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a895e0cc04880c3287ba725ac5484835235f1e066a94ce98fcf4b6133c5cbd2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a21a97385f856ec996105b597d0ae81c56ed28b625ef3f2905e1af6ec8f6a45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://646e01a58908ba2e703aae5eef53e1b981b0f0c3d1d3202487f2a8b94ee3f55d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"ver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:40:26.813730 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814130 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:40:26.814139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:40:26.814143 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:40:26.814145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 03:40:26.821403 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821442 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821482 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821496 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 03:40:26.821509 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821514 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 03:40:26.821612 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 03:40:26.821619 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1206 03:40:26.822049 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16a92d501c9e626fe91e4a6793e426d10eb292cfbab766db8b8cc72e5dae94ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.925186 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-b2knv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a3ae46f-03b5-4999-9d58-c38ee222f68a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76665d4c127445009fd66078827b54607c9ef812c8011d1c0b4cde1c0516853a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk4hj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-b2knv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.971686 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.971795 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.971850 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.971881 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.971900 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:31Z","lastTransitionTime":"2025-12-06T03:40:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:31 crc kubenswrapper[4802]: I1206 03:40:31.988529 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22rrq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03112a5d-5882-4fac-9ae5-13525ff82fe0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0205b089b12c42688203bb7ec0f27dc5f9a577dabdb446b811f184046f85dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trlt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22rrq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.008838 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:32Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.053880 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtw2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3a3bac2-3923-4507-a4de-1556b3ad473e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtw2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:32Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.074541 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.074603 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.074622 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.074648 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.074664 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:32Z","lastTransitionTime":"2025-12-06T03:40:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.092205 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pvm6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:32Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.130474 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-b2knv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a3ae46f-03b5-4999-9d58-c38ee222f68a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76665d4c127445009fd66078827b54607c9ef812c8011d1c0b4cde1c0516853a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk4hj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-b2knv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:32Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.172621 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22rrq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03112a5d-5882-4fac-9ae5-13525ff82fe0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0205b089b12c42688203bb7ec0f27dc5f9a577dabdb446b811f184046f85dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trlt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22rrq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:32Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.177881 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.177952 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.177969 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.177992 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.178010 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:32Z","lastTransitionTime":"2025-12-06T03:40:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.209053 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1721e985-40bd-4b2d-a0e8-d4e365094ba6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a895e0cc04880c3287ba725ac5484835235f1e066a94ce98fcf4b6133c5cbd2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a21a97385f856ec996105b597d0ae81c56ed28b625ef3f2905e1af6ec8f6a45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://646e01a58908ba2e703aae5eef53e1b981b0f0c3d1d3202487f2a8b94ee3f55d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"ver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:40:26.813730 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814130 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:40:26.814139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:40:26.814143 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:40:26.814145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 03:40:26.821403 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821442 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821482 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821496 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 03:40:26.821509 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821514 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 03:40:26.821612 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 03:40:26.821619 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1206 03:40:26.822049 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16a92d501c9e626fe91e4a6793e426d10eb292cfbab766db8b8cc72e5dae94ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:32Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.245183 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kj7hf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd46dfa5-7d4a-49ed-8199-523f10cebdf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5eee8a0cf96ceb3785dac9df31bf8d31715d1c88fbea17c5e582b0a4554bb63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmkcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kj7hf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:32Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.295135 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2466610962332c6ca120fe055e4e8a3fc623088119dc1a412e06b745176f89e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:32Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.297024 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.297055 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.297066 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.297084 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.297096 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:32Z","lastTransitionTime":"2025-12-06T03:40:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.331553 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:32Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.366160 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6df38316-e0d3-4018-8d27-3620eba3a68d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57381e9589a2dc04a4d2733c37985f42594182128a374b7be0f471ea69138b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bf5e41701c7a2d45adc4740ceb4ac1785db17bc7cb15daed439f7556812d1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpxxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:32Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.400385 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.400433 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.400445 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.400461 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.400475 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:32Z","lastTransitionTime":"2025-12-06T03:40:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.406767 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fad9b63d84f41fccf696cd313d7ca7f72b3d3251d6c1b7341573a3fadf1e378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:32Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.446841 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb3f278fff47c900c56d8a563d5ed75a2ae84e2034b0f45aad4083790b9e7946\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b10ca16ac26904ec5e6245344b71d836888cae9b069cae6a7a4831d609752db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:32Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.450015 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.450033 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.450033 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:40:32 crc kubenswrapper[4802]: E1206 03:40:32.450218 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:40:32 crc kubenswrapper[4802]: E1206 03:40:32.450311 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:40:32 crc kubenswrapper[4802]: E1206 03:40:32.450362 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.483793 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2093b6-c4b5-4db8-be16-fbc859885e8d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0d6e582161f0d9126126e03b27211be22f6b5f797154928c92fa13a8118f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03004a29aefe7fe2452fb9f6663c0807afccb055d7ddbf6b907f305e6a19c81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1801ac64db4ba91c2e8a249e3651d3e50ab62e9f2c87312ce9bf7f38518a82cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6e454507cadbb982a566a3692995aaa3c733ff1cd7fb4870476602190e1eeba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:32Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.502127 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.502161 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.502170 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.502183 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.502192 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:32Z","lastTransitionTime":"2025-12-06T03:40:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.526271 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:32Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.604453 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.604521 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.604539 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.604565 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.604583 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:32Z","lastTransitionTime":"2025-12-06T03:40:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.677357 4802 generic.go:334] "Generic (PLEG): container finished" podID="a3a3bac2-3923-4507-a4de-1556b3ad473e" containerID="24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5" exitCode=0 Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.677408 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xtw2k" event={"ID":"a3a3bac2-3923-4507-a4de-1556b3ad473e","Type":"ContainerDied","Data":"24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5"} Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.691733 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6df38316-e0d3-4018-8d27-3620eba3a68d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57381e9589a2dc04a4d2733c37985f42594182128a374b7be0f471ea69138b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bf5e41701c7a2d45adc4740ceb4ac1785db17bc7cb15daed439f7556812d1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpxxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:32Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.705906 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kj7hf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd46dfa5-7d4a-49ed-8199-523f10cebdf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5eee8a0cf96ceb3785dac9df31bf8d31715d1c88fbea17c5e582b0a4554bb63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmkcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kj7hf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:32Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.707365 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.707407 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.707418 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.707433 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.707443 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:32Z","lastTransitionTime":"2025-12-06T03:40:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.721851 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2466610962332c6ca120fe055e4e8a3fc623088119dc1a412e06b745176f89e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:32Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.740040 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:32Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.754516 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:32Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.767820 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fad9b63d84f41fccf696cd313d7ca7f72b3d3251d6c1b7341573a3fadf1e378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:32Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.808975 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb3f278fff47c900c56d8a563d5ed75a2ae84e2034b0f45aad4083790b9e7946\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b10ca16ac26904ec5e6245344b71d836888cae9b069cae6a7a4831d609752db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:32Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.810719 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.810815 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.810833 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.810859 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.810883 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:32Z","lastTransitionTime":"2025-12-06T03:40:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.846563 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2093b6-c4b5-4db8-be16-fbc859885e8d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0d6e582161f0d9126126e03b27211be22f6b5f797154928c92fa13a8118f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03004a29aefe7fe2452fb9f6663c0807afccb055d7ddbf6b907f305e6a19c81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1801ac64db4ba91c2e8a249e3651d3e50ab62e9f2c87312ce9bf7f38518a82cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6e454507cadbb982a566a3692995aaa3c733ff1cd7fb4870476602190e1eeba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:32Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.889881 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pvm6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:32Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.914610 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.914650 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.914665 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.914687 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.914699 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:32Z","lastTransitionTime":"2025-12-06T03:40:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.931690 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:32Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:32 crc kubenswrapper[4802]: I1206 03:40:32.969285 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtw2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3a3bac2-3923-4507-a4de-1556b3ad473e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtw2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:32Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.007860 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-b2knv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a3ae46f-03b5-4999-9d58-c38ee222f68a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76665d4c127445009fd66078827b54607c9ef812c8011d1c0b4cde1c0516853a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk4hj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-b2knv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:33Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.017401 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.017464 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.017481 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.017508 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.017527 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:33Z","lastTransitionTime":"2025-12-06T03:40:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.051887 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22rrq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03112a5d-5882-4fac-9ae5-13525ff82fe0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0205b089b12c42688203bb7ec0f27dc5f9a577dabdb446b811f184046f85dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trlt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22rrq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:33Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.093802 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1721e985-40bd-4b2d-a0e8-d4e365094ba6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a895e0cc04880c3287ba725ac5484835235f1e066a94ce98fcf4b6133c5cbd2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a21a97385f856ec996105b597d0ae81c56ed28b625ef3f2905e1af6ec8f6a45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://646e01a58908ba2e703aae5eef53e1b981b0f0c3d1d3202487f2a8b94ee3f55d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"ver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:40:26.813730 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814130 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:40:26.814139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:40:26.814143 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:40:26.814145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 03:40:26.821403 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821442 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821482 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821496 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 03:40:26.821509 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821514 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 03:40:26.821612 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 03:40:26.821619 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1206 03:40:26.822049 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16a92d501c9e626fe91e4a6793e426d10eb292cfbab766db8b8cc72e5dae94ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:33Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.121048 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.121107 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.121119 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.121140 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.121152 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:33Z","lastTransitionTime":"2025-12-06T03:40:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.223637 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.223806 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.223832 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.223864 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.223888 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:33Z","lastTransitionTime":"2025-12-06T03:40:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.327364 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.327433 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.327450 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.327478 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.327495 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:33Z","lastTransitionTime":"2025-12-06T03:40:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.431426 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.431499 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.431522 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.431550 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.431569 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:33Z","lastTransitionTime":"2025-12-06T03:40:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.534552 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.534612 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.534628 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.534651 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.534667 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:33Z","lastTransitionTime":"2025-12-06T03:40:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.638142 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.638460 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.638482 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.638514 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.638537 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:33Z","lastTransitionTime":"2025-12-06T03:40:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.688618 4802 generic.go:334] "Generic (PLEG): container finished" podID="a3a3bac2-3923-4507-a4de-1556b3ad473e" containerID="fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85" exitCode=0 Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.688683 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xtw2k" event={"ID":"a3a3bac2-3923-4507-a4de-1556b3ad473e","Type":"ContainerDied","Data":"fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85"} Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.720492 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1721e985-40bd-4b2d-a0e8-d4e365094ba6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a895e0cc04880c3287ba725ac5484835235f1e066a94ce98fcf4b6133c5cbd2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a21a97385f856ec996105b597d0ae81c56ed28b625ef3f2905e1af6ec8f6a45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://646e01a58908ba2e703aae5eef53e1b981b0f0c3d1d3202487f2a8b94ee3f55d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"ver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:40:26.813730 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814130 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:40:26.814139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:40:26.814143 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:40:26.814145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 03:40:26.821403 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821442 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821482 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821496 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 03:40:26.821509 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821514 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 03:40:26.821612 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 03:40:26.821619 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1206 03:40:26.822049 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16a92d501c9e626fe91e4a6793e426d10eb292cfbab766db8b8cc72e5dae94ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:33Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.732864 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-b2knv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a3ae46f-03b5-4999-9d58-c38ee222f68a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76665d4c127445009fd66078827b54607c9ef812c8011d1c0b4cde1c0516853a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk4hj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-b2knv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:33Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.740623 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.740682 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.740699 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.740724 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.740742 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:33Z","lastTransitionTime":"2025-12-06T03:40:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.746323 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22rrq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03112a5d-5882-4fac-9ae5-13525ff82fe0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0205b089b12c42688203bb7ec0f27dc5f9a577dabdb446b811f184046f85dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trlt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22rrq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:33Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.759890 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2466610962332c6ca120fe055e4e8a3fc623088119dc1a412e06b745176f89e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:33Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.774738 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:33Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.788878 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6df38316-e0d3-4018-8d27-3620eba3a68d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57381e9589a2dc04a4d2733c37985f42594182128a374b7be0f471ea69138b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bf5e41701c7a2d45adc4740ceb4ac1785db17bc7cb15daed439f7556812d1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpxxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:33Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.801269 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kj7hf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd46dfa5-7d4a-49ed-8199-523f10cebdf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5eee8a0cf96ceb3785dac9df31bf8d31715d1c88fbea17c5e582b0a4554bb63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmkcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kj7hf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:33Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.816577 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2093b6-c4b5-4db8-be16-fbc859885e8d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0d6e582161f0d9126126e03b27211be22f6b5f797154928c92fa13a8118f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03004a29aefe7fe2452fb9f6663c0807afccb055d7ddbf6b907f305e6a19c81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1801ac64db4ba91c2e8a249e3651d3e50ab62e9f2c87312ce9bf7f38518a82cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6e454507cadbb982a566a3692995aaa3c733ff1cd7fb4870476602190e1eeba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:33Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.829826 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:33Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.843436 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.843492 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.843501 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.843515 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.843526 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:33Z","lastTransitionTime":"2025-12-06T03:40:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.843726 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fad9b63d84f41fccf696cd313d7ca7f72b3d3251d6c1b7341573a3fadf1e378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:33Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.859490 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb3f278fff47c900c56d8a563d5ed75a2ae84e2034b0f45aad4083790b9e7946\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b10ca16ac26904ec5e6245344b71d836888cae9b069cae6a7a4831d609752db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:33Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.872377 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:33Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.889303 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtw2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3a3bac2-3923-4507-a4de-1556b3ad473e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtw2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:33Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.915616 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pvm6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:33Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.946886 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.946926 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.946935 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.946952 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:33 crc kubenswrapper[4802]: I1206 03:40:33.946963 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:33Z","lastTransitionTime":"2025-12-06T03:40:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.049573 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.049613 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.049625 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.049644 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.049656 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:34Z","lastTransitionTime":"2025-12-06T03:40:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.151493 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.151537 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.151550 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.151569 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.151590 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:34Z","lastTransitionTime":"2025-12-06T03:40:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.156310 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:40:34 crc kubenswrapper[4802]: E1206 03:40:34.156476 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 03:40:34 crc kubenswrapper[4802]: E1206 03:40:34.156508 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 03:40:34 crc kubenswrapper[4802]: E1206 03:40:34.156528 4802 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:40:34 crc kubenswrapper[4802]: E1206 03:40:34.156598 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 03:40:42.15657275 +0000 UTC m=+35.028481902 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.254339 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.254401 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.254418 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.254440 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.254458 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:34Z","lastTransitionTime":"2025-12-06T03:40:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.256859 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:40:34 crc kubenswrapper[4802]: E1206 03:40:34.257047 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:40:42.257015782 +0000 UTC m=+35.128924944 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.257127 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.257169 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.257212 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:40:34 crc kubenswrapper[4802]: E1206 03:40:34.257322 4802 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 03:40:34 crc kubenswrapper[4802]: E1206 03:40:34.257339 4802 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 03:40:34 crc kubenswrapper[4802]: E1206 03:40:34.257366 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 03:40:34 crc kubenswrapper[4802]: E1206 03:40:34.257390 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 03:40:34 crc kubenswrapper[4802]: E1206 03:40:34.257403 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 03:40:42.257387823 +0000 UTC m=+35.129296985 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 03:40:34 crc kubenswrapper[4802]: E1206 03:40:34.257407 4802 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:40:34 crc kubenswrapper[4802]: E1206 03:40:34.257443 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 03:40:42.257415404 +0000 UTC m=+35.129324596 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 03:40:34 crc kubenswrapper[4802]: E1206 03:40:34.257474 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 03:40:42.257461045 +0000 UTC m=+35.129370247 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.356443 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.356476 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.356485 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.356500 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.356509 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:34Z","lastTransitionTime":"2025-12-06T03:40:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.449789 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.449791 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.449919 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:40:34 crc kubenswrapper[4802]: E1206 03:40:34.450004 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:40:34 crc kubenswrapper[4802]: E1206 03:40:34.450114 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:40:34 crc kubenswrapper[4802]: E1206 03:40:34.450234 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.459211 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.459249 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.459261 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.459277 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.459289 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:34Z","lastTransitionTime":"2025-12-06T03:40:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.561890 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.561949 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.561965 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.561991 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.562008 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:34Z","lastTransitionTime":"2025-12-06T03:40:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.664971 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.665007 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.665015 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.665027 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.665036 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:34Z","lastTransitionTime":"2025-12-06T03:40:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.697049 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xtw2k" event={"ID":"a3a3bac2-3923-4507-a4de-1556b3ad473e","Type":"ContainerStarted","Data":"4fed5ee4189d54456a3216f4fdb1ecf003fa47e29dbfe4b2b2f301edfb897d6f"} Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.704126 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" event={"ID":"eafce4e0-e7fb-4877-b0ab-3283829ba1ef","Type":"ContainerStarted","Data":"e5e40ff23f812195841d652697dded43c59c78b48a80b56d600da57773d73c0d"} Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.704483 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.704807 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.704941 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.714587 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:34Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.735833 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtw2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3a3bac2-3923-4507-a4de-1556b3ad473e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fed5ee4189d54456a3216f4fdb1ecf003fa47e29dbfe4b2b2f301edfb897d6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtw2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:34Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.737595 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.740540 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.759276 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pvm6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:34Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.775593 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.776075 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.776098 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.776128 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.776149 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:34Z","lastTransitionTime":"2025-12-06T03:40:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.794885 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1721e985-40bd-4b2d-a0e8-d4e365094ba6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a895e0cc04880c3287ba725ac5484835235f1e066a94ce98fcf4b6133c5cbd2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a21a97385f856ec996105b597d0ae81c56ed28b625ef3f2905e1af6ec8f6a45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://646e01a58908ba2e703aae5eef53e1b981b0f0c3d1d3202487f2a8b94ee3f55d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"ver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:40:26.813730 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814130 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:40:26.814139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:40:26.814143 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:40:26.814145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 03:40:26.821403 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821442 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821482 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821496 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 03:40:26.821509 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821514 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 03:40:26.821612 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 03:40:26.821619 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1206 03:40:26.822049 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16a92d501c9e626fe91e4a6793e426d10eb292cfbab766db8b8cc72e5dae94ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:34Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.807913 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-b2knv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a3ae46f-03b5-4999-9d58-c38ee222f68a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76665d4c127445009fd66078827b54607c9ef812c8011d1c0b4cde1c0516853a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk4hj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-b2knv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:34Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.824444 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22rrq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03112a5d-5882-4fac-9ae5-13525ff82fe0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0205b089b12c42688203bb7ec0f27dc5f9a577dabdb446b811f184046f85dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trlt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22rrq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:34Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.835819 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2466610962332c6ca120fe055e4e8a3fc623088119dc1a412e06b745176f89e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:34Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.848537 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:34Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.860042 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6df38316-e0d3-4018-8d27-3620eba3a68d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57381e9589a2dc04a4d2733c37985f42594182128a374b7be0f471ea69138b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bf5e41701c7a2d45adc4740ceb4ac1785db17bc7cb15daed439f7556812d1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpxxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:34Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.869617 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kj7hf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd46dfa5-7d4a-49ed-8199-523f10cebdf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5eee8a0cf96ceb3785dac9df31bf8d31715d1c88fbea17c5e582b0a4554bb63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmkcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kj7hf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:34Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.878329 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.878359 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.878367 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.878380 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.878389 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:34Z","lastTransitionTime":"2025-12-06T03:40:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.881013 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2093b6-c4b5-4db8-be16-fbc859885e8d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0d6e582161f0d9126126e03b27211be22f6b5f797154928c92fa13a8118f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03004a29aefe7fe2452fb9f6663c0807afccb055d7ddbf6b907f305e6a19c81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1801ac64db4ba91c2e8a249e3651d3e50ab62e9f2c87312ce9bf7f38518a82cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6e454507cadbb982a566a3692995aaa3c733ff1cd7fb4870476602190e1eeba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:34Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.893297 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:34Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.906319 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fad9b63d84f41fccf696cd313d7ca7f72b3d3251d6c1b7341573a3fadf1e378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:34Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.921625 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb3f278fff47c900c56d8a563d5ed75a2ae84e2034b0f45aad4083790b9e7946\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b10ca16ac26904ec5e6245344b71d836888cae9b069cae6a7a4831d609752db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:34Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.938319 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fad9b63d84f41fccf696cd313d7ca7f72b3d3251d6c1b7341573a3fadf1e378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:34Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.951170 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb3f278fff47c900c56d8a563d5ed75a2ae84e2034b0f45aad4083790b9e7946\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b10ca16ac26904ec5e6245344b71d836888cae9b069cae6a7a4831d609752db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:34Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.963283 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2093b6-c4b5-4db8-be16-fbc859885e8d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0d6e582161f0d9126126e03b27211be22f6b5f797154928c92fa13a8118f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03004a29aefe7fe2452fb9f6663c0807afccb055d7ddbf6b907f305e6a19c81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1801ac64db4ba91c2e8a249e3651d3e50ab62e9f2c87312ce9bf7f38518a82cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6e454507cadbb982a566a3692995aaa3c733ff1cd7fb4870476602190e1eeba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:34Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.980570 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.980608 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.980619 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.980635 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.980644 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:34Z","lastTransitionTime":"2025-12-06T03:40:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.980696 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:34Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:34 crc kubenswrapper[4802]: I1206 03:40:34.994461 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:34Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:35 crc kubenswrapper[4802]: I1206 03:40:35.010279 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtw2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3a3bac2-3923-4507-a4de-1556b3ad473e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fed5ee4189d54456a3216f4fdb1ecf003fa47e29dbfe4b2b2f301edfb897d6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtw2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:35Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:35 crc kubenswrapper[4802]: I1206 03:40:35.038255 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7730de4f033561ad77046e028cd1a3f0bfe2775387cbdfa35f47ea79909ec4cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a158aebd701ef46225d3e37109ae28ab58d107087bf06650c4b6d0055a3bbba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2357f37497fde64937eca7b38693893af745ab278653ebc79406e122a6d4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8941bd25107333b264d3f081a811fdfeea23ad92193fc61367665fd2cd1209c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f9cd4dc75e63f09f60b48b6da284643910bb32ba51f2a808cb4d9803f30042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6796f65ee9c5bd0c18f205697135d17677c5daf3ba378cc992cba795e0188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e40ff23f812195841d652697dded43c59c78b48a80b56d600da57773d73c0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3cb210bfc97a439a7ff98c02c59f8ad16bc400dc214a616a44b43a8b01e2782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pvm6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:35Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:35 crc kubenswrapper[4802]: I1206 03:40:35.050200 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-b2knv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a3ae46f-03b5-4999-9d58-c38ee222f68a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76665d4c127445009fd66078827b54607c9ef812c8011d1c0b4cde1c0516853a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk4hj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-b2knv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:35Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:35 crc kubenswrapper[4802]: I1206 03:40:35.069229 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22rrq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03112a5d-5882-4fac-9ae5-13525ff82fe0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0205b089b12c42688203bb7ec0f27dc5f9a577dabdb446b811f184046f85dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trlt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22rrq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:35Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:35 crc kubenswrapper[4802]: I1206 03:40:35.083158 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:35 crc kubenswrapper[4802]: I1206 03:40:35.083214 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:35 crc kubenswrapper[4802]: I1206 03:40:35.083233 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:35 crc kubenswrapper[4802]: I1206 03:40:35.083258 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:35 crc kubenswrapper[4802]: I1206 03:40:35.083280 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:35Z","lastTransitionTime":"2025-12-06T03:40:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:35 crc kubenswrapper[4802]: I1206 03:40:35.097074 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1721e985-40bd-4b2d-a0e8-d4e365094ba6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a895e0cc04880c3287ba725ac5484835235f1e066a94ce98fcf4b6133c5cbd2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a21a97385f856ec996105b597d0ae81c56ed28b625ef3f2905e1af6ec8f6a45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://646e01a58908ba2e703aae5eef53e1b981b0f0c3d1d3202487f2a8b94ee3f55d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"ver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:40:26.813730 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814130 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:40:26.814139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:40:26.814143 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:40:26.814145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 03:40:26.821403 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821442 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821482 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821496 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 03:40:26.821509 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821514 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 03:40:26.821612 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 03:40:26.821619 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1206 03:40:26.822049 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16a92d501c9e626fe91e4a6793e426d10eb292cfbab766db8b8cc72e5dae94ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:35Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:35 crc kubenswrapper[4802]: I1206 03:40:35.124502 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kj7hf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd46dfa5-7d4a-49ed-8199-523f10cebdf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5eee8a0cf96ceb3785dac9df31bf8d31715d1c88fbea17c5e582b0a4554bb63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmkcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kj7hf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:35Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:35 crc kubenswrapper[4802]: I1206 03:40:35.145817 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2466610962332c6ca120fe055e4e8a3fc623088119dc1a412e06b745176f89e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:35Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:35 crc kubenswrapper[4802]: I1206 03:40:35.170243 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:35Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:35 crc kubenswrapper[4802]: I1206 03:40:35.182284 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6df38316-e0d3-4018-8d27-3620eba3a68d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57381e9589a2dc04a4d2733c37985f42594182128a374b7be0f471ea69138b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bf5e41701c7a2d45adc4740ceb4ac1785db17bc7cb15daed439f7556812d1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpxxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:35Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:35 crc kubenswrapper[4802]: I1206 03:40:35.185841 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:35 crc kubenswrapper[4802]: I1206 03:40:35.185875 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:35 crc kubenswrapper[4802]: I1206 03:40:35.185907 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:35 crc kubenswrapper[4802]: I1206 03:40:35.185924 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:35 crc kubenswrapper[4802]: I1206 03:40:35.185935 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:35Z","lastTransitionTime":"2025-12-06T03:40:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:35 crc kubenswrapper[4802]: I1206 03:40:35.288968 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:35 crc kubenswrapper[4802]: I1206 03:40:35.289018 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:35 crc kubenswrapper[4802]: I1206 03:40:35.289031 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:35 crc kubenswrapper[4802]: I1206 03:40:35.289050 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:35 crc kubenswrapper[4802]: I1206 03:40:35.289063 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:35Z","lastTransitionTime":"2025-12-06T03:40:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:35 crc kubenswrapper[4802]: I1206 03:40:35.392400 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:35 crc kubenswrapper[4802]: I1206 03:40:35.392469 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:35 crc kubenswrapper[4802]: I1206 03:40:35.392492 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:35 crc kubenswrapper[4802]: I1206 03:40:35.392518 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:35 crc kubenswrapper[4802]: I1206 03:40:35.392535 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:35Z","lastTransitionTime":"2025-12-06T03:40:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:35 crc kubenswrapper[4802]: I1206 03:40:35.495630 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:35 crc kubenswrapper[4802]: I1206 03:40:35.495704 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:35 crc kubenswrapper[4802]: I1206 03:40:35.495728 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:35 crc kubenswrapper[4802]: I1206 03:40:35.495799 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:35 crc kubenswrapper[4802]: I1206 03:40:35.495826 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:35Z","lastTransitionTime":"2025-12-06T03:40:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:35 crc kubenswrapper[4802]: I1206 03:40:35.598777 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:35 crc kubenswrapper[4802]: I1206 03:40:35.598843 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:35 crc kubenswrapper[4802]: I1206 03:40:35.598866 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:35 crc kubenswrapper[4802]: I1206 03:40:35.598893 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:35 crc kubenswrapper[4802]: I1206 03:40:35.598912 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:35Z","lastTransitionTime":"2025-12-06T03:40:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:35 crc kubenswrapper[4802]: I1206 03:40:35.702086 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:35 crc kubenswrapper[4802]: I1206 03:40:35.702144 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:35 crc kubenswrapper[4802]: I1206 03:40:35.702155 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:35 crc kubenswrapper[4802]: I1206 03:40:35.702176 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:35 crc kubenswrapper[4802]: I1206 03:40:35.702188 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:35Z","lastTransitionTime":"2025-12-06T03:40:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:35 crc kubenswrapper[4802]: I1206 03:40:35.805419 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:35 crc kubenswrapper[4802]: I1206 03:40:35.805474 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:35 crc kubenswrapper[4802]: I1206 03:40:35.805490 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:35 crc kubenswrapper[4802]: I1206 03:40:35.805511 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:35 crc kubenswrapper[4802]: I1206 03:40:35.805528 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:35Z","lastTransitionTime":"2025-12-06T03:40:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:35 crc kubenswrapper[4802]: I1206 03:40:35.907880 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:35 crc kubenswrapper[4802]: I1206 03:40:35.907916 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:35 crc kubenswrapper[4802]: I1206 03:40:35.907925 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:35 crc kubenswrapper[4802]: I1206 03:40:35.907939 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:35 crc kubenswrapper[4802]: I1206 03:40:35.907947 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:35Z","lastTransitionTime":"2025-12-06T03:40:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:36 crc kubenswrapper[4802]: I1206 03:40:36.010596 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:36 crc kubenswrapper[4802]: I1206 03:40:36.010671 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:36 crc kubenswrapper[4802]: I1206 03:40:36.010681 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:36 crc kubenswrapper[4802]: I1206 03:40:36.010697 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:36 crc kubenswrapper[4802]: I1206 03:40:36.010706 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:36Z","lastTransitionTime":"2025-12-06T03:40:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:36 crc kubenswrapper[4802]: I1206 03:40:36.113232 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:36 crc kubenswrapper[4802]: I1206 03:40:36.113270 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:36 crc kubenswrapper[4802]: I1206 03:40:36.113285 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:36 crc kubenswrapper[4802]: I1206 03:40:36.113300 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:36 crc kubenswrapper[4802]: I1206 03:40:36.113308 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:36Z","lastTransitionTime":"2025-12-06T03:40:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:36 crc kubenswrapper[4802]: I1206 03:40:36.215512 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:36 crc kubenswrapper[4802]: I1206 03:40:36.215918 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:36 crc kubenswrapper[4802]: I1206 03:40:36.216078 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:36 crc kubenswrapper[4802]: I1206 03:40:36.216221 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:36 crc kubenswrapper[4802]: I1206 03:40:36.216369 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:36Z","lastTransitionTime":"2025-12-06T03:40:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:36 crc kubenswrapper[4802]: I1206 03:40:36.319049 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:36 crc kubenswrapper[4802]: I1206 03:40:36.319087 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:36 crc kubenswrapper[4802]: I1206 03:40:36.319098 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:36 crc kubenswrapper[4802]: I1206 03:40:36.319115 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:36 crc kubenswrapper[4802]: I1206 03:40:36.319136 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:36Z","lastTransitionTime":"2025-12-06T03:40:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:36 crc kubenswrapper[4802]: I1206 03:40:36.421456 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:36 crc kubenswrapper[4802]: I1206 03:40:36.421507 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:36 crc kubenswrapper[4802]: I1206 03:40:36.421523 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:36 crc kubenswrapper[4802]: I1206 03:40:36.421548 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:36 crc kubenswrapper[4802]: I1206 03:40:36.421564 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:36Z","lastTransitionTime":"2025-12-06T03:40:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:36 crc kubenswrapper[4802]: I1206 03:40:36.449858 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:40:36 crc kubenswrapper[4802]: I1206 03:40:36.449890 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:40:36 crc kubenswrapper[4802]: E1206 03:40:36.449970 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:40:36 crc kubenswrapper[4802]: I1206 03:40:36.449903 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:40:36 crc kubenswrapper[4802]: E1206 03:40:36.450068 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:40:36 crc kubenswrapper[4802]: E1206 03:40:36.450098 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:40:36 crc kubenswrapper[4802]: I1206 03:40:36.524772 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:36 crc kubenswrapper[4802]: I1206 03:40:36.524818 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:36 crc kubenswrapper[4802]: I1206 03:40:36.524830 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:36 crc kubenswrapper[4802]: I1206 03:40:36.524846 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:36 crc kubenswrapper[4802]: I1206 03:40:36.524858 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:36Z","lastTransitionTime":"2025-12-06T03:40:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:36 crc kubenswrapper[4802]: I1206 03:40:36.628854 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:36 crc kubenswrapper[4802]: I1206 03:40:36.628918 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:36 crc kubenswrapper[4802]: I1206 03:40:36.628937 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:36 crc kubenswrapper[4802]: I1206 03:40:36.628964 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:36 crc kubenswrapper[4802]: I1206 03:40:36.628982 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:36Z","lastTransitionTime":"2025-12-06T03:40:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:36 crc kubenswrapper[4802]: I1206 03:40:36.731315 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:36 crc kubenswrapper[4802]: I1206 03:40:36.731345 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:36 crc kubenswrapper[4802]: I1206 03:40:36.731353 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:36 crc kubenswrapper[4802]: I1206 03:40:36.731366 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:36 crc kubenswrapper[4802]: I1206 03:40:36.731376 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:36Z","lastTransitionTime":"2025-12-06T03:40:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:36 crc kubenswrapper[4802]: I1206 03:40:36.834302 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:36 crc kubenswrapper[4802]: I1206 03:40:36.834388 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:36 crc kubenswrapper[4802]: I1206 03:40:36.834410 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:36 crc kubenswrapper[4802]: I1206 03:40:36.834441 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:36 crc kubenswrapper[4802]: I1206 03:40:36.834465 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:36Z","lastTransitionTime":"2025-12-06T03:40:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:36 crc kubenswrapper[4802]: I1206 03:40:36.937132 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:36 crc kubenswrapper[4802]: I1206 03:40:36.937174 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:36 crc kubenswrapper[4802]: I1206 03:40:36.937185 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:36 crc kubenswrapper[4802]: I1206 03:40:36.937201 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:36 crc kubenswrapper[4802]: I1206 03:40:36.937210 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:36Z","lastTransitionTime":"2025-12-06T03:40:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.039193 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.039229 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.039240 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.039256 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.039270 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:37Z","lastTransitionTime":"2025-12-06T03:40:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.141684 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.141725 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.141736 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.141791 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.141807 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:37Z","lastTransitionTime":"2025-12-06T03:40:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.161867 4802 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.162967 4802 scope.go:117] "RemoveContainer" containerID="2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.244619 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.244974 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.245083 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.245231 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.245365 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:37Z","lastTransitionTime":"2025-12-06T03:40:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.347955 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.347986 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.347994 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.348007 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.348016 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:37Z","lastTransitionTime":"2025-12-06T03:40:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.450929 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.450990 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.451001 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.451016 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.451061 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:37Z","lastTransitionTime":"2025-12-06T03:40:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.465238 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.484609 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtw2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3a3bac2-3923-4507-a4de-1556b3ad473e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fed5ee4189d54456a3216f4fdb1ecf003fa47e29dbfe4b2b2f301edfb897d6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtw2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.514016 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7730de4f033561ad77046e028cd1a3f0bfe2775387cbdfa35f47ea79909ec4cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a158aebd701ef46225d3e37109ae28ab58d107087bf06650c4b6d0055a3bbba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2357f37497fde64937eca7b38693893af745ab278653ebc79406e122a6d4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8941bd25107333b264d3f081a811fdfeea23ad92193fc61367665fd2cd1209c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f9cd4dc75e63f09f60b48b6da284643910bb32ba51f2a808cb4d9803f30042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6796f65ee9c5bd0c18f205697135d17677c5daf3ba378cc992cba795e0188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e40ff23f812195841d652697dded43c59c78b48a80b56d600da57773d73c0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3cb210bfc97a439a7ff98c02c59f8ad16bc400dc214a616a44b43a8b01e2782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pvm6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.539607 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1721e985-40bd-4b2d-a0e8-d4e365094ba6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a895e0cc04880c3287ba725ac5484835235f1e066a94ce98fcf4b6133c5cbd2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a21a97385f856ec996105b597d0ae81c56ed28b625ef3f2905e1af6ec8f6a45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://646e01a58908ba2e703aae5eef53e1b981b0f0c3d1d3202487f2a8b94ee3f55d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"ver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:40:26.813730 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814130 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:40:26.814139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:40:26.814143 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:40:26.814145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 03:40:26.821403 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821442 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821482 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821496 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 03:40:26.821509 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821514 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 03:40:26.821612 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 03:40:26.821619 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1206 03:40:26.822049 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16a92d501c9e626fe91e4a6793e426d10eb292cfbab766db8b8cc72e5dae94ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.553440 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.553469 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.553477 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.553490 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.553500 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:37Z","lastTransitionTime":"2025-12-06T03:40:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.557106 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-b2knv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a3ae46f-03b5-4999-9d58-c38ee222f68a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76665d4c127445009fd66078827b54607c9ef812c8011d1c0b4cde1c0516853a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk4hj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-b2knv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.572730 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22rrq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03112a5d-5882-4fac-9ae5-13525ff82fe0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0205b089b12c42688203bb7ec0f27dc5f9a577dabdb446b811f184046f85dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trlt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22rrq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.587435 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2466610962332c6ca120fe055e4e8a3fc623088119dc1a412e06b745176f89e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.601745 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.617310 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6df38316-e0d3-4018-8d27-3620eba3a68d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57381e9589a2dc04a4d2733c37985f42594182128a374b7be0f471ea69138b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bf5e41701c7a2d45adc4740ceb4ac1785db17bc7cb15daed439f7556812d1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpxxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.633487 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kj7hf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd46dfa5-7d4a-49ed-8199-523f10cebdf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5eee8a0cf96ceb3785dac9df31bf8d31715d1c88fbea17c5e582b0a4554bb63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmkcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kj7hf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.656089 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2093b6-c4b5-4db8-be16-fbc859885e8d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0d6e582161f0d9126126e03b27211be22f6b5f797154928c92fa13a8118f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03004a29aefe7fe2452fb9f6663c0807afccb055d7ddbf6b907f305e6a19c81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1801ac64db4ba91c2e8a249e3651d3e50ab62e9f2c87312ce9bf7f38518a82cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6e454507cadbb982a566a3692995aaa3c733ff1cd7fb4870476602190e1eeba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.656688 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.656732 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.656743 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.656777 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.656794 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:37Z","lastTransitionTime":"2025-12-06T03:40:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.673520 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.688619 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fad9b63d84f41fccf696cd313d7ca7f72b3d3251d6c1b7341573a3fadf1e378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.707322 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb3f278fff47c900c56d8a563d5ed75a2ae84e2034b0f45aad4083790b9e7946\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b10ca16ac26904ec5e6245344b71d836888cae9b069cae6a7a4831d609752db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.719584 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.722437 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"31752b84d86bda30792b29e4c6ed963753cc576c5de784395e8e6dcbebaced47"} Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.722936 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.726599 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pvm6q_eafce4e0-e7fb-4877-b0ab-3283829ba1ef/ovnkube-controller/0.log" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.731652 4802 generic.go:334] "Generic (PLEG): container finished" podID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerID="e5e40ff23f812195841d652697dded43c59c78b48a80b56d600da57773d73c0d" exitCode=1 Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.731695 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" event={"ID":"eafce4e0-e7fb-4877-b0ab-3283829ba1ef","Type":"ContainerDied","Data":"e5e40ff23f812195841d652697dded43c59c78b48a80b56d600da57773d73c0d"} Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.736978 4802 scope.go:117] "RemoveContainer" containerID="e5e40ff23f812195841d652697dded43c59c78b48a80b56d600da57773d73c0d" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.747739 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.758837 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.758900 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.758919 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.758948 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.758965 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:37Z","lastTransitionTime":"2025-12-06T03:40:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.770843 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtw2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3a3bac2-3923-4507-a4de-1556b3ad473e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fed5ee4189d54456a3216f4fdb1ecf003fa47e29dbfe4b2b2f301edfb897d6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtw2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.801987 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7730de4f033561ad77046e028cd1a3f0bfe2775387cbdfa35f47ea79909ec4cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a158aebd701ef46225d3e37109ae28ab58d107087bf06650c4b6d0055a3bbba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2357f37497fde64937eca7b38693893af745ab278653ebc79406e122a6d4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8941bd25107333b264d3f081a811fdfeea23ad92193fc61367665fd2cd1209c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f9cd4dc75e63f09f60b48b6da284643910bb32ba51f2a808cb4d9803f30042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6796f65ee9c5bd0c18f205697135d17677c5daf3ba378cc992cba795e0188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e40ff23f812195841d652697dded43c59c78b48a80b56d600da57773d73c0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3cb210bfc97a439a7ff98c02c59f8ad16bc400dc214a616a44b43a8b01e2782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pvm6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.823393 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1721e985-40bd-4b2d-a0e8-d4e365094ba6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a895e0cc04880c3287ba725ac5484835235f1e066a94ce98fcf4b6133c5cbd2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a21a97385f856ec996105b597d0ae81c56ed28b625ef3f2905e1af6ec8f6a45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://646e01a58908ba2e703aae5eef53e1b981b0f0c3d1d3202487f2a8b94ee3f55d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31752b84d86bda30792b29e4c6ed963753cc576c5de784395e8e6dcbebaced47\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"ver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:40:26.813730 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814130 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:40:26.814139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:40:26.814143 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:40:26.814145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 03:40:26.821403 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821442 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821482 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821496 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 03:40:26.821509 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821514 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 03:40:26.821612 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 03:40:26.821619 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1206 03:40:26.822049 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16a92d501c9e626fe91e4a6793e426d10eb292cfbab766db8b8cc72e5dae94ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.835487 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-b2knv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a3ae46f-03b5-4999-9d58-c38ee222f68a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76665d4c127445009fd66078827b54607c9ef812c8011d1c0b4cde1c0516853a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk4hj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-b2knv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.854569 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22rrq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03112a5d-5882-4fac-9ae5-13525ff82fe0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0205b089b12c42688203bb7ec0f27dc5f9a577dabdb446b811f184046f85dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trlt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22rrq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.861888 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.862044 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.862361 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.862493 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.862606 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:37Z","lastTransitionTime":"2025-12-06T03:40:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.865785 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2466610962332c6ca120fe055e4e8a3fc623088119dc1a412e06b745176f89e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.877076 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.887919 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6df38316-e0d3-4018-8d27-3620eba3a68d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57381e9589a2dc04a4d2733c37985f42594182128a374b7be0f471ea69138b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bf5e41701c7a2d45adc4740ceb4ac1785db17bc7cb15daed439f7556812d1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpxxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.896619 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kj7hf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd46dfa5-7d4a-49ed-8199-523f10cebdf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5eee8a0cf96ceb3785dac9df31bf8d31715d1c88fbea17c5e582b0a4554bb63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmkcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kj7hf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.909097 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2093b6-c4b5-4db8-be16-fbc859885e8d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0d6e582161f0d9126126e03b27211be22f6b5f797154928c92fa13a8118f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03004a29aefe7fe2452fb9f6663c0807afccb055d7ddbf6b907f305e6a19c81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1801ac64db4ba91c2e8a249e3651d3e50ab62e9f2c87312ce9bf7f38518a82cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6e454507cadbb982a566a3692995aaa3c733ff1cd7fb4870476602190e1eeba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.932431 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.952054 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fad9b63d84f41fccf696cd313d7ca7f72b3d3251d6c1b7341573a3fadf1e378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.965252 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.965285 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.965295 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.965319 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.965334 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:37Z","lastTransitionTime":"2025-12-06T03:40:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:37 crc kubenswrapper[4802]: I1206 03:40:37.970006 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb3f278fff47c900c56d8a563d5ed75a2ae84e2034b0f45aad4083790b9e7946\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b10ca16ac26904ec5e6245344b71d836888cae9b069cae6a7a4831d609752db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.001075 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7730de4f033561ad77046e028cd1a3f0bfe2775387cbdfa35f47ea79909ec4cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a158aebd701ef46225d3e37109ae28ab58d107087bf06650c4b6d0055a3bbba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2357f37497fde64937eca7b38693893af745ab278653ebc79406e122a6d4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8941bd25107333b264d3f081a811fdfeea23ad92193fc61367665fd2cd1209c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f9cd4dc75e63f09f60b48b6da284643910bb32ba51f2a808cb4d9803f30042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6796f65ee9c5bd0c18f205697135d17677c5daf3ba378cc992cba795e0188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e40ff23f812195841d652697dded43c59c78b48a80b56d600da57773d73c0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e40ff23f812195841d652697dded43c59c78b48a80b56d600da57773d73c0d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:40:36Z\\\",\\\"message\\\":\\\"lient-go/informers/factory.go:160\\\\nI1206 03:40:36.495935 6089 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 03:40:36.496606 6089 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 03:40:36.496642 6089 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1206 03:40:36.496659 6089 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 03:40:36.496718 6089 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 03:40:36.496805 6089 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 03:40:36.496816 6089 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 03:40:36.496845 6089 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 03:40:36.497027 6089 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 03:40:36.497255 6089 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 03:40:36.497821 6089 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1206 03:40:36.497856 6089 factory.go:656] Stopping watch factory\\\\nI1206 03:40:36.497868 6089 ovnkube.go:599] Stopped ovnkube\\\\nI1206 03:40:3\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3cb210bfc97a439a7ff98c02c59f8ad16bc400dc214a616a44b43a8b01e2782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pvm6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.016935 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:38Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.036256 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtw2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3a3bac2-3923-4507-a4de-1556b3ad473e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fed5ee4189d54456a3216f4fdb1ecf003fa47e29dbfe4b2b2f301edfb897d6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtw2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:38Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.051466 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-b2knv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a3ae46f-03b5-4999-9d58-c38ee222f68a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76665d4c127445009fd66078827b54607c9ef812c8011d1c0b4cde1c0516853a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk4hj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-b2knv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:38Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.067338 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.067392 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.067410 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.067435 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.067453 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:38Z","lastTransitionTime":"2025-12-06T03:40:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.069959 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22rrq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03112a5d-5882-4fac-9ae5-13525ff82fe0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0205b089b12c42688203bb7ec0f27dc5f9a577dabdb446b811f184046f85dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trlt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22rrq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:38Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.090192 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1721e985-40bd-4b2d-a0e8-d4e365094ba6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a895e0cc04880c3287ba725ac5484835235f1e066a94ce98fcf4b6133c5cbd2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a21a97385f856ec996105b597d0ae81c56ed28b625ef3f2905e1af6ec8f6a45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://646e01a58908ba2e703aae5eef53e1b981b0f0c3d1d3202487f2a8b94ee3f55d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31752b84d86bda30792b29e4c6ed963753cc576c5de784395e8e6dcbebaced47\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"ver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:40:26.813730 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814130 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:40:26.814139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:40:26.814143 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:40:26.814145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 03:40:26.821403 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821442 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821482 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821496 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 03:40:26.821509 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821514 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 03:40:26.821612 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 03:40:26.821619 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1206 03:40:26.822049 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16a92d501c9e626fe91e4a6793e426d10eb292cfbab766db8b8cc72e5dae94ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:38Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.103510 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6df38316-e0d3-4018-8d27-3620eba3a68d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57381e9589a2dc04a4d2733c37985f42594182128a374b7be0f471ea69138b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bf5e41701c7a2d45adc4740ceb4ac1785db17bc7cb15daed439f7556812d1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpxxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:38Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.113392 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kj7hf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd46dfa5-7d4a-49ed-8199-523f10cebdf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5eee8a0cf96ceb3785dac9df31bf8d31715d1c88fbea17c5e582b0a4554bb63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmkcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kj7hf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:38Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.127748 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2466610962332c6ca120fe055e4e8a3fc623088119dc1a412e06b745176f89e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:38Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.144021 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:38Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.170243 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.170286 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.170295 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.170311 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.170320 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:38Z","lastTransitionTime":"2025-12-06T03:40:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.173258 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:38Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.190284 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fad9b63d84f41fccf696cd313d7ca7f72b3d3251d6c1b7341573a3fadf1e378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:38Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.204053 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb3f278fff47c900c56d8a563d5ed75a2ae84e2034b0f45aad4083790b9e7946\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b10ca16ac26904ec5e6245344b71d836888cae9b069cae6a7a4831d609752db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:38Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.223462 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2093b6-c4b5-4db8-be16-fbc859885e8d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0d6e582161f0d9126126e03b27211be22f6b5f797154928c92fa13a8118f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03004a29aefe7fe2452fb9f6663c0807afccb055d7ddbf6b907f305e6a19c81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1801ac64db4ba91c2e8a249e3651d3e50ab62e9f2c87312ce9bf7f38518a82cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6e454507cadbb982a566a3692995aaa3c733ff1cd7fb4870476602190e1eeba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:38Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.272295 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.272323 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.272332 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.272347 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.272357 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:38Z","lastTransitionTime":"2025-12-06T03:40:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.375141 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.375185 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.375195 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.375213 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.375224 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:38Z","lastTransitionTime":"2025-12-06T03:40:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.449598 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.449664 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.449691 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:40:38 crc kubenswrapper[4802]: E1206 03:40:38.449721 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:40:38 crc kubenswrapper[4802]: E1206 03:40:38.449835 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:40:38 crc kubenswrapper[4802]: E1206 03:40:38.449925 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.476937 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.476980 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.476995 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.477013 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.477034 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:38Z","lastTransitionTime":"2025-12-06T03:40:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.584169 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.584497 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.584633 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.584815 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.584990 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:38Z","lastTransitionTime":"2025-12-06T03:40:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.687915 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.687960 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.687974 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.687992 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.688005 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:38Z","lastTransitionTime":"2025-12-06T03:40:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.737229 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pvm6q_eafce4e0-e7fb-4877-b0ab-3283829ba1ef/ovnkube-controller/0.log" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.740095 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" event={"ID":"eafce4e0-e7fb-4877-b0ab-3283829ba1ef","Type":"ContainerStarted","Data":"cf845e8893e51078e46991489d9a27c1194eab62a9d0a24eb3336d396e12cdfb"} Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.740528 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.755622 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2093b6-c4b5-4db8-be16-fbc859885e8d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0d6e582161f0d9126126e03b27211be22f6b5f797154928c92fa13a8118f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03004a29aefe7fe2452fb9f6663c0807afccb055d7ddbf6b907f305e6a19c81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1801ac64db4ba91c2e8a249e3651d3e50ab62e9f2c87312ce9bf7f38518a82cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6e454507cadbb982a566a3692995aaa3c733ff1cd7fb4870476602190e1eeba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:38Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.769667 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:38Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.781606 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fad9b63d84f41fccf696cd313d7ca7f72b3d3251d6c1b7341573a3fadf1e378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:38Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.790247 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.790302 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.790314 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.790327 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.790336 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:38Z","lastTransitionTime":"2025-12-06T03:40:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.794123 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb3f278fff47c900c56d8a563d5ed75a2ae84e2034b0f45aad4083790b9e7946\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b10ca16ac26904ec5e6245344b71d836888cae9b069cae6a7a4831d609752db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:38Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.807106 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:38Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.822577 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtw2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3a3bac2-3923-4507-a4de-1556b3ad473e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fed5ee4189d54456a3216f4fdb1ecf003fa47e29dbfe4b2b2f301edfb897d6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtw2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:38Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.849416 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7730de4f033561ad77046e028cd1a3f0bfe2775387cbdfa35f47ea79909ec4cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a158aebd701ef46225d3e37109ae28ab58d107087bf06650c4b6d0055a3bbba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2357f37497fde64937eca7b38693893af745ab278653ebc79406e122a6d4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8941bd25107333b264d3f081a811fdfeea23ad92193fc61367665fd2cd1209c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f9cd4dc75e63f09f60b48b6da284643910bb32ba51f2a808cb4d9803f30042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6796f65ee9c5bd0c18f205697135d17677c5daf3ba378cc992cba795e0188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf845e8893e51078e46991489d9a27c1194eab62a9d0a24eb3336d396e12cdfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e40ff23f812195841d652697dded43c59c78b48a80b56d600da57773d73c0d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:40:36Z\\\",\\\"message\\\":\\\"lient-go/informers/factory.go:160\\\\nI1206 03:40:36.495935 6089 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 03:40:36.496606 6089 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 03:40:36.496642 6089 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1206 03:40:36.496659 6089 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 03:40:36.496718 6089 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 03:40:36.496805 6089 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 03:40:36.496816 6089 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 03:40:36.496845 6089 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 03:40:36.497027 6089 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 03:40:36.497255 6089 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 03:40:36.497821 6089 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1206 03:40:36.497856 6089 factory.go:656] Stopping watch factory\\\\nI1206 03:40:36.497868 6089 ovnkube.go:599] Stopped ovnkube\\\\nI1206 03:40:3\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3cb210bfc97a439a7ff98c02c59f8ad16bc400dc214a616a44b43a8b01e2782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pvm6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:38Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.872478 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1721e985-40bd-4b2d-a0e8-d4e365094ba6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a895e0cc04880c3287ba725ac5484835235f1e066a94ce98fcf4b6133c5cbd2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a21a97385f856ec996105b597d0ae81c56ed28b625ef3f2905e1af6ec8f6a45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://646e01a58908ba2e703aae5eef53e1b981b0f0c3d1d3202487f2a8b94ee3f55d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31752b84d86bda30792b29e4c6ed963753cc576c5de784395e8e6dcbebaced47\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"ver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:40:26.813730 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814130 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:40:26.814139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:40:26.814143 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:40:26.814145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 03:40:26.821403 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821442 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821482 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821496 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 03:40:26.821509 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821514 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 03:40:26.821612 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 03:40:26.821619 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1206 03:40:26.822049 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16a92d501c9e626fe91e4a6793e426d10eb292cfbab766db8b8cc72e5dae94ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:38Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.887632 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-b2knv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a3ae46f-03b5-4999-9d58-c38ee222f68a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76665d4c127445009fd66078827b54607c9ef812c8011d1c0b4cde1c0516853a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk4hj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-b2knv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:38Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.892688 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.892723 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.892732 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.892767 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.892776 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:38Z","lastTransitionTime":"2025-12-06T03:40:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.902199 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22rrq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03112a5d-5882-4fac-9ae5-13525ff82fe0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0205b089b12c42688203bb7ec0f27dc5f9a577dabdb446b811f184046f85dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trlt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22rrq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:38Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.920507 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2466610962332c6ca120fe055e4e8a3fc623088119dc1a412e06b745176f89e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:38Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.938544 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:38Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.953681 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6df38316-e0d3-4018-8d27-3620eba3a68d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57381e9589a2dc04a4d2733c37985f42594182128a374b7be0f471ea69138b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bf5e41701c7a2d45adc4740ceb4ac1785db17bc7cb15daed439f7556812d1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpxxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:38Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.967127 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kj7hf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd46dfa5-7d4a-49ed-8199-523f10cebdf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5eee8a0cf96ceb3785dac9df31bf8d31715d1c88fbea17c5e582b0a4554bb63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmkcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kj7hf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:38Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.995747 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.996099 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.996246 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.996415 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:38 crc kubenswrapper[4802]: I1206 03:40:38.996546 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:38Z","lastTransitionTime":"2025-12-06T03:40:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.056189 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.056477 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.056599 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.056736 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.056933 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:39Z","lastTransitionTime":"2025-12-06T03:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:39 crc kubenswrapper[4802]: E1206 03:40:39.076491 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0e01ca3a-758f-44bc-b4a2-30c1bc04fda0\\\",\\\"systemUUID\\\":\\\"c7c815e4-d2a4-4244-b035-f988fc95e215\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.083181 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.083228 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.083240 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.083258 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.083270 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:39Z","lastTransitionTime":"2025-12-06T03:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:39 crc kubenswrapper[4802]: E1206 03:40:39.102100 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0e01ca3a-758f-44bc-b4a2-30c1bc04fda0\\\",\\\"systemUUID\\\":\\\"c7c815e4-d2a4-4244-b035-f988fc95e215\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.106887 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.106950 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.106968 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.106994 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.107011 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:39Z","lastTransitionTime":"2025-12-06T03:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:39 crc kubenswrapper[4802]: E1206 03:40:39.128033 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0e01ca3a-758f-44bc-b4a2-30c1bc04fda0\\\",\\\"systemUUID\\\":\\\"c7c815e4-d2a4-4244-b035-f988fc95e215\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.133081 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.133118 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.133129 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.133147 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.133161 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:39Z","lastTransitionTime":"2025-12-06T03:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.147516 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ngspk"] Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.148308 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ngspk" Dec 06 03:40:39 crc kubenswrapper[4802]: E1206 03:40:39.149941 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0e01ca3a-758f-44bc-b4a2-30c1bc04fda0\\\",\\\"systemUUID\\\":\\\"c7c815e4-d2a4-4244-b035-f988fc95e215\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.150466 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.150667 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.157449 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.157499 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.157513 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.157532 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.157544 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:39Z","lastTransitionTime":"2025-12-06T03:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.167015 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2093b6-c4b5-4db8-be16-fbc859885e8d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0d6e582161f0d9126126e03b27211be22f6b5f797154928c92fa13a8118f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03004a29aefe7fe2452fb9f6663c0807afccb055d7ddbf6b907f305e6a19c81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1801ac64db4ba91c2e8a249e3651d3e50ab62e9f2c87312ce9bf7f38518a82cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6e454507cadbb982a566a3692995aaa3c733ff1cd7fb4870476602190e1eeba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:39 crc kubenswrapper[4802]: E1206 03:40:39.173090 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0e01ca3a-758f-44bc-b4a2-30c1bc04fda0\\\",\\\"systemUUID\\\":\\\"c7c815e4-d2a4-4244-b035-f988fc95e215\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:39 crc kubenswrapper[4802]: E1206 03:40:39.173440 4802 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.175463 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.175490 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.175499 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.175512 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.175520 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:39Z","lastTransitionTime":"2025-12-06T03:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.183068 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.198032 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fad9b63d84f41fccf696cd313d7ca7f72b3d3251d6c1b7341573a3fadf1e378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.205908 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/34808ab2-d21b-40b6-8fe3-e18bf9e5fc46-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-ngspk\" (UID: \"34808ab2-d21b-40b6-8fe3-e18bf9e5fc46\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ngspk" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.206306 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/34808ab2-d21b-40b6-8fe3-e18bf9e5fc46-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-ngspk\" (UID: \"34808ab2-d21b-40b6-8fe3-e18bf9e5fc46\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ngspk" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.206406 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/34808ab2-d21b-40b6-8fe3-e18bf9e5fc46-env-overrides\") pod \"ovnkube-control-plane-749d76644c-ngspk\" (UID: \"34808ab2-d21b-40b6-8fe3-e18bf9e5fc46\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ngspk" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.206491 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qkf8j\" (UniqueName: \"kubernetes.io/projected/34808ab2-d21b-40b6-8fe3-e18bf9e5fc46-kube-api-access-qkf8j\") pod \"ovnkube-control-plane-749d76644c-ngspk\" (UID: \"34808ab2-d21b-40b6-8fe3-e18bf9e5fc46\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ngspk" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.217483 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb3f278fff47c900c56d8a563d5ed75a2ae84e2034b0f45aad4083790b9e7946\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b10ca16ac26904ec5e6245344b71d836888cae9b069cae6a7a4831d609752db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.228906 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.240631 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtw2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3a3bac2-3923-4507-a4de-1556b3ad473e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fed5ee4189d54456a3216f4fdb1ecf003fa47e29dbfe4b2b2f301edfb897d6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtw2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.258915 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7730de4f033561ad77046e028cd1a3f0bfe2775387cbdfa35f47ea79909ec4cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a158aebd701ef46225d3e37109ae28ab58d107087bf06650c4b6d0055a3bbba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2357f37497fde64937eca7b38693893af745ab278653ebc79406e122a6d4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8941bd25107333b264d3f081a811fdfeea23ad92193fc61367665fd2cd1209c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f9cd4dc75e63f09f60b48b6da284643910bb32ba51f2a808cb4d9803f30042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6796f65ee9c5bd0c18f205697135d17677c5daf3ba378cc992cba795e0188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf845e8893e51078e46991489d9a27c1194eab62a9d0a24eb3336d396e12cdfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e40ff23f812195841d652697dded43c59c78b48a80b56d600da57773d73c0d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:40:36Z\\\",\\\"message\\\":\\\"lient-go/informers/factory.go:160\\\\nI1206 03:40:36.495935 6089 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 03:40:36.496606 6089 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 03:40:36.496642 6089 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1206 03:40:36.496659 6089 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 03:40:36.496718 6089 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 03:40:36.496805 6089 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 03:40:36.496816 6089 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 03:40:36.496845 6089 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 03:40:36.497027 6089 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 03:40:36.497255 6089 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 03:40:36.497821 6089 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1206 03:40:36.497856 6089 factory.go:656] Stopping watch factory\\\\nI1206 03:40:36.497868 6089 ovnkube.go:599] Stopped ovnkube\\\\nI1206 03:40:3\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3cb210bfc97a439a7ff98c02c59f8ad16bc400dc214a616a44b43a8b01e2782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pvm6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.271368 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1721e985-40bd-4b2d-a0e8-d4e365094ba6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a895e0cc04880c3287ba725ac5484835235f1e066a94ce98fcf4b6133c5cbd2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a21a97385f856ec996105b597d0ae81c56ed28b625ef3f2905e1af6ec8f6a45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://646e01a58908ba2e703aae5eef53e1b981b0f0c3d1d3202487f2a8b94ee3f55d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31752b84d86bda30792b29e4c6ed963753cc576c5de784395e8e6dcbebaced47\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"ver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:40:26.813730 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814130 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:40:26.814139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:40:26.814143 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:40:26.814145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 03:40:26.821403 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821442 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821482 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821496 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 03:40:26.821509 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821514 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 03:40:26.821612 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 03:40:26.821619 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1206 03:40:26.822049 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16a92d501c9e626fe91e4a6793e426d10eb292cfbab766db8b8cc72e5dae94ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.278204 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.278363 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.278503 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.278628 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.278720 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:39Z","lastTransitionTime":"2025-12-06T03:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.284299 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-b2knv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a3ae46f-03b5-4999-9d58-c38ee222f68a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76665d4c127445009fd66078827b54607c9ef812c8011d1c0b4cde1c0516853a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk4hj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-b2knv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.298096 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22rrq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03112a5d-5882-4fac-9ae5-13525ff82fe0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0205b089b12c42688203bb7ec0f27dc5f9a577dabdb446b811f184046f85dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trlt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22rrq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.307687 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/34808ab2-d21b-40b6-8fe3-e18bf9e5fc46-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-ngspk\" (UID: \"34808ab2-d21b-40b6-8fe3-e18bf9e5fc46\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ngspk" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.307782 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/34808ab2-d21b-40b6-8fe3-e18bf9e5fc46-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-ngspk\" (UID: \"34808ab2-d21b-40b6-8fe3-e18bf9e5fc46\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ngspk" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.307813 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qkf8j\" (UniqueName: \"kubernetes.io/projected/34808ab2-d21b-40b6-8fe3-e18bf9e5fc46-kube-api-access-qkf8j\") pod \"ovnkube-control-plane-749d76644c-ngspk\" (UID: \"34808ab2-d21b-40b6-8fe3-e18bf9e5fc46\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ngspk" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.307846 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/34808ab2-d21b-40b6-8fe3-e18bf9e5fc46-env-overrides\") pod \"ovnkube-control-plane-749d76644c-ngspk\" (UID: \"34808ab2-d21b-40b6-8fe3-e18bf9e5fc46\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ngspk" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.308451 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/34808ab2-d21b-40b6-8fe3-e18bf9e5fc46-env-overrides\") pod \"ovnkube-control-plane-749d76644c-ngspk\" (UID: \"34808ab2-d21b-40b6-8fe3-e18bf9e5fc46\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ngspk" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.308857 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/34808ab2-d21b-40b6-8fe3-e18bf9e5fc46-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-ngspk\" (UID: \"34808ab2-d21b-40b6-8fe3-e18bf9e5fc46\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ngspk" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.309803 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ngspk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34808ab2-d21b-40b6-8fe3-e18bf9e5fc46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qkf8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qkf8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ngspk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.314162 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/34808ab2-d21b-40b6-8fe3-e18bf9e5fc46-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-ngspk\" (UID: \"34808ab2-d21b-40b6-8fe3-e18bf9e5fc46\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ngspk" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.325566 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2466610962332c6ca120fe055e4e8a3fc623088119dc1a412e06b745176f89e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.325944 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qkf8j\" (UniqueName: \"kubernetes.io/projected/34808ab2-d21b-40b6-8fe3-e18bf9e5fc46-kube-api-access-qkf8j\") pod \"ovnkube-control-plane-749d76644c-ngspk\" (UID: \"34808ab2-d21b-40b6-8fe3-e18bf9e5fc46\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ngspk" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.340953 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.353190 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6df38316-e0d3-4018-8d27-3620eba3a68d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57381e9589a2dc04a4d2733c37985f42594182128a374b7be0f471ea69138b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bf5e41701c7a2d45adc4740ceb4ac1785db17bc7cb15daed439f7556812d1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpxxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.363556 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kj7hf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd46dfa5-7d4a-49ed-8199-523f10cebdf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5eee8a0cf96ceb3785dac9df31bf8d31715d1c88fbea17c5e582b0a4554bb63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmkcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kj7hf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.382098 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.382173 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.382225 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.382254 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.382273 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:39Z","lastTransitionTime":"2025-12-06T03:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.463989 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ngspk" Dec 06 03:40:39 crc kubenswrapper[4802]: W1206 03:40:39.479222 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod34808ab2_d21b_40b6_8fe3_e18bf9e5fc46.slice/crio-d477fad27fe11824b4d2c97d48f111b414ad715715e5a9248c17a307392536b9 WatchSource:0}: Error finding container d477fad27fe11824b4d2c97d48f111b414ad715715e5a9248c17a307392536b9: Status 404 returned error can't find the container with id d477fad27fe11824b4d2c97d48f111b414ad715715e5a9248c17a307392536b9 Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.484148 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.484170 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.484180 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.484194 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.484203 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:39Z","lastTransitionTime":"2025-12-06T03:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.587121 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.587189 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.587212 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.587237 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.587253 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:39Z","lastTransitionTime":"2025-12-06T03:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.690415 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.690482 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.690500 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.690527 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.690553 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:39Z","lastTransitionTime":"2025-12-06T03:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.745621 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ngspk" event={"ID":"34808ab2-d21b-40b6-8fe3-e18bf9e5fc46","Type":"ContainerStarted","Data":"d477fad27fe11824b4d2c97d48f111b414ad715715e5a9248c17a307392536b9"} Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.748666 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pvm6q_eafce4e0-e7fb-4877-b0ab-3283829ba1ef/ovnkube-controller/1.log" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.749726 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pvm6q_eafce4e0-e7fb-4877-b0ab-3283829ba1ef/ovnkube-controller/0.log" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.754000 4802 generic.go:334] "Generic (PLEG): container finished" podID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerID="cf845e8893e51078e46991489d9a27c1194eab62a9d0a24eb3336d396e12cdfb" exitCode=1 Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.754057 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" event={"ID":"eafce4e0-e7fb-4877-b0ab-3283829ba1ef","Type":"ContainerDied","Data":"cf845e8893e51078e46991489d9a27c1194eab62a9d0a24eb3336d396e12cdfb"} Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.754111 4802 scope.go:117] "RemoveContainer" containerID="e5e40ff23f812195841d652697dded43c59c78b48a80b56d600da57773d73c0d" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.755452 4802 scope.go:117] "RemoveContainer" containerID="cf845e8893e51078e46991489d9a27c1194eab62a9d0a24eb3336d396e12cdfb" Dec 06 03:40:39 crc kubenswrapper[4802]: E1206 03:40:39.755848 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-pvm6q_openshift-ovn-kubernetes(eafce4e0-e7fb-4877-b0ab-3283829ba1ef)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" podUID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.775356 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2093b6-c4b5-4db8-be16-fbc859885e8d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0d6e582161f0d9126126e03b27211be22f6b5f797154928c92fa13a8118f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03004a29aefe7fe2452fb9f6663c0807afccb055d7ddbf6b907f305e6a19c81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1801ac64db4ba91c2e8a249e3651d3e50ab62e9f2c87312ce9bf7f38518a82cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6e454507cadbb982a566a3692995aaa3c733ff1cd7fb4870476602190e1eeba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.793724 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.793846 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.793873 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.793909 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.793933 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:39Z","lastTransitionTime":"2025-12-06T03:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.796070 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.815823 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fad9b63d84f41fccf696cd313d7ca7f72b3d3251d6c1b7341573a3fadf1e378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.834159 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb3f278fff47c900c56d8a563d5ed75a2ae84e2034b0f45aad4083790b9e7946\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b10ca16ac26904ec5e6245344b71d836888cae9b069cae6a7a4831d609752db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.856334 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.881326 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtw2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3a3bac2-3923-4507-a4de-1556b3ad473e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fed5ee4189d54456a3216f4fdb1ecf003fa47e29dbfe4b2b2f301edfb897d6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtw2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.897331 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.897407 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.897427 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.897452 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.897469 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:39Z","lastTransitionTime":"2025-12-06T03:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.916116 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7730de4f033561ad77046e028cd1a3f0bfe2775387cbdfa35f47ea79909ec4cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a158aebd701ef46225d3e37109ae28ab58d107087bf06650c4b6d0055a3bbba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2357f37497fde64937eca7b38693893af745ab278653ebc79406e122a6d4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8941bd25107333b264d3f081a811fdfeea23ad92193fc61367665fd2cd1209c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f9cd4dc75e63f09f60b48b6da284643910bb32ba51f2a808cb4d9803f30042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6796f65ee9c5bd0c18f205697135d17677c5daf3ba378cc992cba795e0188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf845e8893e51078e46991489d9a27c1194eab62a9d0a24eb3336d396e12cdfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e40ff23f812195841d652697dded43c59c78b48a80b56d600da57773d73c0d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:40:36Z\\\",\\\"message\\\":\\\"lient-go/informers/factory.go:160\\\\nI1206 03:40:36.495935 6089 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 03:40:36.496606 6089 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 03:40:36.496642 6089 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1206 03:40:36.496659 6089 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 03:40:36.496718 6089 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 03:40:36.496805 6089 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 03:40:36.496816 6089 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 03:40:36.496845 6089 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 03:40:36.497027 6089 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 03:40:36.497255 6089 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 03:40:36.497821 6089 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1206 03:40:36.497856 6089 factory.go:656] Stopping watch factory\\\\nI1206 03:40:36.497868 6089 ovnkube.go:599] Stopped ovnkube\\\\nI1206 03:40:3\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf845e8893e51078e46991489d9a27c1194eab62a9d0a24eb3336d396e12cdfb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:40:38Z\\\",\\\"message\\\":\\\"Map:map[10.217.4.110:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {f9232b32-e89f-4c8e-acc4-c6801b70dcb0}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:NB_Global Row:map[] Rows:[] Columns:[] Mutations:[{Column:nb_cfg Mutator:+= Value:1}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {6011affd-30a6-4be6-872d-e4cf1ca780cf}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 03:40:38.703973 6236 ovnkube.go:599] Stopped ovnkube\\\\nI1206 03:40:38.704014 6236 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1206 03:40:38.703921 6236 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-controller]} name:Service_openshift-machine-config-operator/machine-config-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.16:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {3f1b9878-e751-4e46-a226-ce007d2c4aa7}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1206 03:40:38.704093 6236 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3cb210bfc97a439a7ff98c02c59f8ad16bc400dc214a616a44b43a8b01e2782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pvm6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.934703 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1721e985-40bd-4b2d-a0e8-d4e365094ba6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a895e0cc04880c3287ba725ac5484835235f1e066a94ce98fcf4b6133c5cbd2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a21a97385f856ec996105b597d0ae81c56ed28b625ef3f2905e1af6ec8f6a45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://646e01a58908ba2e703aae5eef53e1b981b0f0c3d1d3202487f2a8b94ee3f55d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31752b84d86bda30792b29e4c6ed963753cc576c5de784395e8e6dcbebaced47\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"ver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:40:26.813730 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814130 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:40:26.814139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:40:26.814143 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:40:26.814145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 03:40:26.821403 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821442 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821482 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821496 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 03:40:26.821509 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821514 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 03:40:26.821612 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 03:40:26.821619 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1206 03:40:26.822049 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16a92d501c9e626fe91e4a6793e426d10eb292cfbab766db8b8cc72e5dae94ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.948196 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-b2knv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a3ae46f-03b5-4999-9d58-c38ee222f68a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76665d4c127445009fd66078827b54607c9ef812c8011d1c0b4cde1c0516853a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk4hj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-b2knv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.964519 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22rrq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03112a5d-5882-4fac-9ae5-13525ff82fe0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0205b089b12c42688203bb7ec0f27dc5f9a577dabdb446b811f184046f85dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trlt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22rrq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:39 crc kubenswrapper[4802]: I1206 03:40:39.985487 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ngspk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34808ab2-d21b-40b6-8fe3-e18bf9e5fc46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qkf8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qkf8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ngspk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.000363 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.000410 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.000423 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.000441 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.000454 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:40Z","lastTransitionTime":"2025-12-06T03:40:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.001659 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2466610962332c6ca120fe055e4e8a3fc623088119dc1a412e06b745176f89e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.016313 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:40Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.027990 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6df38316-e0d3-4018-8d27-3620eba3a68d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57381e9589a2dc04a4d2733c37985f42594182128a374b7be0f471ea69138b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bf5e41701c7a2d45adc4740ceb4ac1785db17bc7cb15daed439f7556812d1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpxxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:40Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.038283 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kj7hf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd46dfa5-7d4a-49ed-8199-523f10cebdf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5eee8a0cf96ceb3785dac9df31bf8d31715d1c88fbea17c5e582b0a4554bb63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmkcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kj7hf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:40Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.104216 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.104269 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.104280 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.104296 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.104309 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:40Z","lastTransitionTime":"2025-12-06T03:40:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.207584 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.207659 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.207685 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.207715 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.207736 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:40Z","lastTransitionTime":"2025-12-06T03:40:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.311568 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.311795 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.311813 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.311828 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.311839 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:40Z","lastTransitionTime":"2025-12-06T03:40:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.415554 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.415625 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.415648 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.415677 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.415698 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:40Z","lastTransitionTime":"2025-12-06T03:40:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.449732 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.449845 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.449735 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:40:40 crc kubenswrapper[4802]: E1206 03:40:40.450006 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:40:40 crc kubenswrapper[4802]: E1206 03:40:40.450138 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:40:40 crc kubenswrapper[4802]: E1206 03:40:40.450319 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.518705 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.518825 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.518850 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.518873 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.518891 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:40Z","lastTransitionTime":"2025-12-06T03:40:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.623394 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.623473 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.623493 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.623519 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.623538 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:40Z","lastTransitionTime":"2025-12-06T03:40:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.726334 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.726376 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.726385 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.726400 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.726410 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:40Z","lastTransitionTime":"2025-12-06T03:40:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.759059 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ngspk" event={"ID":"34808ab2-d21b-40b6-8fe3-e18bf9e5fc46","Type":"ContainerStarted","Data":"52509d21f5b4bfbd458162cba176ff82c627c7c4671204addf9d76f7392fd760"} Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.761908 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pvm6q_eafce4e0-e7fb-4877-b0ab-3283829ba1ef/ovnkube-controller/1.log" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.768970 4802 scope.go:117] "RemoveContainer" containerID="cf845e8893e51078e46991489d9a27c1194eab62a9d0a24eb3336d396e12cdfb" Dec 06 03:40:40 crc kubenswrapper[4802]: E1206 03:40:40.769260 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-pvm6q_openshift-ovn-kubernetes(eafce4e0-e7fb-4877-b0ab-3283829ba1ef)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" podUID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.784308 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1721e985-40bd-4b2d-a0e8-d4e365094ba6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a895e0cc04880c3287ba725ac5484835235f1e066a94ce98fcf4b6133c5cbd2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a21a97385f856ec996105b597d0ae81c56ed28b625ef3f2905e1af6ec8f6a45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://646e01a58908ba2e703aae5eef53e1b981b0f0c3d1d3202487f2a8b94ee3f55d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31752b84d86bda30792b29e4c6ed963753cc576c5de784395e8e6dcbebaced47\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"ver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:40:26.813730 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814130 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:40:26.814139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:40:26.814143 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:40:26.814145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 03:40:26.821403 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821442 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821482 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821496 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 03:40:26.821509 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821514 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 03:40:26.821612 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 03:40:26.821619 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1206 03:40:26.822049 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16a92d501c9e626fe91e4a6793e426d10eb292cfbab766db8b8cc72e5dae94ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:40Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.799577 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-b2knv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a3ae46f-03b5-4999-9d58-c38ee222f68a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76665d4c127445009fd66078827b54607c9ef812c8011d1c0b4cde1c0516853a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk4hj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-b2knv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:40Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.813977 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22rrq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03112a5d-5882-4fac-9ae5-13525ff82fe0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0205b089b12c42688203bb7ec0f27dc5f9a577dabdb446b811f184046f85dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trlt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22rrq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:40Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.825734 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ngspk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34808ab2-d21b-40b6-8fe3-e18bf9e5fc46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qkf8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qkf8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ngspk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:40Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.828136 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.828219 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.828235 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.828258 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.828274 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:40Z","lastTransitionTime":"2025-12-06T03:40:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.840614 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:40Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.852478 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6df38316-e0d3-4018-8d27-3620eba3a68d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57381e9589a2dc04a4d2733c37985f42594182128a374b7be0f471ea69138b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bf5e41701c7a2d45adc4740ceb4ac1785db17bc7cb15daed439f7556812d1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpxxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:40Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.862872 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kj7hf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd46dfa5-7d4a-49ed-8199-523f10cebdf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5eee8a0cf96ceb3785dac9df31bf8d31715d1c88fbea17c5e582b0a4554bb63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmkcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kj7hf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:40Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.879507 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2466610962332c6ca120fe055e4e8a3fc623088119dc1a412e06b745176f89e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:40Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.892573 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2093b6-c4b5-4db8-be16-fbc859885e8d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0d6e582161f0d9126126e03b27211be22f6b5f797154928c92fa13a8118f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03004a29aefe7fe2452fb9f6663c0807afccb055d7ddbf6b907f305e6a19c81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1801ac64db4ba91c2e8a249e3651d3e50ab62e9f2c87312ce9bf7f38518a82cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6e454507cadbb982a566a3692995aaa3c733ff1cd7fb4870476602190e1eeba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:40Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.909084 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:40Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.931187 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.931222 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.931231 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.931247 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.931258 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:40Z","lastTransitionTime":"2025-12-06T03:40:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.944362 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fad9b63d84f41fccf696cd313d7ca7f72b3d3251d6c1b7341573a3fadf1e378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:40Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.968732 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb3f278fff47c900c56d8a563d5ed75a2ae84e2034b0f45aad4083790b9e7946\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b10ca16ac26904ec5e6245344b71d836888cae9b069cae6a7a4831d609752db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:40Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:40 crc kubenswrapper[4802]: I1206 03:40:40.985080 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtw2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3a3bac2-3923-4507-a4de-1556b3ad473e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fed5ee4189d54456a3216f4fdb1ecf003fa47e29dbfe4b2b2f301edfb897d6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtw2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:40Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.000869 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-5ltx2"] Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.001266 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:40:41 crc kubenswrapper[4802]: E1206 03:40:41.001332 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5ltx2" podUID="8881599e-71e5-4578-b3f9-21f4fa2f57aa" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.011272 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7730de4f033561ad77046e028cd1a3f0bfe2775387cbdfa35f47ea79909ec4cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a158aebd701ef46225d3e37109ae28ab58d107087bf06650c4b6d0055a3bbba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2357f37497fde64937eca7b38693893af745ab278653ebc79406e122a6d4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8941bd25107333b264d3f081a811fdfeea23ad92193fc61367665fd2cd1209c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f9cd4dc75e63f09f60b48b6da284643910bb32ba51f2a808cb4d9803f30042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6796f65ee9c5bd0c18f205697135d17677c5daf3ba378cc992cba795e0188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf845e8893e51078e46991489d9a27c1194eab62a9d0a24eb3336d396e12cdfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf845e8893e51078e46991489d9a27c1194eab62a9d0a24eb3336d396e12cdfb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:40:38Z\\\",\\\"message\\\":\\\"Map:map[10.217.4.110:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {f9232b32-e89f-4c8e-acc4-c6801b70dcb0}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:NB_Global Row:map[] Rows:[] Columns:[] Mutations:[{Column:nb_cfg Mutator:+= Value:1}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {6011affd-30a6-4be6-872d-e4cf1ca780cf}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 03:40:38.703973 6236 ovnkube.go:599] Stopped ovnkube\\\\nI1206 03:40:38.704014 6236 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1206 03:40:38.703921 6236 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-controller]} name:Service_openshift-machine-config-operator/machine-config-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.16:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {3f1b9878-e751-4e46-a226-ce007d2c4aa7}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1206 03:40:38.704093 6236 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-pvm6q_openshift-ovn-kubernetes(eafce4e0-e7fb-4877-b0ab-3283829ba1ef)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3cb210bfc97a439a7ff98c02c59f8ad16bc400dc214a616a44b43a8b01e2782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pvm6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:41Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.026341 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:41Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.026707 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8881599e-71e5-4578-b3f9-21f4fa2f57aa-metrics-certs\") pod \"network-metrics-daemon-5ltx2\" (UID: \"8881599e-71e5-4578-b3f9-21f4fa2f57aa\") " pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.026803 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbm6t\" (UniqueName: \"kubernetes.io/projected/8881599e-71e5-4578-b3f9-21f4fa2f57aa-kube-api-access-lbm6t\") pod \"network-metrics-daemon-5ltx2\" (UID: \"8881599e-71e5-4578-b3f9-21f4fa2f57aa\") " pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.033786 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.033832 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.033846 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.033862 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.033873 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:41Z","lastTransitionTime":"2025-12-06T03:40:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.040970 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:41Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.055208 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtw2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3a3bac2-3923-4507-a4de-1556b3ad473e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fed5ee4189d54456a3216f4fdb1ecf003fa47e29dbfe4b2b2f301edfb897d6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtw2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:41Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.074980 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7730de4f033561ad77046e028cd1a3f0bfe2775387cbdfa35f47ea79909ec4cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a158aebd701ef46225d3e37109ae28ab58d107087bf06650c4b6d0055a3bbba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2357f37497fde64937eca7b38693893af745ab278653ebc79406e122a6d4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8941bd25107333b264d3f081a811fdfeea23ad92193fc61367665fd2cd1209c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f9cd4dc75e63f09f60b48b6da284643910bb32ba51f2a808cb4d9803f30042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6796f65ee9c5bd0c18f205697135d17677c5daf3ba378cc992cba795e0188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf845e8893e51078e46991489d9a27c1194eab62a9d0a24eb3336d396e12cdfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf845e8893e51078e46991489d9a27c1194eab62a9d0a24eb3336d396e12cdfb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:40:38Z\\\",\\\"message\\\":\\\"Map:map[10.217.4.110:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {f9232b32-e89f-4c8e-acc4-c6801b70dcb0}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:NB_Global Row:map[] Rows:[] Columns:[] Mutations:[{Column:nb_cfg Mutator:+= Value:1}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {6011affd-30a6-4be6-872d-e4cf1ca780cf}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 03:40:38.703973 6236 ovnkube.go:599] Stopped ovnkube\\\\nI1206 03:40:38.704014 6236 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1206 03:40:38.703921 6236 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-controller]} name:Service_openshift-machine-config-operator/machine-config-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.16:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {3f1b9878-e751-4e46-a226-ce007d2c4aa7}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1206 03:40:38.704093 6236 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-pvm6q_openshift-ovn-kubernetes(eafce4e0-e7fb-4877-b0ab-3283829ba1ef)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3cb210bfc97a439a7ff98c02c59f8ad16bc400dc214a616a44b43a8b01e2782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pvm6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:41Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.084911 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-b2knv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a3ae46f-03b5-4999-9d58-c38ee222f68a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76665d4c127445009fd66078827b54607c9ef812c8011d1c0b4cde1c0516853a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk4hj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-b2knv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:41Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.099155 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22rrq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03112a5d-5882-4fac-9ae5-13525ff82fe0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0205b089b12c42688203bb7ec0f27dc5f9a577dabdb446b811f184046f85dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trlt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22rrq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:41Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.109040 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ngspk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34808ab2-d21b-40b6-8fe3-e18bf9e5fc46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qkf8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qkf8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ngspk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:41Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.117349 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5ltx2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8881599e-71e5-4578-b3f9-21f4fa2f57aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbm6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbm6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5ltx2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:41Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.127653 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbm6t\" (UniqueName: \"kubernetes.io/projected/8881599e-71e5-4578-b3f9-21f4fa2f57aa-kube-api-access-lbm6t\") pod \"network-metrics-daemon-5ltx2\" (UID: \"8881599e-71e5-4578-b3f9-21f4fa2f57aa\") " pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.127719 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8881599e-71e5-4578-b3f9-21f4fa2f57aa-metrics-certs\") pod \"network-metrics-daemon-5ltx2\" (UID: \"8881599e-71e5-4578-b3f9-21f4fa2f57aa\") " pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:40:41 crc kubenswrapper[4802]: E1206 03:40:41.127843 4802 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 03:40:41 crc kubenswrapper[4802]: E1206 03:40:41.127892 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8881599e-71e5-4578-b3f9-21f4fa2f57aa-metrics-certs podName:8881599e-71e5-4578-b3f9-21f4fa2f57aa nodeName:}" failed. No retries permitted until 2025-12-06 03:40:41.627876132 +0000 UTC m=+34.499785294 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8881599e-71e5-4578-b3f9-21f4fa2f57aa-metrics-certs") pod "network-metrics-daemon-5ltx2" (UID: "8881599e-71e5-4578-b3f9-21f4fa2f57aa") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.128543 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1721e985-40bd-4b2d-a0e8-d4e365094ba6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a895e0cc04880c3287ba725ac5484835235f1e066a94ce98fcf4b6133c5cbd2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a21a97385f856ec996105b597d0ae81c56ed28b625ef3f2905e1af6ec8f6a45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://646e01a58908ba2e703aae5eef53e1b981b0f0c3d1d3202487f2a8b94ee3f55d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31752b84d86bda30792b29e4c6ed963753cc576c5de784395e8e6dcbebaced47\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"ver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:40:26.813730 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814130 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:40:26.814139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:40:26.814143 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:40:26.814145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 03:40:26.821403 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821442 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821482 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821496 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 03:40:26.821509 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821514 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 03:40:26.821612 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 03:40:26.821619 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1206 03:40:26.822049 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16a92d501c9e626fe91e4a6793e426d10eb292cfbab766db8b8cc72e5dae94ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:41Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.136021 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.136149 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.136239 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.136334 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.136416 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:41Z","lastTransitionTime":"2025-12-06T03:40:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.137705 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kj7hf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd46dfa5-7d4a-49ed-8199-523f10cebdf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5eee8a0cf96ceb3785dac9df31bf8d31715d1c88fbea17c5e582b0a4554bb63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmkcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kj7hf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:41Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.147243 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2466610962332c6ca120fe055e4e8a3fc623088119dc1a412e06b745176f89e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:41Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.147431 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbm6t\" (UniqueName: \"kubernetes.io/projected/8881599e-71e5-4578-b3f9-21f4fa2f57aa-kube-api-access-lbm6t\") pod \"network-metrics-daemon-5ltx2\" (UID: \"8881599e-71e5-4578-b3f9-21f4fa2f57aa\") " pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.159305 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:41Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.168423 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6df38316-e0d3-4018-8d27-3620eba3a68d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57381e9589a2dc04a4d2733c37985f42594182128a374b7be0f471ea69138b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bf5e41701c7a2d45adc4740ceb4ac1785db17bc7cb15daed439f7556812d1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpxxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:41Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.181106 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fad9b63d84f41fccf696cd313d7ca7f72b3d3251d6c1b7341573a3fadf1e378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:41Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.193213 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb3f278fff47c900c56d8a563d5ed75a2ae84e2034b0f45aad4083790b9e7946\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b10ca16ac26904ec5e6245344b71d836888cae9b069cae6a7a4831d609752db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:41Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.208103 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2093b6-c4b5-4db8-be16-fbc859885e8d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0d6e582161f0d9126126e03b27211be22f6b5f797154928c92fa13a8118f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03004a29aefe7fe2452fb9f6663c0807afccb055d7ddbf6b907f305e6a19c81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1801ac64db4ba91c2e8a249e3651d3e50ab62e9f2c87312ce9bf7f38518a82cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6e454507cadbb982a566a3692995aaa3c733ff1cd7fb4870476602190e1eeba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:41Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.224596 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:41Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.238678 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.238732 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.238762 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.238783 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.238798 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:41Z","lastTransitionTime":"2025-12-06T03:40:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.341499 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.341548 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.341559 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.341579 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.341591 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:41Z","lastTransitionTime":"2025-12-06T03:40:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.444791 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.444862 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.444884 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.444912 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.444930 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:41Z","lastTransitionTime":"2025-12-06T03:40:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.548239 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.548278 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.548290 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.548310 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.548322 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:41Z","lastTransitionTime":"2025-12-06T03:40:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.633581 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8881599e-71e5-4578-b3f9-21f4fa2f57aa-metrics-certs\") pod \"network-metrics-daemon-5ltx2\" (UID: \"8881599e-71e5-4578-b3f9-21f4fa2f57aa\") " pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:40:41 crc kubenswrapper[4802]: E1206 03:40:41.633833 4802 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 03:40:41 crc kubenswrapper[4802]: E1206 03:40:41.633945 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8881599e-71e5-4578-b3f9-21f4fa2f57aa-metrics-certs podName:8881599e-71e5-4578-b3f9-21f4fa2f57aa nodeName:}" failed. No retries permitted until 2025-12-06 03:40:42.633919924 +0000 UTC m=+35.505829186 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8881599e-71e5-4578-b3f9-21f4fa2f57aa-metrics-certs") pod "network-metrics-daemon-5ltx2" (UID: "8881599e-71e5-4578-b3f9-21f4fa2f57aa") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.658207 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.658270 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.658288 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.658315 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.658332 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:41Z","lastTransitionTime":"2025-12-06T03:40:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.761745 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.761826 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.761843 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.761862 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.761880 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:41Z","lastTransitionTime":"2025-12-06T03:40:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.774459 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ngspk" event={"ID":"34808ab2-d21b-40b6-8fe3-e18bf9e5fc46","Type":"ContainerStarted","Data":"9685a4d5d4f40ff4f59b6da6193785f340890453d1e422d3421a621dc7bd02f5"} Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.792828 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5ltx2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8881599e-71e5-4578-b3f9-21f4fa2f57aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbm6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbm6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5ltx2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:41Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.813213 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1721e985-40bd-4b2d-a0e8-d4e365094ba6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a895e0cc04880c3287ba725ac5484835235f1e066a94ce98fcf4b6133c5cbd2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a21a97385f856ec996105b597d0ae81c56ed28b625ef3f2905e1af6ec8f6a45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://646e01a58908ba2e703aae5eef53e1b981b0f0c3d1d3202487f2a8b94ee3f55d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31752b84d86bda30792b29e4c6ed963753cc576c5de784395e8e6dcbebaced47\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"ver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:40:26.813730 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814130 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:40:26.814139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:40:26.814143 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:40:26.814145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 03:40:26.821403 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821442 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821482 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821496 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 03:40:26.821509 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821514 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 03:40:26.821612 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 03:40:26.821619 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1206 03:40:26.822049 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16a92d501c9e626fe91e4a6793e426d10eb292cfbab766db8b8cc72e5dae94ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:41Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.830159 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-b2knv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a3ae46f-03b5-4999-9d58-c38ee222f68a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76665d4c127445009fd66078827b54607c9ef812c8011d1c0b4cde1c0516853a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk4hj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-b2knv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:41Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.845817 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22rrq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03112a5d-5882-4fac-9ae5-13525ff82fe0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0205b089b12c42688203bb7ec0f27dc5f9a577dabdb446b811f184046f85dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trlt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22rrq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:41Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.859744 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ngspk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34808ab2-d21b-40b6-8fe3-e18bf9e5fc46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52509d21f5b4bfbd458162cba176ff82c627c7c4671204addf9d76f7392fd760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qkf8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9685a4d5d4f40ff4f59b6da6193785f340890453d1e422d3421a621dc7bd02f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qkf8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ngspk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:41Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.864099 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.864161 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.864183 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.864210 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.864228 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:41Z","lastTransitionTime":"2025-12-06T03:40:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.873024 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2466610962332c6ca120fe055e4e8a3fc623088119dc1a412e06b745176f89e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:41Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.887285 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:41Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.901274 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6df38316-e0d3-4018-8d27-3620eba3a68d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57381e9589a2dc04a4d2733c37985f42594182128a374b7be0f471ea69138b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bf5e41701c7a2d45adc4740ceb4ac1785db17bc7cb15daed439f7556812d1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpxxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:41Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.916040 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kj7hf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd46dfa5-7d4a-49ed-8199-523f10cebdf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5eee8a0cf96ceb3785dac9df31bf8d31715d1c88fbea17c5e582b0a4554bb63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmkcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kj7hf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:41Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.933619 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2093b6-c4b5-4db8-be16-fbc859885e8d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0d6e582161f0d9126126e03b27211be22f6b5f797154928c92fa13a8118f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03004a29aefe7fe2452fb9f6663c0807afccb055d7ddbf6b907f305e6a19c81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1801ac64db4ba91c2e8a249e3651d3e50ab62e9f2c87312ce9bf7f38518a82cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6e454507cadbb982a566a3692995aaa3c733ff1cd7fb4870476602190e1eeba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:41Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.951442 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:41Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.966592 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.966664 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.966688 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.966720 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.966744 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:41Z","lastTransitionTime":"2025-12-06T03:40:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.973320 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fad9b63d84f41fccf696cd313d7ca7f72b3d3251d6c1b7341573a3fadf1e378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:41Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:41 crc kubenswrapper[4802]: I1206 03:40:41.992878 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb3f278fff47c900c56d8a563d5ed75a2ae84e2034b0f45aad4083790b9e7946\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b10ca16ac26904ec5e6245344b71d836888cae9b069cae6a7a4831d609752db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:41Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:42 crc kubenswrapper[4802]: I1206 03:40:42.008700 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:42Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:42 crc kubenswrapper[4802]: I1206 03:40:42.025564 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtw2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3a3bac2-3923-4507-a4de-1556b3ad473e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fed5ee4189d54456a3216f4fdb1ecf003fa47e29dbfe4b2b2f301edfb897d6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtw2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:42Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:42 crc kubenswrapper[4802]: I1206 03:40:42.055462 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7730de4f033561ad77046e028cd1a3f0bfe2775387cbdfa35f47ea79909ec4cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a158aebd701ef46225d3e37109ae28ab58d107087bf06650c4b6d0055a3bbba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2357f37497fde64937eca7b38693893af745ab278653ebc79406e122a6d4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8941bd25107333b264d3f081a811fdfeea23ad92193fc61367665fd2cd1209c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f9cd4dc75e63f09f60b48b6da284643910bb32ba51f2a808cb4d9803f30042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6796f65ee9c5bd0c18f205697135d17677c5daf3ba378cc992cba795e0188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf845e8893e51078e46991489d9a27c1194eab62a9d0a24eb3336d396e12cdfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf845e8893e51078e46991489d9a27c1194eab62a9d0a24eb3336d396e12cdfb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:40:38Z\\\",\\\"message\\\":\\\"Map:map[10.217.4.110:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {f9232b32-e89f-4c8e-acc4-c6801b70dcb0}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:NB_Global Row:map[] Rows:[] Columns:[] Mutations:[{Column:nb_cfg Mutator:+= Value:1}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {6011affd-30a6-4be6-872d-e4cf1ca780cf}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 03:40:38.703973 6236 ovnkube.go:599] Stopped ovnkube\\\\nI1206 03:40:38.704014 6236 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1206 03:40:38.703921 6236 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-controller]} name:Service_openshift-machine-config-operator/machine-config-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.16:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {3f1b9878-e751-4e46-a226-ce007d2c4aa7}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1206 03:40:38.704093 6236 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-pvm6q_openshift-ovn-kubernetes(eafce4e0-e7fb-4877-b0ab-3283829ba1ef)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3cb210bfc97a439a7ff98c02c59f8ad16bc400dc214a616a44b43a8b01e2782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pvm6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:42Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:42 crc kubenswrapper[4802]: I1206 03:40:42.069151 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:42 crc kubenswrapper[4802]: I1206 03:40:42.069214 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:42 crc kubenswrapper[4802]: I1206 03:40:42.069232 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:42 crc kubenswrapper[4802]: I1206 03:40:42.069257 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:42 crc kubenswrapper[4802]: I1206 03:40:42.069276 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:42Z","lastTransitionTime":"2025-12-06T03:40:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:42 crc kubenswrapper[4802]: I1206 03:40:42.172299 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:42 crc kubenswrapper[4802]: I1206 03:40:42.172422 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:42 crc kubenswrapper[4802]: I1206 03:40:42.172451 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:42 crc kubenswrapper[4802]: I1206 03:40:42.172529 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:42 crc kubenswrapper[4802]: I1206 03:40:42.172553 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:42Z","lastTransitionTime":"2025-12-06T03:40:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:42 crc kubenswrapper[4802]: I1206 03:40:42.241006 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:40:42 crc kubenswrapper[4802]: E1206 03:40:42.241234 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 03:40:42 crc kubenswrapper[4802]: E1206 03:40:42.241278 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 03:40:42 crc kubenswrapper[4802]: E1206 03:40:42.241305 4802 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:40:42 crc kubenswrapper[4802]: E1206 03:40:42.241394 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 03:40:58.241371756 +0000 UTC m=+51.113280948 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:40:42 crc kubenswrapper[4802]: I1206 03:40:42.276224 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:42 crc kubenswrapper[4802]: I1206 03:40:42.276354 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:42 crc kubenswrapper[4802]: I1206 03:40:42.276380 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:42 crc kubenswrapper[4802]: I1206 03:40:42.276405 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:42 crc kubenswrapper[4802]: I1206 03:40:42.276422 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:42Z","lastTransitionTime":"2025-12-06T03:40:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:42 crc kubenswrapper[4802]: I1206 03:40:42.341745 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:40:42 crc kubenswrapper[4802]: I1206 03:40:42.341879 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:40:42 crc kubenswrapper[4802]: I1206 03:40:42.341957 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:40:42 crc kubenswrapper[4802]: E1206 03:40:42.341999 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:40:58.341971284 +0000 UTC m=+51.213880466 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:40:42 crc kubenswrapper[4802]: I1206 03:40:42.342093 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:40:42 crc kubenswrapper[4802]: E1206 03:40:42.342141 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 03:40:42 crc kubenswrapper[4802]: E1206 03:40:42.342166 4802 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 03:40:42 crc kubenswrapper[4802]: E1206 03:40:42.342277 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 03:40:58.342248911 +0000 UTC m=+51.214158103 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 03:40:42 crc kubenswrapper[4802]: E1206 03:40:42.342324 4802 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 03:40:42 crc kubenswrapper[4802]: E1206 03:40:42.342472 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 03:40:58.342442887 +0000 UTC m=+51.214352079 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 03:40:42 crc kubenswrapper[4802]: E1206 03:40:42.342173 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 03:40:42 crc kubenswrapper[4802]: E1206 03:40:42.342518 4802 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:40:42 crc kubenswrapper[4802]: E1206 03:40:42.342583 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 03:40:58.34256697 +0000 UTC m=+51.214476152 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:40:42 crc kubenswrapper[4802]: I1206 03:40:42.379526 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:42 crc kubenswrapper[4802]: I1206 03:40:42.379587 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:42 crc kubenswrapper[4802]: I1206 03:40:42.379609 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:42 crc kubenswrapper[4802]: I1206 03:40:42.379636 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:42 crc kubenswrapper[4802]: I1206 03:40:42.379657 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:42Z","lastTransitionTime":"2025-12-06T03:40:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:42 crc kubenswrapper[4802]: I1206 03:40:42.449626 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:40:42 crc kubenswrapper[4802]: I1206 03:40:42.449682 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:40:42 crc kubenswrapper[4802]: I1206 03:40:42.449696 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:40:42 crc kubenswrapper[4802]: I1206 03:40:42.449626 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:40:42 crc kubenswrapper[4802]: E1206 03:40:42.449876 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5ltx2" podUID="8881599e-71e5-4578-b3f9-21f4fa2f57aa" Dec 06 03:40:42 crc kubenswrapper[4802]: E1206 03:40:42.450018 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:40:42 crc kubenswrapper[4802]: E1206 03:40:42.450122 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:40:42 crc kubenswrapper[4802]: E1206 03:40:42.450265 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:40:42 crc kubenswrapper[4802]: I1206 03:40:42.482550 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:42 crc kubenswrapper[4802]: I1206 03:40:42.482618 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:42 crc kubenswrapper[4802]: I1206 03:40:42.482641 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:42 crc kubenswrapper[4802]: I1206 03:40:42.482680 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:42 crc kubenswrapper[4802]: I1206 03:40:42.482708 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:42Z","lastTransitionTime":"2025-12-06T03:40:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:42 crc kubenswrapper[4802]: I1206 03:40:42.585840 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:42 crc kubenswrapper[4802]: I1206 03:40:42.585884 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:42 crc kubenswrapper[4802]: I1206 03:40:42.585897 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:42 crc kubenswrapper[4802]: I1206 03:40:42.585914 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:42 crc kubenswrapper[4802]: I1206 03:40:42.585926 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:42Z","lastTransitionTime":"2025-12-06T03:40:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:42 crc kubenswrapper[4802]: I1206 03:40:42.644576 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8881599e-71e5-4578-b3f9-21f4fa2f57aa-metrics-certs\") pod \"network-metrics-daemon-5ltx2\" (UID: \"8881599e-71e5-4578-b3f9-21f4fa2f57aa\") " pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:40:42 crc kubenswrapper[4802]: E1206 03:40:42.644744 4802 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 03:40:42 crc kubenswrapper[4802]: E1206 03:40:42.644859 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8881599e-71e5-4578-b3f9-21f4fa2f57aa-metrics-certs podName:8881599e-71e5-4578-b3f9-21f4fa2f57aa nodeName:}" failed. No retries permitted until 2025-12-06 03:40:44.644839375 +0000 UTC m=+37.516748537 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8881599e-71e5-4578-b3f9-21f4fa2f57aa-metrics-certs") pod "network-metrics-daemon-5ltx2" (UID: "8881599e-71e5-4578-b3f9-21f4fa2f57aa") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 03:40:42 crc kubenswrapper[4802]: I1206 03:40:42.688935 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:42 crc kubenswrapper[4802]: I1206 03:40:42.688985 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:42 crc kubenswrapper[4802]: I1206 03:40:42.689045 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:42 crc kubenswrapper[4802]: I1206 03:40:42.689069 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:42 crc kubenswrapper[4802]: I1206 03:40:42.689086 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:42Z","lastTransitionTime":"2025-12-06T03:40:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:42 crc kubenswrapper[4802]: I1206 03:40:42.791704 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:42 crc kubenswrapper[4802]: I1206 03:40:42.791801 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:42 crc kubenswrapper[4802]: I1206 03:40:42.791820 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:42 crc kubenswrapper[4802]: I1206 03:40:42.791844 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:42 crc kubenswrapper[4802]: I1206 03:40:42.791864 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:42Z","lastTransitionTime":"2025-12-06T03:40:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:42 crc kubenswrapper[4802]: I1206 03:40:42.895371 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:42 crc kubenswrapper[4802]: I1206 03:40:42.895417 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:42 crc kubenswrapper[4802]: I1206 03:40:42.895632 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:42 crc kubenswrapper[4802]: I1206 03:40:42.895660 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:42 crc kubenswrapper[4802]: I1206 03:40:42.895677 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:42Z","lastTransitionTime":"2025-12-06T03:40:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:43 crc kubenswrapper[4802]: I1206 03:40:43.000377 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:43 crc kubenswrapper[4802]: I1206 03:40:43.000431 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:43 crc kubenswrapper[4802]: I1206 03:40:43.000453 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:43 crc kubenswrapper[4802]: I1206 03:40:43.000549 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:43 crc kubenswrapper[4802]: I1206 03:40:43.000580 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:43Z","lastTransitionTime":"2025-12-06T03:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:43 crc kubenswrapper[4802]: I1206 03:40:43.103689 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:43 crc kubenswrapper[4802]: I1206 03:40:43.103783 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:43 crc kubenswrapper[4802]: I1206 03:40:43.103812 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:43 crc kubenswrapper[4802]: I1206 03:40:43.103842 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:43 crc kubenswrapper[4802]: I1206 03:40:43.103865 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:43Z","lastTransitionTime":"2025-12-06T03:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:43 crc kubenswrapper[4802]: I1206 03:40:43.207162 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:43 crc kubenswrapper[4802]: I1206 03:40:43.207243 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:43 crc kubenswrapper[4802]: I1206 03:40:43.207275 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:43 crc kubenswrapper[4802]: I1206 03:40:43.207315 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:43 crc kubenswrapper[4802]: I1206 03:40:43.207336 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:43Z","lastTransitionTime":"2025-12-06T03:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:43 crc kubenswrapper[4802]: I1206 03:40:43.310985 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:43 crc kubenswrapper[4802]: I1206 03:40:43.311054 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:43 crc kubenswrapper[4802]: I1206 03:40:43.311073 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:43 crc kubenswrapper[4802]: I1206 03:40:43.311100 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:43 crc kubenswrapper[4802]: I1206 03:40:43.311122 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:43Z","lastTransitionTime":"2025-12-06T03:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:43 crc kubenswrapper[4802]: I1206 03:40:43.413923 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:43 crc kubenswrapper[4802]: I1206 03:40:43.414027 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:43 crc kubenswrapper[4802]: I1206 03:40:43.414045 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:43 crc kubenswrapper[4802]: I1206 03:40:43.414069 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:43 crc kubenswrapper[4802]: I1206 03:40:43.414088 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:43Z","lastTransitionTime":"2025-12-06T03:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:43 crc kubenswrapper[4802]: I1206 03:40:43.516451 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:43 crc kubenswrapper[4802]: I1206 03:40:43.516505 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:43 crc kubenswrapper[4802]: I1206 03:40:43.516522 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:43 crc kubenswrapper[4802]: I1206 03:40:43.516547 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:43 crc kubenswrapper[4802]: I1206 03:40:43.516564 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:43Z","lastTransitionTime":"2025-12-06T03:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:43 crc kubenswrapper[4802]: I1206 03:40:43.619927 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:43 crc kubenswrapper[4802]: I1206 03:40:43.619981 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:43 crc kubenswrapper[4802]: I1206 03:40:43.619999 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:43 crc kubenswrapper[4802]: I1206 03:40:43.620024 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:43 crc kubenswrapper[4802]: I1206 03:40:43.620044 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:43Z","lastTransitionTime":"2025-12-06T03:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:43 crc kubenswrapper[4802]: I1206 03:40:43.722407 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:43 crc kubenswrapper[4802]: I1206 03:40:43.722548 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:43 crc kubenswrapper[4802]: I1206 03:40:43.722571 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:43 crc kubenswrapper[4802]: I1206 03:40:43.722595 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:43 crc kubenswrapper[4802]: I1206 03:40:43.722615 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:43Z","lastTransitionTime":"2025-12-06T03:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:43 crc kubenswrapper[4802]: I1206 03:40:43.825710 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:43 crc kubenswrapper[4802]: I1206 03:40:43.825794 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:43 crc kubenswrapper[4802]: I1206 03:40:43.825817 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:43 crc kubenswrapper[4802]: I1206 03:40:43.825894 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:43 crc kubenswrapper[4802]: I1206 03:40:43.825976 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:43Z","lastTransitionTime":"2025-12-06T03:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:43 crc kubenswrapper[4802]: I1206 03:40:43.929397 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:43 crc kubenswrapper[4802]: I1206 03:40:43.929460 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:43 crc kubenswrapper[4802]: I1206 03:40:43.929478 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:43 crc kubenswrapper[4802]: I1206 03:40:43.929505 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:43 crc kubenswrapper[4802]: I1206 03:40:43.929522 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:43Z","lastTransitionTime":"2025-12-06T03:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:44 crc kubenswrapper[4802]: I1206 03:40:44.032520 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:44 crc kubenswrapper[4802]: I1206 03:40:44.032587 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:44 crc kubenswrapper[4802]: I1206 03:40:44.032643 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:44 crc kubenswrapper[4802]: I1206 03:40:44.032678 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:44 crc kubenswrapper[4802]: I1206 03:40:44.032701 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:44Z","lastTransitionTime":"2025-12-06T03:40:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:44 crc kubenswrapper[4802]: I1206 03:40:44.136095 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:44 crc kubenswrapper[4802]: I1206 03:40:44.136185 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:44 crc kubenswrapper[4802]: I1206 03:40:44.136198 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:44 crc kubenswrapper[4802]: I1206 03:40:44.136225 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:44 crc kubenswrapper[4802]: I1206 03:40:44.136239 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:44Z","lastTransitionTime":"2025-12-06T03:40:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:44 crc kubenswrapper[4802]: I1206 03:40:44.239307 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:44 crc kubenswrapper[4802]: I1206 03:40:44.239453 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:44 crc kubenswrapper[4802]: I1206 03:40:44.239474 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:44 crc kubenswrapper[4802]: I1206 03:40:44.239498 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:44 crc kubenswrapper[4802]: I1206 03:40:44.239599 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:44Z","lastTransitionTime":"2025-12-06T03:40:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:44 crc kubenswrapper[4802]: I1206 03:40:44.342570 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:44 crc kubenswrapper[4802]: I1206 03:40:44.342659 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:44 crc kubenswrapper[4802]: I1206 03:40:44.342678 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:44 crc kubenswrapper[4802]: I1206 03:40:44.342699 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:44 crc kubenswrapper[4802]: I1206 03:40:44.342718 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:44Z","lastTransitionTime":"2025-12-06T03:40:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:44 crc kubenswrapper[4802]: I1206 03:40:44.445571 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:44 crc kubenswrapper[4802]: I1206 03:40:44.445658 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:44 crc kubenswrapper[4802]: I1206 03:40:44.445681 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:44 crc kubenswrapper[4802]: I1206 03:40:44.445719 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:44 crc kubenswrapper[4802]: I1206 03:40:44.445789 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:44Z","lastTransitionTime":"2025-12-06T03:40:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:44 crc kubenswrapper[4802]: I1206 03:40:44.449973 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:40:44 crc kubenswrapper[4802]: I1206 03:40:44.450021 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:40:44 crc kubenswrapper[4802]: I1206 03:40:44.450021 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:40:44 crc kubenswrapper[4802]: E1206 03:40:44.450168 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5ltx2" podUID="8881599e-71e5-4578-b3f9-21f4fa2f57aa" Dec 06 03:40:44 crc kubenswrapper[4802]: E1206 03:40:44.450270 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:40:44 crc kubenswrapper[4802]: E1206 03:40:44.450367 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:40:44 crc kubenswrapper[4802]: I1206 03:40:44.450652 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:40:44 crc kubenswrapper[4802]: E1206 03:40:44.451030 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:40:44 crc kubenswrapper[4802]: I1206 03:40:44.549291 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:44 crc kubenswrapper[4802]: I1206 03:40:44.549414 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:44 crc kubenswrapper[4802]: I1206 03:40:44.549444 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:44 crc kubenswrapper[4802]: I1206 03:40:44.549475 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:44 crc kubenswrapper[4802]: I1206 03:40:44.549499 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:44Z","lastTransitionTime":"2025-12-06T03:40:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:44 crc kubenswrapper[4802]: I1206 03:40:44.652178 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:44 crc kubenswrapper[4802]: I1206 03:40:44.652227 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:44 crc kubenswrapper[4802]: I1206 03:40:44.652246 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:44 crc kubenswrapper[4802]: I1206 03:40:44.652269 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:44 crc kubenswrapper[4802]: I1206 03:40:44.652287 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:44Z","lastTransitionTime":"2025-12-06T03:40:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:44 crc kubenswrapper[4802]: I1206 03:40:44.668573 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8881599e-71e5-4578-b3f9-21f4fa2f57aa-metrics-certs\") pod \"network-metrics-daemon-5ltx2\" (UID: \"8881599e-71e5-4578-b3f9-21f4fa2f57aa\") " pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:40:44 crc kubenswrapper[4802]: E1206 03:40:44.668793 4802 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 03:40:44 crc kubenswrapper[4802]: E1206 03:40:44.668895 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8881599e-71e5-4578-b3f9-21f4fa2f57aa-metrics-certs podName:8881599e-71e5-4578-b3f9-21f4fa2f57aa nodeName:}" failed. No retries permitted until 2025-12-06 03:40:48.668868387 +0000 UTC m=+41.540777569 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8881599e-71e5-4578-b3f9-21f4fa2f57aa-metrics-certs") pod "network-metrics-daemon-5ltx2" (UID: "8881599e-71e5-4578-b3f9-21f4fa2f57aa") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 03:40:44 crc kubenswrapper[4802]: I1206 03:40:44.754377 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:44 crc kubenswrapper[4802]: I1206 03:40:44.754423 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:44 crc kubenswrapper[4802]: I1206 03:40:44.754442 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:44 crc kubenswrapper[4802]: I1206 03:40:44.754465 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:44 crc kubenswrapper[4802]: I1206 03:40:44.754482 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:44Z","lastTransitionTime":"2025-12-06T03:40:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:44 crc kubenswrapper[4802]: I1206 03:40:44.857394 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:44 crc kubenswrapper[4802]: I1206 03:40:44.857427 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:44 crc kubenswrapper[4802]: I1206 03:40:44.857438 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:44 crc kubenswrapper[4802]: I1206 03:40:44.857451 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:44 crc kubenswrapper[4802]: I1206 03:40:44.857462 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:44Z","lastTransitionTime":"2025-12-06T03:40:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:44 crc kubenswrapper[4802]: I1206 03:40:44.963674 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:44 crc kubenswrapper[4802]: I1206 03:40:44.963742 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:44 crc kubenswrapper[4802]: I1206 03:40:44.963783 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:44 crc kubenswrapper[4802]: I1206 03:40:44.963803 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:44 crc kubenswrapper[4802]: I1206 03:40:44.963818 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:44Z","lastTransitionTime":"2025-12-06T03:40:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:45 crc kubenswrapper[4802]: I1206 03:40:45.067144 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:45 crc kubenswrapper[4802]: I1206 03:40:45.067591 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:45 crc kubenswrapper[4802]: I1206 03:40:45.067791 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:45 crc kubenswrapper[4802]: I1206 03:40:45.067991 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:45 crc kubenswrapper[4802]: I1206 03:40:45.068144 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:45Z","lastTransitionTime":"2025-12-06T03:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:45 crc kubenswrapper[4802]: I1206 03:40:45.171904 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:45 crc kubenswrapper[4802]: I1206 03:40:45.171968 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:45 crc kubenswrapper[4802]: I1206 03:40:45.171986 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:45 crc kubenswrapper[4802]: I1206 03:40:45.172014 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:45 crc kubenswrapper[4802]: I1206 03:40:45.172032 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:45Z","lastTransitionTime":"2025-12-06T03:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:45 crc kubenswrapper[4802]: I1206 03:40:45.274825 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:45 crc kubenswrapper[4802]: I1206 03:40:45.274896 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:45 crc kubenswrapper[4802]: I1206 03:40:45.274908 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:45 crc kubenswrapper[4802]: I1206 03:40:45.274927 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:45 crc kubenswrapper[4802]: I1206 03:40:45.274939 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:45Z","lastTransitionTime":"2025-12-06T03:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:45 crc kubenswrapper[4802]: I1206 03:40:45.378124 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:45 crc kubenswrapper[4802]: I1206 03:40:45.378187 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:45 crc kubenswrapper[4802]: I1206 03:40:45.378204 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:45 crc kubenswrapper[4802]: I1206 03:40:45.378231 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:45 crc kubenswrapper[4802]: I1206 03:40:45.378253 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:45Z","lastTransitionTime":"2025-12-06T03:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:45 crc kubenswrapper[4802]: I1206 03:40:45.481495 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:45 crc kubenswrapper[4802]: I1206 03:40:45.481569 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:45 crc kubenswrapper[4802]: I1206 03:40:45.481589 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:45 crc kubenswrapper[4802]: I1206 03:40:45.481620 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:45 crc kubenswrapper[4802]: I1206 03:40:45.481720 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:45Z","lastTransitionTime":"2025-12-06T03:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:45 crc kubenswrapper[4802]: I1206 03:40:45.584840 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:45 crc kubenswrapper[4802]: I1206 03:40:45.584997 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:45 crc kubenswrapper[4802]: I1206 03:40:45.585020 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:45 crc kubenswrapper[4802]: I1206 03:40:45.585044 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:45 crc kubenswrapper[4802]: I1206 03:40:45.585098 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:45Z","lastTransitionTime":"2025-12-06T03:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:45 crc kubenswrapper[4802]: I1206 03:40:45.687920 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:45 crc kubenswrapper[4802]: I1206 03:40:45.687983 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:45 crc kubenswrapper[4802]: I1206 03:40:45.688000 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:45 crc kubenswrapper[4802]: I1206 03:40:45.688053 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:45 crc kubenswrapper[4802]: I1206 03:40:45.688071 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:45Z","lastTransitionTime":"2025-12-06T03:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:45 crc kubenswrapper[4802]: I1206 03:40:45.790699 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:45 crc kubenswrapper[4802]: I1206 03:40:45.790785 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:45 crc kubenswrapper[4802]: I1206 03:40:45.790807 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:45 crc kubenswrapper[4802]: I1206 03:40:45.790831 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:45 crc kubenswrapper[4802]: I1206 03:40:45.790848 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:45Z","lastTransitionTime":"2025-12-06T03:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:45 crc kubenswrapper[4802]: I1206 03:40:45.894298 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:45 crc kubenswrapper[4802]: I1206 03:40:45.894360 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:45 crc kubenswrapper[4802]: I1206 03:40:45.894383 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:45 crc kubenswrapper[4802]: I1206 03:40:45.894413 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:45 crc kubenswrapper[4802]: I1206 03:40:45.894437 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:45Z","lastTransitionTime":"2025-12-06T03:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:45 crc kubenswrapper[4802]: I1206 03:40:45.997931 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:45 crc kubenswrapper[4802]: I1206 03:40:45.998002 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:45 crc kubenswrapper[4802]: I1206 03:40:45.998019 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:45 crc kubenswrapper[4802]: I1206 03:40:45.998046 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:45 crc kubenswrapper[4802]: I1206 03:40:45.998065 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:45Z","lastTransitionTime":"2025-12-06T03:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:46 crc kubenswrapper[4802]: I1206 03:40:46.100703 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:46 crc kubenswrapper[4802]: I1206 03:40:46.100786 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:46 crc kubenswrapper[4802]: I1206 03:40:46.100804 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:46 crc kubenswrapper[4802]: I1206 03:40:46.100827 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:46 crc kubenswrapper[4802]: I1206 03:40:46.100844 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:46Z","lastTransitionTime":"2025-12-06T03:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:46 crc kubenswrapper[4802]: I1206 03:40:46.204590 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:46 crc kubenswrapper[4802]: I1206 03:40:46.204685 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:46 crc kubenswrapper[4802]: I1206 03:40:46.204708 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:46 crc kubenswrapper[4802]: I1206 03:40:46.204739 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:46 crc kubenswrapper[4802]: I1206 03:40:46.204952 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:46Z","lastTransitionTime":"2025-12-06T03:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:46 crc kubenswrapper[4802]: I1206 03:40:46.308742 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:46 crc kubenswrapper[4802]: I1206 03:40:46.309261 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:46 crc kubenswrapper[4802]: I1206 03:40:46.309413 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:46 crc kubenswrapper[4802]: I1206 03:40:46.309570 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:46 crc kubenswrapper[4802]: I1206 03:40:46.309817 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:46Z","lastTransitionTime":"2025-12-06T03:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:46 crc kubenswrapper[4802]: I1206 03:40:46.413004 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:46 crc kubenswrapper[4802]: I1206 03:40:46.413323 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:46 crc kubenswrapper[4802]: I1206 03:40:46.413452 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:46 crc kubenswrapper[4802]: I1206 03:40:46.413628 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:46 crc kubenswrapper[4802]: I1206 03:40:46.413804 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:46Z","lastTransitionTime":"2025-12-06T03:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:46 crc kubenswrapper[4802]: I1206 03:40:46.449317 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:40:46 crc kubenswrapper[4802]: I1206 03:40:46.449488 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:40:46 crc kubenswrapper[4802]: E1206 03:40:46.449638 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:40:46 crc kubenswrapper[4802]: I1206 03:40:46.450013 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:40:46 crc kubenswrapper[4802]: I1206 03:40:46.450069 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:40:46 crc kubenswrapper[4802]: E1206 03:40:46.450266 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:40:46 crc kubenswrapper[4802]: E1206 03:40:46.450342 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5ltx2" podUID="8881599e-71e5-4578-b3f9-21f4fa2f57aa" Dec 06 03:40:46 crc kubenswrapper[4802]: E1206 03:40:46.450472 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:40:46 crc kubenswrapper[4802]: I1206 03:40:46.517597 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:46 crc kubenswrapper[4802]: I1206 03:40:46.517660 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:46 crc kubenswrapper[4802]: I1206 03:40:46.517677 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:46 crc kubenswrapper[4802]: I1206 03:40:46.517706 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:46 crc kubenswrapper[4802]: I1206 03:40:46.517728 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:46Z","lastTransitionTime":"2025-12-06T03:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:46 crc kubenswrapper[4802]: I1206 03:40:46.620436 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:46 crc kubenswrapper[4802]: I1206 03:40:46.620500 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:46 crc kubenswrapper[4802]: I1206 03:40:46.620522 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:46 crc kubenswrapper[4802]: I1206 03:40:46.620547 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:46 crc kubenswrapper[4802]: I1206 03:40:46.620564 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:46Z","lastTransitionTime":"2025-12-06T03:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:46 crc kubenswrapper[4802]: I1206 03:40:46.724075 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:46 crc kubenswrapper[4802]: I1206 03:40:46.724139 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:46 crc kubenswrapper[4802]: I1206 03:40:46.724156 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:46 crc kubenswrapper[4802]: I1206 03:40:46.724180 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:46 crc kubenswrapper[4802]: I1206 03:40:46.724197 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:46Z","lastTransitionTime":"2025-12-06T03:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:46 crc kubenswrapper[4802]: I1206 03:40:46.827394 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:46 crc kubenswrapper[4802]: I1206 03:40:46.827462 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:46 crc kubenswrapper[4802]: I1206 03:40:46.827484 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:46 crc kubenswrapper[4802]: I1206 03:40:46.827515 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:46 crc kubenswrapper[4802]: I1206 03:40:46.827537 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:46Z","lastTransitionTime":"2025-12-06T03:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:46 crc kubenswrapper[4802]: I1206 03:40:46.930464 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:46 crc kubenswrapper[4802]: I1206 03:40:46.930502 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:46 crc kubenswrapper[4802]: I1206 03:40:46.930541 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:46 crc kubenswrapper[4802]: I1206 03:40:46.930560 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:46 crc kubenswrapper[4802]: I1206 03:40:46.930571 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:46Z","lastTransitionTime":"2025-12-06T03:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.033222 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.033268 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.033281 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.033301 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.033313 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:47Z","lastTransitionTime":"2025-12-06T03:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.139052 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.139134 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.139158 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.139189 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.139222 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:47Z","lastTransitionTime":"2025-12-06T03:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.242962 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.243025 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.243043 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.243068 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.243086 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:47Z","lastTransitionTime":"2025-12-06T03:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.345612 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.345659 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.345670 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.345687 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.345701 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:47Z","lastTransitionTime":"2025-12-06T03:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.449091 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.449156 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.449262 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.449447 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.449535 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:47Z","lastTransitionTime":"2025-12-06T03:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.470451 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22rrq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03112a5d-5882-4fac-9ae5-13525ff82fe0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0205b089b12c42688203bb7ec0f27dc5f9a577dabdb446b811f184046f85dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trlt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22rrq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:47Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.488294 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ngspk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34808ab2-d21b-40b6-8fe3-e18bf9e5fc46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52509d21f5b4bfbd458162cba176ff82c627c7c4671204addf9d76f7392fd760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qkf8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9685a4d5d4f40ff4f59b6da6193785f340890453d1e422d3421a621dc7bd02f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qkf8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ngspk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:47Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.504980 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5ltx2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8881599e-71e5-4578-b3f9-21f4fa2f57aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbm6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbm6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5ltx2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:47Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.527605 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1721e985-40bd-4b2d-a0e8-d4e365094ba6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a895e0cc04880c3287ba725ac5484835235f1e066a94ce98fcf4b6133c5cbd2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a21a97385f856ec996105b597d0ae81c56ed28b625ef3f2905e1af6ec8f6a45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://646e01a58908ba2e703aae5eef53e1b981b0f0c3d1d3202487f2a8b94ee3f55d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31752b84d86bda30792b29e4c6ed963753cc576c5de784395e8e6dcbebaced47\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"ver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:40:26.813730 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814130 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:40:26.814139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:40:26.814143 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:40:26.814145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 03:40:26.821403 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821442 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821482 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821496 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 03:40:26.821509 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821514 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 03:40:26.821612 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 03:40:26.821619 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1206 03:40:26.822049 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16a92d501c9e626fe91e4a6793e426d10eb292cfbab766db8b8cc72e5dae94ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:47Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.542246 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-b2knv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a3ae46f-03b5-4999-9d58-c38ee222f68a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76665d4c127445009fd66078827b54607c9ef812c8011d1c0b4cde1c0516853a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk4hj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-b2knv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:47Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.551682 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.551717 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.551725 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.551738 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.551763 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:47Z","lastTransitionTime":"2025-12-06T03:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.558194 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2466610962332c6ca120fe055e4e8a3fc623088119dc1a412e06b745176f89e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:47Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.574314 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:47Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.585182 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6df38316-e0d3-4018-8d27-3620eba3a68d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57381e9589a2dc04a4d2733c37985f42594182128a374b7be0f471ea69138b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bf5e41701c7a2d45adc4740ceb4ac1785db17bc7cb15daed439f7556812d1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpxxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:47Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.594636 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kj7hf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd46dfa5-7d4a-49ed-8199-523f10cebdf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5eee8a0cf96ceb3785dac9df31bf8d31715d1c88fbea17c5e582b0a4554bb63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmkcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kj7hf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:47Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.605210 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb3f278fff47c900c56d8a563d5ed75a2ae84e2034b0f45aad4083790b9e7946\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b10ca16ac26904ec5e6245344b71d836888cae9b069cae6a7a4831d609752db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:47Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.617401 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2093b6-c4b5-4db8-be16-fbc859885e8d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0d6e582161f0d9126126e03b27211be22f6b5f797154928c92fa13a8118f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03004a29aefe7fe2452fb9f6663c0807afccb055d7ddbf6b907f305e6a19c81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1801ac64db4ba91c2e8a249e3651d3e50ab62e9f2c87312ce9bf7f38518a82cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6e454507cadbb982a566a3692995aaa3c733ff1cd7fb4870476602190e1eeba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:47Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.628917 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:47Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.640573 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fad9b63d84f41fccf696cd313d7ca7f72b3d3251d6c1b7341573a3fadf1e378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:47Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.653200 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.653236 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.653247 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.653263 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.653274 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:47Z","lastTransitionTime":"2025-12-06T03:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.654370 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:47Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.677216 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtw2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3a3bac2-3923-4507-a4de-1556b3ad473e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fed5ee4189d54456a3216f4fdb1ecf003fa47e29dbfe4b2b2f301edfb897d6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtw2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:47Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.701632 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7730de4f033561ad77046e028cd1a3f0bfe2775387cbdfa35f47ea79909ec4cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a158aebd701ef46225d3e37109ae28ab58d107087bf06650c4b6d0055a3bbba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2357f37497fde64937eca7b38693893af745ab278653ebc79406e122a6d4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8941bd25107333b264d3f081a811fdfeea23ad92193fc61367665fd2cd1209c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f9cd4dc75e63f09f60b48b6da284643910bb32ba51f2a808cb4d9803f30042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6796f65ee9c5bd0c18f205697135d17677c5daf3ba378cc992cba795e0188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf845e8893e51078e46991489d9a27c1194eab62a9d0a24eb3336d396e12cdfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf845e8893e51078e46991489d9a27c1194eab62a9d0a24eb3336d396e12cdfb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:40:38Z\\\",\\\"message\\\":\\\"Map:map[10.217.4.110:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {f9232b32-e89f-4c8e-acc4-c6801b70dcb0}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:NB_Global Row:map[] Rows:[] Columns:[] Mutations:[{Column:nb_cfg Mutator:+= Value:1}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {6011affd-30a6-4be6-872d-e4cf1ca780cf}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 03:40:38.703973 6236 ovnkube.go:599] Stopped ovnkube\\\\nI1206 03:40:38.704014 6236 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1206 03:40:38.703921 6236 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-controller]} name:Service_openshift-machine-config-operator/machine-config-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.16:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {3f1b9878-e751-4e46-a226-ce007d2c4aa7}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1206 03:40:38.704093 6236 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-pvm6q_openshift-ovn-kubernetes(eafce4e0-e7fb-4877-b0ab-3283829ba1ef)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3cb210bfc97a439a7ff98c02c59f8ad16bc400dc214a616a44b43a8b01e2782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pvm6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:47Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.755859 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.755928 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.755946 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.755973 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.755991 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:47Z","lastTransitionTime":"2025-12-06T03:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.859097 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.859176 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.859198 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.859231 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.859252 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:47Z","lastTransitionTime":"2025-12-06T03:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.962962 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.963054 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.963076 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.963113 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:47 crc kubenswrapper[4802]: I1206 03:40:47.963142 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:47Z","lastTransitionTime":"2025-12-06T03:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:48 crc kubenswrapper[4802]: I1206 03:40:48.066284 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:48 crc kubenswrapper[4802]: I1206 03:40:48.066346 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:48 crc kubenswrapper[4802]: I1206 03:40:48.066356 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:48 crc kubenswrapper[4802]: I1206 03:40:48.066377 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:48 crc kubenswrapper[4802]: I1206 03:40:48.066388 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:48Z","lastTransitionTime":"2025-12-06T03:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:48 crc kubenswrapper[4802]: I1206 03:40:48.169575 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:48 crc kubenswrapper[4802]: I1206 03:40:48.169631 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:48 crc kubenswrapper[4802]: I1206 03:40:48.169648 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:48 crc kubenswrapper[4802]: I1206 03:40:48.169672 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:48 crc kubenswrapper[4802]: I1206 03:40:48.169689 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:48Z","lastTransitionTime":"2025-12-06T03:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:48 crc kubenswrapper[4802]: I1206 03:40:48.272395 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:48 crc kubenswrapper[4802]: I1206 03:40:48.272449 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:48 crc kubenswrapper[4802]: I1206 03:40:48.272466 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:48 crc kubenswrapper[4802]: I1206 03:40:48.272491 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:48 crc kubenswrapper[4802]: I1206 03:40:48.272508 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:48Z","lastTransitionTime":"2025-12-06T03:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:48 crc kubenswrapper[4802]: I1206 03:40:48.375289 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:48 crc kubenswrapper[4802]: I1206 03:40:48.375359 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:48 crc kubenswrapper[4802]: I1206 03:40:48.375377 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:48 crc kubenswrapper[4802]: I1206 03:40:48.375402 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:48 crc kubenswrapper[4802]: I1206 03:40:48.375421 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:48Z","lastTransitionTime":"2025-12-06T03:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:48 crc kubenswrapper[4802]: I1206 03:40:48.449608 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:40:48 crc kubenswrapper[4802]: I1206 03:40:48.449667 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:40:48 crc kubenswrapper[4802]: I1206 03:40:48.449724 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:40:48 crc kubenswrapper[4802]: I1206 03:40:48.449626 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:40:48 crc kubenswrapper[4802]: E1206 03:40:48.449925 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5ltx2" podUID="8881599e-71e5-4578-b3f9-21f4fa2f57aa" Dec 06 03:40:48 crc kubenswrapper[4802]: E1206 03:40:48.450150 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:40:48 crc kubenswrapper[4802]: E1206 03:40:48.450324 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:40:48 crc kubenswrapper[4802]: E1206 03:40:48.450510 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:40:48 crc kubenswrapper[4802]: I1206 03:40:48.478862 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:48 crc kubenswrapper[4802]: I1206 03:40:48.478939 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:48 crc kubenswrapper[4802]: I1206 03:40:48.478965 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:48 crc kubenswrapper[4802]: I1206 03:40:48.478993 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:48 crc kubenswrapper[4802]: I1206 03:40:48.479013 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:48Z","lastTransitionTime":"2025-12-06T03:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:48 crc kubenswrapper[4802]: I1206 03:40:48.584052 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:48 crc kubenswrapper[4802]: I1206 03:40:48.584316 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:48 crc kubenswrapper[4802]: I1206 03:40:48.584425 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:48 crc kubenswrapper[4802]: I1206 03:40:48.584546 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:48 crc kubenswrapper[4802]: I1206 03:40:48.584658 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:48Z","lastTransitionTime":"2025-12-06T03:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:48 crc kubenswrapper[4802]: I1206 03:40:48.687797 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:48 crc kubenswrapper[4802]: I1206 03:40:48.687832 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:48 crc kubenswrapper[4802]: I1206 03:40:48.687839 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:48 crc kubenswrapper[4802]: I1206 03:40:48.687853 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:48 crc kubenswrapper[4802]: I1206 03:40:48.687862 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:48Z","lastTransitionTime":"2025-12-06T03:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:48 crc kubenswrapper[4802]: I1206 03:40:48.710672 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8881599e-71e5-4578-b3f9-21f4fa2f57aa-metrics-certs\") pod \"network-metrics-daemon-5ltx2\" (UID: \"8881599e-71e5-4578-b3f9-21f4fa2f57aa\") " pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:40:48 crc kubenswrapper[4802]: E1206 03:40:48.710859 4802 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 03:40:48 crc kubenswrapper[4802]: E1206 03:40:48.710914 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8881599e-71e5-4578-b3f9-21f4fa2f57aa-metrics-certs podName:8881599e-71e5-4578-b3f9-21f4fa2f57aa nodeName:}" failed. No retries permitted until 2025-12-06 03:40:56.710900413 +0000 UTC m=+49.582809565 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8881599e-71e5-4578-b3f9-21f4fa2f57aa-metrics-certs") pod "network-metrics-daemon-5ltx2" (UID: "8881599e-71e5-4578-b3f9-21f4fa2f57aa") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 03:40:48 crc kubenswrapper[4802]: I1206 03:40:48.791089 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:48 crc kubenswrapper[4802]: I1206 03:40:48.791171 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:48 crc kubenswrapper[4802]: I1206 03:40:48.791196 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:48 crc kubenswrapper[4802]: I1206 03:40:48.791222 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:48 crc kubenswrapper[4802]: I1206 03:40:48.791242 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:48Z","lastTransitionTime":"2025-12-06T03:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:48 crc kubenswrapper[4802]: I1206 03:40:48.894742 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:48 crc kubenswrapper[4802]: I1206 03:40:48.894860 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:48 crc kubenswrapper[4802]: I1206 03:40:48.894879 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:48 crc kubenswrapper[4802]: I1206 03:40:48.894905 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:48 crc kubenswrapper[4802]: I1206 03:40:48.894923 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:48Z","lastTransitionTime":"2025-12-06T03:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:48 crc kubenswrapper[4802]: I1206 03:40:48.997559 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:48 crc kubenswrapper[4802]: I1206 03:40:48.997621 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:48 crc kubenswrapper[4802]: I1206 03:40:48.997676 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:48 crc kubenswrapper[4802]: I1206 03:40:48.997699 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:48 crc kubenswrapper[4802]: I1206 03:40:48.997717 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:48Z","lastTransitionTime":"2025-12-06T03:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.101291 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.101368 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.101387 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.101410 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.101430 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:49Z","lastTransitionTime":"2025-12-06T03:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.204074 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.204190 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.204215 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.204275 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.204300 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:49Z","lastTransitionTime":"2025-12-06T03:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.231152 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.231252 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.231280 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.231308 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.231337 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:49Z","lastTransitionTime":"2025-12-06T03:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:49 crc kubenswrapper[4802]: E1206 03:40:49.253977 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0e01ca3a-758f-44bc-b4a2-30c1bc04fda0\\\",\\\"systemUUID\\\":\\\"c7c815e4-d2a4-4244-b035-f988fc95e215\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:49Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.259458 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.259517 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.259534 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.259557 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.259574 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:49Z","lastTransitionTime":"2025-12-06T03:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:49 crc kubenswrapper[4802]: E1206 03:40:49.279637 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0e01ca3a-758f-44bc-b4a2-30c1bc04fda0\\\",\\\"systemUUID\\\":\\\"c7c815e4-d2a4-4244-b035-f988fc95e215\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:49Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.285346 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.285421 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.285444 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.285476 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.285502 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:49Z","lastTransitionTime":"2025-12-06T03:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:49 crc kubenswrapper[4802]: E1206 03:40:49.309323 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0e01ca3a-758f-44bc-b4a2-30c1bc04fda0\\\",\\\"systemUUID\\\":\\\"c7c815e4-d2a4-4244-b035-f988fc95e215\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:49Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.315436 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.315497 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.315514 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.315542 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.315560 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:49Z","lastTransitionTime":"2025-12-06T03:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:49 crc kubenswrapper[4802]: E1206 03:40:49.335554 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0e01ca3a-758f-44bc-b4a2-30c1bc04fda0\\\",\\\"systemUUID\\\":\\\"c7c815e4-d2a4-4244-b035-f988fc95e215\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:49Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.340543 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.340586 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.340603 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.340627 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.340646 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:49Z","lastTransitionTime":"2025-12-06T03:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:49 crc kubenswrapper[4802]: E1206 03:40:49.361902 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0e01ca3a-758f-44bc-b4a2-30c1bc04fda0\\\",\\\"systemUUID\\\":\\\"c7c815e4-d2a4-4244-b035-f988fc95e215\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:49Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:49 crc kubenswrapper[4802]: E1206 03:40:49.362185 4802 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.364847 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.364914 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.364937 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.364966 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.364990 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:49Z","lastTransitionTime":"2025-12-06T03:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.467339 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.467383 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.467393 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.467404 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.467413 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:49Z","lastTransitionTime":"2025-12-06T03:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.570593 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.570663 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.570685 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.570714 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.570735 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:49Z","lastTransitionTime":"2025-12-06T03:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.673905 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.674240 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.674444 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.674637 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.674840 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:49Z","lastTransitionTime":"2025-12-06T03:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.777861 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.777899 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.777911 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.777929 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.777943 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:49Z","lastTransitionTime":"2025-12-06T03:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.881510 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.882064 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.882221 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.882360 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.882482 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:49Z","lastTransitionTime":"2025-12-06T03:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.985788 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.985889 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.985913 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.985943 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:49 crc kubenswrapper[4802]: I1206 03:40:49.985965 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:49Z","lastTransitionTime":"2025-12-06T03:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:50 crc kubenswrapper[4802]: I1206 03:40:50.094686 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:50 crc kubenswrapper[4802]: I1206 03:40:50.094773 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:50 crc kubenswrapper[4802]: I1206 03:40:50.094794 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:50 crc kubenswrapper[4802]: I1206 03:40:50.094851 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:50 crc kubenswrapper[4802]: I1206 03:40:50.094870 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:50Z","lastTransitionTime":"2025-12-06T03:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:50 crc kubenswrapper[4802]: I1206 03:40:50.197442 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:50 crc kubenswrapper[4802]: I1206 03:40:50.197485 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:50 crc kubenswrapper[4802]: I1206 03:40:50.197494 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:50 crc kubenswrapper[4802]: I1206 03:40:50.197508 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:50 crc kubenswrapper[4802]: I1206 03:40:50.197517 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:50Z","lastTransitionTime":"2025-12-06T03:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:50 crc kubenswrapper[4802]: I1206 03:40:50.300120 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:50 crc kubenswrapper[4802]: I1206 03:40:50.300168 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:50 crc kubenswrapper[4802]: I1206 03:40:50.300180 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:50 crc kubenswrapper[4802]: I1206 03:40:50.300197 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:50 crc kubenswrapper[4802]: I1206 03:40:50.300211 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:50Z","lastTransitionTime":"2025-12-06T03:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:50 crc kubenswrapper[4802]: I1206 03:40:50.402451 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:50 crc kubenswrapper[4802]: I1206 03:40:50.402486 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:50 crc kubenswrapper[4802]: I1206 03:40:50.402495 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:50 crc kubenswrapper[4802]: I1206 03:40:50.402506 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:50 crc kubenswrapper[4802]: I1206 03:40:50.402514 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:50Z","lastTransitionTime":"2025-12-06T03:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:50 crc kubenswrapper[4802]: I1206 03:40:50.449717 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:40:50 crc kubenswrapper[4802]: I1206 03:40:50.449830 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:40:50 crc kubenswrapper[4802]: I1206 03:40:50.449880 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:40:50 crc kubenswrapper[4802]: E1206 03:40:50.449929 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5ltx2" podUID="8881599e-71e5-4578-b3f9-21f4fa2f57aa" Dec 06 03:40:50 crc kubenswrapper[4802]: E1206 03:40:50.450095 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:40:50 crc kubenswrapper[4802]: I1206 03:40:50.450177 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:40:50 crc kubenswrapper[4802]: E1206 03:40:50.450298 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:40:50 crc kubenswrapper[4802]: E1206 03:40:50.450326 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:40:50 crc kubenswrapper[4802]: I1206 03:40:50.505340 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:50 crc kubenswrapper[4802]: I1206 03:40:50.505387 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:50 crc kubenswrapper[4802]: I1206 03:40:50.505399 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:50 crc kubenswrapper[4802]: I1206 03:40:50.505418 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:50 crc kubenswrapper[4802]: I1206 03:40:50.505432 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:50Z","lastTransitionTime":"2025-12-06T03:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:50 crc kubenswrapper[4802]: I1206 03:40:50.612240 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:50 crc kubenswrapper[4802]: I1206 03:40:50.612298 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:50 crc kubenswrapper[4802]: I1206 03:40:50.612314 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:50 crc kubenswrapper[4802]: I1206 03:40:50.612339 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:50 crc kubenswrapper[4802]: I1206 03:40:50.612356 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:50Z","lastTransitionTime":"2025-12-06T03:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:50 crc kubenswrapper[4802]: I1206 03:40:50.714905 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:50 crc kubenswrapper[4802]: I1206 03:40:50.714960 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:50 crc kubenswrapper[4802]: I1206 03:40:50.714973 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:50 crc kubenswrapper[4802]: I1206 03:40:50.714991 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:50 crc kubenswrapper[4802]: I1206 03:40:50.715003 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:50Z","lastTransitionTime":"2025-12-06T03:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:50 crc kubenswrapper[4802]: I1206 03:40:50.817402 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:50 crc kubenswrapper[4802]: I1206 03:40:50.817463 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:50 crc kubenswrapper[4802]: I1206 03:40:50.817483 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:50 crc kubenswrapper[4802]: I1206 03:40:50.817511 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:50 crc kubenswrapper[4802]: I1206 03:40:50.817556 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:50Z","lastTransitionTime":"2025-12-06T03:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:50 crc kubenswrapper[4802]: I1206 03:40:50.920646 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:50 crc kubenswrapper[4802]: I1206 03:40:50.920709 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:50 crc kubenswrapper[4802]: I1206 03:40:50.920727 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:50 crc kubenswrapper[4802]: I1206 03:40:50.920778 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:50 crc kubenswrapper[4802]: I1206 03:40:50.920797 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:50Z","lastTransitionTime":"2025-12-06T03:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:51 crc kubenswrapper[4802]: I1206 03:40:51.023608 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:51 crc kubenswrapper[4802]: I1206 03:40:51.023688 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:51 crc kubenswrapper[4802]: I1206 03:40:51.023712 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:51 crc kubenswrapper[4802]: I1206 03:40:51.023742 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:51 crc kubenswrapper[4802]: I1206 03:40:51.023873 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:51Z","lastTransitionTime":"2025-12-06T03:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:51 crc kubenswrapper[4802]: I1206 03:40:51.126120 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:51 crc kubenswrapper[4802]: I1206 03:40:51.126178 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:51 crc kubenswrapper[4802]: I1206 03:40:51.126196 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:51 crc kubenswrapper[4802]: I1206 03:40:51.126219 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:51 crc kubenswrapper[4802]: I1206 03:40:51.126236 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:51Z","lastTransitionTime":"2025-12-06T03:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:51 crc kubenswrapper[4802]: I1206 03:40:51.229283 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:51 crc kubenswrapper[4802]: I1206 03:40:51.229367 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:51 crc kubenswrapper[4802]: I1206 03:40:51.229391 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:51 crc kubenswrapper[4802]: I1206 03:40:51.229421 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:51 crc kubenswrapper[4802]: I1206 03:40:51.229443 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:51Z","lastTransitionTime":"2025-12-06T03:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:51 crc kubenswrapper[4802]: I1206 03:40:51.332003 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:51 crc kubenswrapper[4802]: I1206 03:40:51.332083 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:51 crc kubenswrapper[4802]: I1206 03:40:51.332095 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:51 crc kubenswrapper[4802]: I1206 03:40:51.332115 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:51 crc kubenswrapper[4802]: I1206 03:40:51.332127 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:51Z","lastTransitionTime":"2025-12-06T03:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:51 crc kubenswrapper[4802]: I1206 03:40:51.435278 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:51 crc kubenswrapper[4802]: I1206 03:40:51.435325 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:51 crc kubenswrapper[4802]: I1206 03:40:51.435343 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:51 crc kubenswrapper[4802]: I1206 03:40:51.435362 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:51 crc kubenswrapper[4802]: I1206 03:40:51.435376 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:51Z","lastTransitionTime":"2025-12-06T03:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:51 crc kubenswrapper[4802]: I1206 03:40:51.538305 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:51 crc kubenswrapper[4802]: I1206 03:40:51.538356 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:51 crc kubenswrapper[4802]: I1206 03:40:51.538369 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:51 crc kubenswrapper[4802]: I1206 03:40:51.538386 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:51 crc kubenswrapper[4802]: I1206 03:40:51.538404 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:51Z","lastTransitionTime":"2025-12-06T03:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:51 crc kubenswrapper[4802]: I1206 03:40:51.641890 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:51 crc kubenswrapper[4802]: I1206 03:40:51.641931 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:51 crc kubenswrapper[4802]: I1206 03:40:51.641943 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:51 crc kubenswrapper[4802]: I1206 03:40:51.641961 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:51 crc kubenswrapper[4802]: I1206 03:40:51.641975 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:51Z","lastTransitionTime":"2025-12-06T03:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:51 crc kubenswrapper[4802]: I1206 03:40:51.745047 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:51 crc kubenswrapper[4802]: I1206 03:40:51.745106 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:51 crc kubenswrapper[4802]: I1206 03:40:51.745125 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:51 crc kubenswrapper[4802]: I1206 03:40:51.745151 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:51 crc kubenswrapper[4802]: I1206 03:40:51.745167 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:51Z","lastTransitionTime":"2025-12-06T03:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:51 crc kubenswrapper[4802]: I1206 03:40:51.848090 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:51 crc kubenswrapper[4802]: I1206 03:40:51.848146 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:51 crc kubenswrapper[4802]: I1206 03:40:51.848163 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:51 crc kubenswrapper[4802]: I1206 03:40:51.848188 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:51 crc kubenswrapper[4802]: I1206 03:40:51.848207 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:51Z","lastTransitionTime":"2025-12-06T03:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:51 crc kubenswrapper[4802]: I1206 03:40:51.950441 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:51 crc kubenswrapper[4802]: I1206 03:40:51.950493 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:51 crc kubenswrapper[4802]: I1206 03:40:51.950509 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:51 crc kubenswrapper[4802]: I1206 03:40:51.950527 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:51 crc kubenswrapper[4802]: I1206 03:40:51.950539 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:51Z","lastTransitionTime":"2025-12-06T03:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:52 crc kubenswrapper[4802]: I1206 03:40:52.053972 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:52 crc kubenswrapper[4802]: I1206 03:40:52.054017 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:52 crc kubenswrapper[4802]: I1206 03:40:52.054031 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:52 crc kubenswrapper[4802]: I1206 03:40:52.054048 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:52 crc kubenswrapper[4802]: I1206 03:40:52.054060 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:52Z","lastTransitionTime":"2025-12-06T03:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:52 crc kubenswrapper[4802]: I1206 03:40:52.157996 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:52 crc kubenswrapper[4802]: I1206 03:40:52.158988 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:52 crc kubenswrapper[4802]: I1206 03:40:52.159066 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:52 crc kubenswrapper[4802]: I1206 03:40:52.159117 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:52 crc kubenswrapper[4802]: I1206 03:40:52.159143 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:52Z","lastTransitionTime":"2025-12-06T03:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:52 crc kubenswrapper[4802]: I1206 03:40:52.261962 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:52 crc kubenswrapper[4802]: I1206 03:40:52.262007 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:52 crc kubenswrapper[4802]: I1206 03:40:52.262018 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:52 crc kubenswrapper[4802]: I1206 03:40:52.262033 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:52 crc kubenswrapper[4802]: I1206 03:40:52.262045 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:52Z","lastTransitionTime":"2025-12-06T03:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:52 crc kubenswrapper[4802]: I1206 03:40:52.365255 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:52 crc kubenswrapper[4802]: I1206 03:40:52.365316 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:52 crc kubenswrapper[4802]: I1206 03:40:52.365334 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:52 crc kubenswrapper[4802]: I1206 03:40:52.365358 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:52 crc kubenswrapper[4802]: I1206 03:40:52.365375 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:52Z","lastTransitionTime":"2025-12-06T03:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:52 crc kubenswrapper[4802]: I1206 03:40:52.449990 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:40:52 crc kubenswrapper[4802]: I1206 03:40:52.450013 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:40:52 crc kubenswrapper[4802]: I1206 03:40:52.450107 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:40:52 crc kubenswrapper[4802]: I1206 03:40:52.450201 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:40:52 crc kubenswrapper[4802]: E1206 03:40:52.450402 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:40:52 crc kubenswrapper[4802]: E1206 03:40:52.450551 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:40:52 crc kubenswrapper[4802]: E1206 03:40:52.450747 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5ltx2" podUID="8881599e-71e5-4578-b3f9-21f4fa2f57aa" Dec 06 03:40:52 crc kubenswrapper[4802]: E1206 03:40:52.450991 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:40:52 crc kubenswrapper[4802]: I1206 03:40:52.467788 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:52 crc kubenswrapper[4802]: I1206 03:40:52.467853 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:52 crc kubenswrapper[4802]: I1206 03:40:52.467873 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:52 crc kubenswrapper[4802]: I1206 03:40:52.467897 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:52 crc kubenswrapper[4802]: I1206 03:40:52.467916 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:52Z","lastTransitionTime":"2025-12-06T03:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:52 crc kubenswrapper[4802]: I1206 03:40:52.570656 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:52 crc kubenswrapper[4802]: I1206 03:40:52.570740 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:52 crc kubenswrapper[4802]: I1206 03:40:52.570795 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:52 crc kubenswrapper[4802]: I1206 03:40:52.570822 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:52 crc kubenswrapper[4802]: I1206 03:40:52.570840 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:52Z","lastTransitionTime":"2025-12-06T03:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:52 crc kubenswrapper[4802]: I1206 03:40:52.673663 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:52 crc kubenswrapper[4802]: I1206 03:40:52.673736 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:52 crc kubenswrapper[4802]: I1206 03:40:52.673791 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:52 crc kubenswrapper[4802]: I1206 03:40:52.673829 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:52 crc kubenswrapper[4802]: I1206 03:40:52.673847 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:52Z","lastTransitionTime":"2025-12-06T03:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:52 crc kubenswrapper[4802]: I1206 03:40:52.777336 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:52 crc kubenswrapper[4802]: I1206 03:40:52.777743 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:52 crc kubenswrapper[4802]: I1206 03:40:52.777937 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:52 crc kubenswrapper[4802]: I1206 03:40:52.778117 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:52 crc kubenswrapper[4802]: I1206 03:40:52.778260 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:52Z","lastTransitionTime":"2025-12-06T03:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:52 crc kubenswrapper[4802]: I1206 03:40:52.882440 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:52 crc kubenswrapper[4802]: I1206 03:40:52.882951 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:52 crc kubenswrapper[4802]: I1206 03:40:52.883180 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:52 crc kubenswrapper[4802]: I1206 03:40:52.883381 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:52 crc kubenswrapper[4802]: I1206 03:40:52.883565 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:52Z","lastTransitionTime":"2025-12-06T03:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:52 crc kubenswrapper[4802]: I1206 03:40:52.986963 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:52 crc kubenswrapper[4802]: I1206 03:40:52.987053 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:52 crc kubenswrapper[4802]: I1206 03:40:52.987072 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:52 crc kubenswrapper[4802]: I1206 03:40:52.987100 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:52 crc kubenswrapper[4802]: I1206 03:40:52.987119 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:52Z","lastTransitionTime":"2025-12-06T03:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.089512 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.089572 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.089588 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.089609 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.089653 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:53Z","lastTransitionTime":"2025-12-06T03:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.192479 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.192517 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.192528 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.192544 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.192555 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:53Z","lastTransitionTime":"2025-12-06T03:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.295622 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.295661 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.295672 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.295686 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.295696 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:53Z","lastTransitionTime":"2025-12-06T03:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.397786 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.397828 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.397838 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.397854 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.397895 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:53Z","lastTransitionTime":"2025-12-06T03:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.451262 4802 scope.go:117] "RemoveContainer" containerID="cf845e8893e51078e46991489d9a27c1194eab62a9d0a24eb3336d396e12cdfb" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.500509 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.500829 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.500845 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.500864 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.500875 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:53Z","lastTransitionTime":"2025-12-06T03:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.518028 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.540515 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fad9b63d84f41fccf696cd313d7ca7f72b3d3251d6c1b7341573a3fadf1e378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:53Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.558621 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb3f278fff47c900c56d8a563d5ed75a2ae84e2034b0f45aad4083790b9e7946\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b10ca16ac26904ec5e6245344b71d836888cae9b069cae6a7a4831d609752db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:53Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.578545 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2093b6-c4b5-4db8-be16-fbc859885e8d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0d6e582161f0d9126126e03b27211be22f6b5f797154928c92fa13a8118f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03004a29aefe7fe2452fb9f6663c0807afccb055d7ddbf6b907f305e6a19c81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1801ac64db4ba91c2e8a249e3651d3e50ab62e9f2c87312ce9bf7f38518a82cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6e454507cadbb982a566a3692995aaa3c733ff1cd7fb4870476602190e1eeba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:53Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.600457 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:53Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.603371 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.603429 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.603448 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.603471 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.603488 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:53Z","lastTransitionTime":"2025-12-06T03:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.614535 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:53Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.640003 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtw2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3a3bac2-3923-4507-a4de-1556b3ad473e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fed5ee4189d54456a3216f4fdb1ecf003fa47e29dbfe4b2b2f301edfb897d6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtw2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:53Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.668308 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7730de4f033561ad77046e028cd1a3f0bfe2775387cbdfa35f47ea79909ec4cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a158aebd701ef46225d3e37109ae28ab58d107087bf06650c4b6d0055a3bbba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2357f37497fde64937eca7b38693893af745ab278653ebc79406e122a6d4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8941bd25107333b264d3f081a811fdfeea23ad92193fc61367665fd2cd1209c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f9cd4dc75e63f09f60b48b6da284643910bb32ba51f2a808cb4d9803f30042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6796f65ee9c5bd0c18f205697135d17677c5daf3ba378cc992cba795e0188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf845e8893e51078e46991489d9a27c1194eab62a9d0a24eb3336d396e12cdfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf845e8893e51078e46991489d9a27c1194eab62a9d0a24eb3336d396e12cdfb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:40:38Z\\\",\\\"message\\\":\\\"Map:map[10.217.4.110:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {f9232b32-e89f-4c8e-acc4-c6801b70dcb0}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:NB_Global Row:map[] Rows:[] Columns:[] Mutations:[{Column:nb_cfg Mutator:+= Value:1}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {6011affd-30a6-4be6-872d-e4cf1ca780cf}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 03:40:38.703973 6236 ovnkube.go:599] Stopped ovnkube\\\\nI1206 03:40:38.704014 6236 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1206 03:40:38.703921 6236 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-controller]} name:Service_openshift-machine-config-operator/machine-config-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.16:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {3f1b9878-e751-4e46-a226-ce007d2c4aa7}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1206 03:40:38.704093 6236 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-pvm6q_openshift-ovn-kubernetes(eafce4e0-e7fb-4877-b0ab-3283829ba1ef)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3cb210bfc97a439a7ff98c02c59f8ad16bc400dc214a616a44b43a8b01e2782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pvm6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:53Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.683322 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-b2knv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a3ae46f-03b5-4999-9d58-c38ee222f68a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76665d4c127445009fd66078827b54607c9ef812c8011d1c0b4cde1c0516853a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk4hj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-b2knv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:53Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.702479 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22rrq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03112a5d-5882-4fac-9ae5-13525ff82fe0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0205b089b12c42688203bb7ec0f27dc5f9a577dabdb446b811f184046f85dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trlt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22rrq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:53Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.706566 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.706631 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.706655 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.706687 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.706709 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:53Z","lastTransitionTime":"2025-12-06T03:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.716719 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ngspk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34808ab2-d21b-40b6-8fe3-e18bf9e5fc46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52509d21f5b4bfbd458162cba176ff82c627c7c4671204addf9d76f7392fd760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qkf8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9685a4d5d4f40ff4f59b6da6193785f340890453d1e422d3421a621dc7bd02f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qkf8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ngspk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:53Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.730738 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5ltx2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8881599e-71e5-4578-b3f9-21f4fa2f57aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbm6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbm6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5ltx2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:53Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.751949 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1721e985-40bd-4b2d-a0e8-d4e365094ba6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a895e0cc04880c3287ba725ac5484835235f1e066a94ce98fcf4b6133c5cbd2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a21a97385f856ec996105b597d0ae81c56ed28b625ef3f2905e1af6ec8f6a45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://646e01a58908ba2e703aae5eef53e1b981b0f0c3d1d3202487f2a8b94ee3f55d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31752b84d86bda30792b29e4c6ed963753cc576c5de784395e8e6dcbebaced47\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"ver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:40:26.813730 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814130 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:40:26.814139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:40:26.814143 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:40:26.814145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 03:40:26.821403 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821442 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821482 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821496 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 03:40:26.821509 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821514 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 03:40:26.821612 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 03:40:26.821619 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1206 03:40:26.822049 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16a92d501c9e626fe91e4a6793e426d10eb292cfbab766db8b8cc72e5dae94ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:53Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.768009 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kj7hf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd46dfa5-7d4a-49ed-8199-523f10cebdf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5eee8a0cf96ceb3785dac9df31bf8d31715d1c88fbea17c5e582b0a4554bb63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmkcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kj7hf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:53Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.786693 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2466610962332c6ca120fe055e4e8a3fc623088119dc1a412e06b745176f89e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:53Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.810447 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.810492 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.810511 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.810535 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.810555 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:53Z","lastTransitionTime":"2025-12-06T03:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.813901 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:53Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.819518 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pvm6q_eafce4e0-e7fb-4877-b0ab-3283829ba1ef/ovnkube-controller/1.log" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.824925 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" event={"ID":"eafce4e0-e7fb-4877-b0ab-3283829ba1ef","Type":"ContainerStarted","Data":"3870b61bce487c69724f668c38fc02190ff97b5e3b1434d66ed624412d18319b"} Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.825405 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.835975 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6df38316-e0d3-4018-8d27-3620eba3a68d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57381e9589a2dc04a4d2733c37985f42594182128a374b7be0f471ea69138b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bf5e41701c7a2d45adc4740ceb4ac1785db17bc7cb15daed439f7556812d1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpxxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:53Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.866523 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7730de4f033561ad77046e028cd1a3f0bfe2775387cbdfa35f47ea79909ec4cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a158aebd701ef46225d3e37109ae28ab58d107087bf06650c4b6d0055a3bbba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2357f37497fde64937eca7b38693893af745ab278653ebc79406e122a6d4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8941bd25107333b264d3f081a811fdfeea23ad92193fc61367665fd2cd1209c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f9cd4dc75e63f09f60b48b6da284643910bb32ba51f2a808cb4d9803f30042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6796f65ee9c5bd0c18f205697135d17677c5daf3ba378cc992cba795e0188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3870b61bce487c69724f668c38fc02190ff97b5e3b1434d66ed624412d18319b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf845e8893e51078e46991489d9a27c1194eab62a9d0a24eb3336d396e12cdfb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:40:38Z\\\",\\\"message\\\":\\\"Map:map[10.217.4.110:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {f9232b32-e89f-4c8e-acc4-c6801b70dcb0}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:NB_Global Row:map[] Rows:[] Columns:[] Mutations:[{Column:nb_cfg Mutator:+= Value:1}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {6011affd-30a6-4be6-872d-e4cf1ca780cf}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 03:40:38.703973 6236 ovnkube.go:599] Stopped ovnkube\\\\nI1206 03:40:38.704014 6236 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1206 03:40:38.703921 6236 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-controller]} name:Service_openshift-machine-config-operator/machine-config-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.16:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {3f1b9878-e751-4e46-a226-ce007d2c4aa7}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1206 03:40:38.704093 6236 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3cb210bfc97a439a7ff98c02c59f8ad16bc400dc214a616a44b43a8b01e2782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pvm6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:53Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.883463 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:53Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.898336 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtw2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3a3bac2-3923-4507-a4de-1556b3ad473e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fed5ee4189d54456a3216f4fdb1ecf003fa47e29dbfe4b2b2f301edfb897d6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtw2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:53Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.909978 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-b2knv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a3ae46f-03b5-4999-9d58-c38ee222f68a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76665d4c127445009fd66078827b54607c9ef812c8011d1c0b4cde1c0516853a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk4hj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-b2knv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:53Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.913143 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.913177 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.913187 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.913203 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.913215 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:53Z","lastTransitionTime":"2025-12-06T03:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.923409 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22rrq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03112a5d-5882-4fac-9ae5-13525ff82fe0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0205b089b12c42688203bb7ec0f27dc5f9a577dabdb446b811f184046f85dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trlt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22rrq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:53Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.935376 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ngspk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34808ab2-d21b-40b6-8fe3-e18bf9e5fc46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52509d21f5b4bfbd458162cba176ff82c627c7c4671204addf9d76f7392fd760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qkf8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9685a4d5d4f40ff4f59b6da6193785f340890453d1e422d3421a621dc7bd02f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qkf8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ngspk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:53Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.948978 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5ltx2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8881599e-71e5-4578-b3f9-21f4fa2f57aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbm6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbm6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5ltx2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:53Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.965794 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1721e985-40bd-4b2d-a0e8-d4e365094ba6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a895e0cc04880c3287ba725ac5484835235f1e066a94ce98fcf4b6133c5cbd2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a21a97385f856ec996105b597d0ae81c56ed28b625ef3f2905e1af6ec8f6a45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://646e01a58908ba2e703aae5eef53e1b981b0f0c3d1d3202487f2a8b94ee3f55d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31752b84d86bda30792b29e4c6ed963753cc576c5de784395e8e6dcbebaced47\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"ver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:40:26.813730 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814130 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:40:26.814139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:40:26.814143 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:40:26.814145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 03:40:26.821403 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821442 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821482 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821496 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 03:40:26.821509 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821514 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 03:40:26.821612 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 03:40:26.821619 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1206 03:40:26.822049 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16a92d501c9e626fe91e4a6793e426d10eb292cfbab766db8b8cc72e5dae94ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:53Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.979646 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6df38316-e0d3-4018-8d27-3620eba3a68d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57381e9589a2dc04a4d2733c37985f42594182128a374b7be0f471ea69138b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bf5e41701c7a2d45adc4740ceb4ac1785db17bc7cb15daed439f7556812d1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpxxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:53Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:53 crc kubenswrapper[4802]: I1206 03:40:53.996057 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kj7hf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd46dfa5-7d4a-49ed-8199-523f10cebdf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5eee8a0cf96ceb3785dac9df31bf8d31715d1c88fbea17c5e582b0a4554bb63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmkcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kj7hf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:53Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.014297 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2466610962332c6ca120fe055e4e8a3fc623088119dc1a412e06b745176f89e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:54Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.015616 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.015660 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.015676 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.015698 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.015716 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:54Z","lastTransitionTime":"2025-12-06T03:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.031480 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:54Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.051695 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:54Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.071375 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fad9b63d84f41fccf696cd313d7ca7f72b3d3251d6c1b7341573a3fadf1e378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:54Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.089539 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb3f278fff47c900c56d8a563d5ed75a2ae84e2034b0f45aad4083790b9e7946\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b10ca16ac26904ec5e6245344b71d836888cae9b069cae6a7a4831d609752db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:54Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.118047 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2093b6-c4b5-4db8-be16-fbc859885e8d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0d6e582161f0d9126126e03b27211be22f6b5f797154928c92fa13a8118f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03004a29aefe7fe2452fb9f6663c0807afccb055d7ddbf6b907f305e6a19c81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1801ac64db4ba91c2e8a249e3651d3e50ab62e9f2c87312ce9bf7f38518a82cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6e454507cadbb982a566a3692995aaa3c733ff1cd7fb4870476602190e1eeba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:54Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.118128 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.118179 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.118200 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.118229 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.118252 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:54Z","lastTransitionTime":"2025-12-06T03:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.220416 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.220496 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.220506 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.220530 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.220539 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:54Z","lastTransitionTime":"2025-12-06T03:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.322261 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.322301 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.322315 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.322335 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.322349 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:54Z","lastTransitionTime":"2025-12-06T03:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.424657 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.424720 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.424736 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.424789 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.424809 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:54Z","lastTransitionTime":"2025-12-06T03:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.449380 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.449437 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.449483 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.449517 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:40:54 crc kubenswrapper[4802]: E1206 03:40:54.449652 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:40:54 crc kubenswrapper[4802]: E1206 03:40:54.449814 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:40:54 crc kubenswrapper[4802]: E1206 03:40:54.449932 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:40:54 crc kubenswrapper[4802]: E1206 03:40:54.450024 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5ltx2" podUID="8881599e-71e5-4578-b3f9-21f4fa2f57aa" Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.527158 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.527187 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.527194 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.527206 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.527215 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:54Z","lastTransitionTime":"2025-12-06T03:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.633110 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.633162 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.633174 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.633191 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.633207 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:54Z","lastTransitionTime":"2025-12-06T03:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.737150 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.737213 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.737230 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.737258 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.737276 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:54Z","lastTransitionTime":"2025-12-06T03:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.840175 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.840256 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.840288 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.840316 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.840337 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:54Z","lastTransitionTime":"2025-12-06T03:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.943637 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.943682 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.943697 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.943720 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:54 crc kubenswrapper[4802]: I1206 03:40:54.943737 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:54Z","lastTransitionTime":"2025-12-06T03:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.047402 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.047463 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.047480 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.047504 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.047523 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:55Z","lastTransitionTime":"2025-12-06T03:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.150852 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.150906 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.150922 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.150945 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.150962 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:55Z","lastTransitionTime":"2025-12-06T03:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.254244 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.254290 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.254300 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.254319 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.254332 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:55Z","lastTransitionTime":"2025-12-06T03:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.357741 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.357831 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.357848 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.357873 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.357893 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:55Z","lastTransitionTime":"2025-12-06T03:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.460980 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.461042 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.461061 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.461084 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.461103 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:55Z","lastTransitionTime":"2025-12-06T03:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.563398 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.563442 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.563454 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.563469 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.563481 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:55Z","lastTransitionTime":"2025-12-06T03:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.667015 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.667078 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.667091 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.667111 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.667125 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:55Z","lastTransitionTime":"2025-12-06T03:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.770143 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.770213 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.770232 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.770258 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.770275 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:55Z","lastTransitionTime":"2025-12-06T03:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.835497 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pvm6q_eafce4e0-e7fb-4877-b0ab-3283829ba1ef/ovnkube-controller/2.log" Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.836683 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pvm6q_eafce4e0-e7fb-4877-b0ab-3283829ba1ef/ovnkube-controller/1.log" Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.840647 4802 generic.go:334] "Generic (PLEG): container finished" podID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerID="3870b61bce487c69724f668c38fc02190ff97b5e3b1434d66ed624412d18319b" exitCode=1 Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.840721 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" event={"ID":"eafce4e0-e7fb-4877-b0ab-3283829ba1ef","Type":"ContainerDied","Data":"3870b61bce487c69724f668c38fc02190ff97b5e3b1434d66ed624412d18319b"} Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.840829 4802 scope.go:117] "RemoveContainer" containerID="cf845e8893e51078e46991489d9a27c1194eab62a9d0a24eb3336d396e12cdfb" Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.843141 4802 scope.go:117] "RemoveContainer" containerID="3870b61bce487c69724f668c38fc02190ff97b5e3b1434d66ed624412d18319b" Dec 06 03:40:55 crc kubenswrapper[4802]: E1206 03:40:55.843413 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pvm6q_openshift-ovn-kubernetes(eafce4e0-e7fb-4877-b0ab-3283829ba1ef)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" podUID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.870134 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2093b6-c4b5-4db8-be16-fbc859885e8d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0d6e582161f0d9126126e03b27211be22f6b5f797154928c92fa13a8118f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03004a29aefe7fe2452fb9f6663c0807afccb055d7ddbf6b907f305e6a19c81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1801ac64db4ba91c2e8a249e3651d3e50ab62e9f2c87312ce9bf7f38518a82cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6e454507cadbb982a566a3692995aaa3c733ff1cd7fb4870476602190e1eeba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:55Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.874023 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.874079 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.874099 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.874124 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.874141 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:55Z","lastTransitionTime":"2025-12-06T03:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.890233 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:55Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.908897 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fad9b63d84f41fccf696cd313d7ca7f72b3d3251d6c1b7341573a3fadf1e378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:55Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.930315 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb3f278fff47c900c56d8a563d5ed75a2ae84e2034b0f45aad4083790b9e7946\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b10ca16ac26904ec5e6245344b71d836888cae9b069cae6a7a4831d609752db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:55Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.949229 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:55Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.977527 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtw2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3a3bac2-3923-4507-a4de-1556b3ad473e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fed5ee4189d54456a3216f4fdb1ecf003fa47e29dbfe4b2b2f301edfb897d6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtw2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:55Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.978173 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.978234 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.978252 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.978278 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:55 crc kubenswrapper[4802]: I1206 03:40:55.978296 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:55Z","lastTransitionTime":"2025-12-06T03:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.010892 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7730de4f033561ad77046e028cd1a3f0bfe2775387cbdfa35f47ea79909ec4cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a158aebd701ef46225d3e37109ae28ab58d107087bf06650c4b6d0055a3bbba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2357f37497fde64937eca7b38693893af745ab278653ebc79406e122a6d4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8941bd25107333b264d3f081a811fdfeea23ad92193fc61367665fd2cd1209c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f9cd4dc75e63f09f60b48b6da284643910bb32ba51f2a808cb4d9803f30042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6796f65ee9c5bd0c18f205697135d17677c5daf3ba378cc992cba795e0188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3870b61bce487c69724f668c38fc02190ff97b5e3b1434d66ed624412d18319b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf845e8893e51078e46991489d9a27c1194eab62a9d0a24eb3336d396e12cdfb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:40:38Z\\\",\\\"message\\\":\\\"Map:map[10.217.4.110:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {f9232b32-e89f-4c8e-acc4-c6801b70dcb0}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:NB_Global Row:map[] Rows:[] Columns:[] Mutations:[{Column:nb_cfg Mutator:+= Value:1}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {6011affd-30a6-4be6-872d-e4cf1ca780cf}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 03:40:38.703973 6236 ovnkube.go:599] Stopped ovnkube\\\\nI1206 03:40:38.704014 6236 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1206 03:40:38.703921 6236 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-controller]} name:Service_openshift-machine-config-operator/machine-config-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.16:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {3f1b9878-e751-4e46-a226-ce007d2c4aa7}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1206 03:40:38.704093 6236 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3870b61bce487c69724f668c38fc02190ff97b5e3b1434d66ed624412d18319b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:40:54Z\\\",\\\"message\\\":\\\"/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 03:40:54.580644 6463 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 03:40:54.580730 6463 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 03:40:54.580797 6463 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 03:40:54.580836 6463 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 03:40:54.580845 6463 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 03:40:54.580895 6463 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1206 03:40:54.580943 6463 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1206 03:40:54.580950 6463 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1206 03:40:54.581001 6463 factory.go:656] Stopping watch factory\\\\nI1206 03:40:54.581036 6463 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 03:40:54.581050 6463 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 03:40:54.581063 6463 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 03:40:54.581074 6463 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 03:40:54.581085 6463 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1206 03:40:54.581102 6463 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3cb210bfc97a439a7ff98c02c59f8ad16bc400dc214a616a44b43a8b01e2782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pvm6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:56Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.034567 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1721e985-40bd-4b2d-a0e8-d4e365094ba6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a895e0cc04880c3287ba725ac5484835235f1e066a94ce98fcf4b6133c5cbd2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a21a97385f856ec996105b597d0ae81c56ed28b625ef3f2905e1af6ec8f6a45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://646e01a58908ba2e703aae5eef53e1b981b0f0c3d1d3202487f2a8b94ee3f55d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31752b84d86bda30792b29e4c6ed963753cc576c5de784395e8e6dcbebaced47\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"ver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:40:26.813730 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814130 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:40:26.814139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:40:26.814143 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:40:26.814145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 03:40:26.821403 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821442 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821482 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821496 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 03:40:26.821509 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821514 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 03:40:26.821612 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 03:40:26.821619 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1206 03:40:26.822049 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16a92d501c9e626fe91e4a6793e426d10eb292cfbab766db8b8cc72e5dae94ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:56Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.054300 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-b2knv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a3ae46f-03b5-4999-9d58-c38ee222f68a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76665d4c127445009fd66078827b54607c9ef812c8011d1c0b4cde1c0516853a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk4hj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-b2knv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:56Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.077803 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22rrq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03112a5d-5882-4fac-9ae5-13525ff82fe0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0205b089b12c42688203bb7ec0f27dc5f9a577dabdb446b811f184046f85dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trlt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22rrq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:56Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.081091 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.081169 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.081190 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.081221 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.081243 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:56Z","lastTransitionTime":"2025-12-06T03:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.097440 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ngspk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34808ab2-d21b-40b6-8fe3-e18bf9e5fc46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52509d21f5b4bfbd458162cba176ff82c627c7c4671204addf9d76f7392fd760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qkf8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9685a4d5d4f40ff4f59b6da6193785f340890453d1e422d3421a621dc7bd02f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qkf8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ngspk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:56Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.115108 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5ltx2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8881599e-71e5-4578-b3f9-21f4fa2f57aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbm6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbm6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5ltx2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:56Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.144637 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2466610962332c6ca120fe055e4e8a3fc623088119dc1a412e06b745176f89e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:56Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.170839 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:56Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.184701 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.184800 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.184828 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.184860 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.184879 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:56Z","lastTransitionTime":"2025-12-06T03:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.192376 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6df38316-e0d3-4018-8d27-3620eba3a68d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57381e9589a2dc04a4d2733c37985f42594182128a374b7be0f471ea69138b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bf5e41701c7a2d45adc4740ceb4ac1785db17bc7cb15daed439f7556812d1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpxxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:56Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.209416 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kj7hf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd46dfa5-7d4a-49ed-8199-523f10cebdf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5eee8a0cf96ceb3785dac9df31bf8d31715d1c88fbea17c5e582b0a4554bb63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmkcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kj7hf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:56Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.288634 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.288675 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.288685 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.288701 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.288713 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:56Z","lastTransitionTime":"2025-12-06T03:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.390907 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.390955 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.390970 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.390987 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.390998 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:56Z","lastTransitionTime":"2025-12-06T03:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.449564 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.449598 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.449576 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.449654 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:40:56 crc kubenswrapper[4802]: E1206 03:40:56.449918 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:40:56 crc kubenswrapper[4802]: E1206 03:40:56.450050 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5ltx2" podUID="8881599e-71e5-4578-b3f9-21f4fa2f57aa" Dec 06 03:40:56 crc kubenswrapper[4802]: E1206 03:40:56.450128 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:40:56 crc kubenswrapper[4802]: E1206 03:40:56.450269 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.494167 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.494234 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.494250 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.494276 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.494293 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:56Z","lastTransitionTime":"2025-12-06T03:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.597356 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.597442 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.597467 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.597499 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.597530 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:56Z","lastTransitionTime":"2025-12-06T03:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.700869 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.700925 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.700943 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.700967 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.700986 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:56Z","lastTransitionTime":"2025-12-06T03:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.803732 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.803858 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.803882 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.803913 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.803936 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:56Z","lastTransitionTime":"2025-12-06T03:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.807366 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8881599e-71e5-4578-b3f9-21f4fa2f57aa-metrics-certs\") pod \"network-metrics-daemon-5ltx2\" (UID: \"8881599e-71e5-4578-b3f9-21f4fa2f57aa\") " pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:40:56 crc kubenswrapper[4802]: E1206 03:40:56.807545 4802 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 03:40:56 crc kubenswrapper[4802]: E1206 03:40:56.807645 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8881599e-71e5-4578-b3f9-21f4fa2f57aa-metrics-certs podName:8881599e-71e5-4578-b3f9-21f4fa2f57aa nodeName:}" failed. No retries permitted until 2025-12-06 03:41:12.807617699 +0000 UTC m=+65.679526881 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8881599e-71e5-4578-b3f9-21f4fa2f57aa-metrics-certs") pod "network-metrics-daemon-5ltx2" (UID: "8881599e-71e5-4578-b3f9-21f4fa2f57aa") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.847422 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pvm6q_eafce4e0-e7fb-4877-b0ab-3283829ba1ef/ovnkube-controller/2.log" Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.906422 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.906480 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.906501 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.906527 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:56 crc kubenswrapper[4802]: I1206 03:40:56.906546 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:56Z","lastTransitionTime":"2025-12-06T03:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.009991 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.010074 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.010100 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.010162 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.010191 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:57Z","lastTransitionTime":"2025-12-06T03:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.112780 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.112822 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.112831 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.112845 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.112855 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:57Z","lastTransitionTime":"2025-12-06T03:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.215651 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.215731 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.215777 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.215824 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.215856 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:57Z","lastTransitionTime":"2025-12-06T03:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.318841 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.318908 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.318927 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.318952 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.318969 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:57Z","lastTransitionTime":"2025-12-06T03:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.421096 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.421143 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.421170 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.421189 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.421201 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:57Z","lastTransitionTime":"2025-12-06T03:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.467843 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kj7hf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd46dfa5-7d4a-49ed-8199-523f10cebdf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5eee8a0cf96ceb3785dac9df31bf8d31715d1c88fbea17c5e582b0a4554bb63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmkcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kj7hf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:57Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.484436 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2466610962332c6ca120fe055e4e8a3fc623088119dc1a412e06b745176f89e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:57Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.500690 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:57Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.518389 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6df38316-e0d3-4018-8d27-3620eba3a68d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57381e9589a2dc04a4d2733c37985f42594182128a374b7be0f471ea69138b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bf5e41701c7a2d45adc4740ceb4ac1785db17bc7cb15daed439f7556812d1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpxxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:57Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.523311 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.523351 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.523370 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.523391 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.523405 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:57Z","lastTransitionTime":"2025-12-06T03:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.538290 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fad9b63d84f41fccf696cd313d7ca7f72b3d3251d6c1b7341573a3fadf1e378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:57Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.552226 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb3f278fff47c900c56d8a563d5ed75a2ae84e2034b0f45aad4083790b9e7946\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b10ca16ac26904ec5e6245344b71d836888cae9b069cae6a7a4831d609752db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:57Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.565117 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2093b6-c4b5-4db8-be16-fbc859885e8d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0d6e582161f0d9126126e03b27211be22f6b5f797154928c92fa13a8118f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03004a29aefe7fe2452fb9f6663c0807afccb055d7ddbf6b907f305e6a19c81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1801ac64db4ba91c2e8a249e3651d3e50ab62e9f2c87312ce9bf7f38518a82cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6e454507cadbb982a566a3692995aaa3c733ff1cd7fb4870476602190e1eeba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:57Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.576459 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:57Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.589578 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:57Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.608064 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtw2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3a3bac2-3923-4507-a4de-1556b3ad473e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fed5ee4189d54456a3216f4fdb1ecf003fa47e29dbfe4b2b2f301edfb897d6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtw2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:57Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.642085 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.642168 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.642194 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.642221 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.642238 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:57Z","lastTransitionTime":"2025-12-06T03:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.669040 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7730de4f033561ad77046e028cd1a3f0bfe2775387cbdfa35f47ea79909ec4cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a158aebd701ef46225d3e37109ae28ab58d107087bf06650c4b6d0055a3bbba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2357f37497fde64937eca7b38693893af745ab278653ebc79406e122a6d4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8941bd25107333b264d3f081a811fdfeea23ad92193fc61367665fd2cd1209c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f9cd4dc75e63f09f60b48b6da284643910bb32ba51f2a808cb4d9803f30042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6796f65ee9c5bd0c18f205697135d17677c5daf3ba378cc992cba795e0188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3870b61bce487c69724f668c38fc02190ff97b5e3b1434d66ed624412d18319b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf845e8893e51078e46991489d9a27c1194eab62a9d0a24eb3336d396e12cdfb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:40:38Z\\\",\\\"message\\\":\\\"Map:map[10.217.4.110:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {f9232b32-e89f-4c8e-acc4-c6801b70dcb0}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:NB_Global Row:map[] Rows:[] Columns:[] Mutations:[{Column:nb_cfg Mutator:+= Value:1}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {6011affd-30a6-4be6-872d-e4cf1ca780cf}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 03:40:38.703973 6236 ovnkube.go:599] Stopped ovnkube\\\\nI1206 03:40:38.704014 6236 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1206 03:40:38.703921 6236 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-controller]} name:Service_openshift-machine-config-operator/machine-config-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.16:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {3f1b9878-e751-4e46-a226-ce007d2c4aa7}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1206 03:40:38.704093 6236 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3870b61bce487c69724f668c38fc02190ff97b5e3b1434d66ed624412d18319b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:40:54Z\\\",\\\"message\\\":\\\"/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 03:40:54.580644 6463 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 03:40:54.580730 6463 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 03:40:54.580797 6463 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 03:40:54.580836 6463 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 03:40:54.580845 6463 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 03:40:54.580895 6463 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1206 03:40:54.580943 6463 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1206 03:40:54.580950 6463 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1206 03:40:54.581001 6463 factory.go:656] Stopping watch factory\\\\nI1206 03:40:54.581036 6463 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 03:40:54.581050 6463 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 03:40:54.581063 6463 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 03:40:54.581074 6463 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 03:40:54.581085 6463 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1206 03:40:54.581102 6463 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3cb210bfc97a439a7ff98c02c59f8ad16bc400dc214a616a44b43a8b01e2782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pvm6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:57Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.683974 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-b2knv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a3ae46f-03b5-4999-9d58-c38ee222f68a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76665d4c127445009fd66078827b54607c9ef812c8011d1c0b4cde1c0516853a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk4hj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-b2knv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:57Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.711856 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22rrq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03112a5d-5882-4fac-9ae5-13525ff82fe0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0205b089b12c42688203bb7ec0f27dc5f9a577dabdb446b811f184046f85dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trlt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22rrq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:57Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.723456 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ngspk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34808ab2-d21b-40b6-8fe3-e18bf9e5fc46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52509d21f5b4bfbd458162cba176ff82c627c7c4671204addf9d76f7392fd760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qkf8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9685a4d5d4f40ff4f59b6da6193785f340890453d1e422d3421a621dc7bd02f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qkf8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ngspk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:57Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.735336 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5ltx2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8881599e-71e5-4578-b3f9-21f4fa2f57aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbm6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbm6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5ltx2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:57Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.744517 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.744739 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.744818 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.744925 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.744994 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:57Z","lastTransitionTime":"2025-12-06T03:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.752541 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1721e985-40bd-4b2d-a0e8-d4e365094ba6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a895e0cc04880c3287ba725ac5484835235f1e066a94ce98fcf4b6133c5cbd2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a21a97385f856ec996105b597d0ae81c56ed28b625ef3f2905e1af6ec8f6a45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://646e01a58908ba2e703aae5eef53e1b981b0f0c3d1d3202487f2a8b94ee3f55d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31752b84d86bda30792b29e4c6ed963753cc576c5de784395e8e6dcbebaced47\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"ver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:40:26.813730 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814130 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:40:26.814139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:40:26.814143 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:40:26.814145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 03:40:26.821403 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821442 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821482 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821496 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 03:40:26.821509 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821514 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 03:40:26.821612 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 03:40:26.821619 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1206 03:40:26.822049 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16a92d501c9e626fe91e4a6793e426d10eb292cfbab766db8b8cc72e5dae94ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:57Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.847880 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.848189 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.848281 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.848352 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.848419 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:57Z","lastTransitionTime":"2025-12-06T03:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.951799 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.951867 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.951888 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.951917 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:57 crc kubenswrapper[4802]: I1206 03:40:57.951939 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:57Z","lastTransitionTime":"2025-12-06T03:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.055249 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.055651 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.055849 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.056003 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.056168 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:58Z","lastTransitionTime":"2025-12-06T03:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.159551 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.159607 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.159629 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.159658 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.159681 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:58Z","lastTransitionTime":"2025-12-06T03:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.262790 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.262838 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.262854 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.262877 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.262893 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:58Z","lastTransitionTime":"2025-12-06T03:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.324118 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:40:58 crc kubenswrapper[4802]: E1206 03:40:58.324358 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 03:40:58 crc kubenswrapper[4802]: E1206 03:40:58.325516 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 03:40:58 crc kubenswrapper[4802]: E1206 03:40:58.325538 4802 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:40:58 crc kubenswrapper[4802]: E1206 03:40:58.325622 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 03:41:30.32559956 +0000 UTC m=+83.197508752 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.366195 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.366253 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.366269 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.366293 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.366310 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:58Z","lastTransitionTime":"2025-12-06T03:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.426608 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.426831 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:40:58 crc kubenswrapper[4802]: E1206 03:40:58.426905 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:41:30.426856465 +0000 UTC m=+83.298765677 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.426987 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:40:58 crc kubenswrapper[4802]: E1206 03:40:58.427020 4802 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 03:40:58 crc kubenswrapper[4802]: E1206 03:40:58.427117 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 03:41:30.427088782 +0000 UTC m=+83.298997994 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.427181 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:40:58 crc kubenswrapper[4802]: E1206 03:40:58.427220 4802 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 03:40:58 crc kubenswrapper[4802]: E1206 03:40:58.427313 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 03:41:30.427286217 +0000 UTC m=+83.299195409 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 03:40:58 crc kubenswrapper[4802]: E1206 03:40:58.427354 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 03:40:58 crc kubenswrapper[4802]: E1206 03:40:58.427385 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 03:40:58 crc kubenswrapper[4802]: E1206 03:40:58.427407 4802 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:40:58 crc kubenswrapper[4802]: E1206 03:40:58.427464 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 03:41:30.427447161 +0000 UTC m=+83.299356353 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.449531 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.449584 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.449583 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:40:58 crc kubenswrapper[4802]: E1206 03:40:58.449736 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5ltx2" podUID="8881599e-71e5-4578-b3f9-21f4fa2f57aa" Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.449531 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:40:58 crc kubenswrapper[4802]: E1206 03:40:58.449901 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:40:58 crc kubenswrapper[4802]: E1206 03:40:58.450009 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:40:58 crc kubenswrapper[4802]: E1206 03:40:58.450083 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.469839 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.469922 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.469940 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.469964 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.469983 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:58Z","lastTransitionTime":"2025-12-06T03:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.573271 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.573313 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.573330 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.573354 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.573416 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:58Z","lastTransitionTime":"2025-12-06T03:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.676327 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.676395 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.676418 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.676446 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.676470 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:58Z","lastTransitionTime":"2025-12-06T03:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.779088 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.779151 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.779174 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.779203 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.779226 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:58Z","lastTransitionTime":"2025-12-06T03:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.882359 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.882420 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.882437 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.882460 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.882477 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:58Z","lastTransitionTime":"2025-12-06T03:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.985309 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.985370 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.985391 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.985420 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:58 crc kubenswrapper[4802]: I1206 03:40:58.985447 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:58Z","lastTransitionTime":"2025-12-06T03:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.088290 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.088609 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.088631 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.088660 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.088682 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:59Z","lastTransitionTime":"2025-12-06T03:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.191739 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.192123 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.192297 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.192444 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.192570 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:59Z","lastTransitionTime":"2025-12-06T03:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.295840 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.295901 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.295919 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.295942 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.295959 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:59Z","lastTransitionTime":"2025-12-06T03:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.399097 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.399156 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.399175 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.399197 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.399214 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:59Z","lastTransitionTime":"2025-12-06T03:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.503163 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.503510 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.503653 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.503819 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.503965 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:59Z","lastTransitionTime":"2025-12-06T03:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.606680 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.606736 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.606788 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.606815 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.606834 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:59Z","lastTransitionTime":"2025-12-06T03:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.641912 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.642365 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.642527 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.642679 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.642853 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:59Z","lastTransitionTime":"2025-12-06T03:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:59 crc kubenswrapper[4802]: E1206 03:40:59.667004 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0e01ca3a-758f-44bc-b4a2-30c1bc04fda0\\\",\\\"systemUUID\\\":\\\"c7c815e4-d2a4-4244-b035-f988fc95e215\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:59Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.672810 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.673088 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.673265 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.673470 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.673614 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:59Z","lastTransitionTime":"2025-12-06T03:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:59 crc kubenswrapper[4802]: E1206 03:40:59.695184 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0e01ca3a-758f-44bc-b4a2-30c1bc04fda0\\\",\\\"systemUUID\\\":\\\"c7c815e4-d2a4-4244-b035-f988fc95e215\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:59Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.701118 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.701177 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.701195 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.701220 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.701238 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:59Z","lastTransitionTime":"2025-12-06T03:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:59 crc kubenswrapper[4802]: E1206 03:40:59.720787 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0e01ca3a-758f-44bc-b4a2-30c1bc04fda0\\\",\\\"systemUUID\\\":\\\"c7c815e4-d2a4-4244-b035-f988fc95e215\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:59Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.725507 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.725556 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.725574 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.725598 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.725614 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:59Z","lastTransitionTime":"2025-12-06T03:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:59 crc kubenswrapper[4802]: E1206 03:40:59.744731 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0e01ca3a-758f-44bc-b4a2-30c1bc04fda0\\\",\\\"systemUUID\\\":\\\"c7c815e4-d2a4-4244-b035-f988fc95e215\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:59Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.750296 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.750571 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.750853 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.751023 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.751182 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:59Z","lastTransitionTime":"2025-12-06T03:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:59 crc kubenswrapper[4802]: E1206 03:40:59.772738 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0e01ca3a-758f-44bc-b4a2-30c1bc04fda0\\\",\\\"systemUUID\\\":\\\"c7c815e4-d2a4-4244-b035-f988fc95e215\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:40:59Z is after 2025-08-24T17:21:41Z" Dec 06 03:40:59 crc kubenswrapper[4802]: E1206 03:40:59.773433 4802 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.775409 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.775461 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.775478 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.775503 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.775520 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:59Z","lastTransitionTime":"2025-12-06T03:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.879055 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.879128 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.879152 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.879184 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.879207 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:59Z","lastTransitionTime":"2025-12-06T03:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.983236 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.983298 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.983317 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.983340 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:40:59 crc kubenswrapper[4802]: I1206 03:40:59.983358 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:40:59Z","lastTransitionTime":"2025-12-06T03:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:00 crc kubenswrapper[4802]: I1206 03:41:00.086124 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:00 crc kubenswrapper[4802]: I1206 03:41:00.086188 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:00 crc kubenswrapper[4802]: I1206 03:41:00.086199 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:00 crc kubenswrapper[4802]: I1206 03:41:00.086218 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:00 crc kubenswrapper[4802]: I1206 03:41:00.086229 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:00Z","lastTransitionTime":"2025-12-06T03:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:00 crc kubenswrapper[4802]: I1206 03:41:00.189244 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:00 crc kubenswrapper[4802]: I1206 03:41:00.189306 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:00 crc kubenswrapper[4802]: I1206 03:41:00.189323 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:00 crc kubenswrapper[4802]: I1206 03:41:00.189348 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:00 crc kubenswrapper[4802]: I1206 03:41:00.189365 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:00Z","lastTransitionTime":"2025-12-06T03:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:00 crc kubenswrapper[4802]: I1206 03:41:00.292899 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:00 crc kubenswrapper[4802]: I1206 03:41:00.292970 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:00 crc kubenswrapper[4802]: I1206 03:41:00.292988 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:00 crc kubenswrapper[4802]: I1206 03:41:00.293013 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:00 crc kubenswrapper[4802]: I1206 03:41:00.293034 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:00Z","lastTransitionTime":"2025-12-06T03:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:00 crc kubenswrapper[4802]: I1206 03:41:00.396042 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:00 crc kubenswrapper[4802]: I1206 03:41:00.396253 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:00 crc kubenswrapper[4802]: I1206 03:41:00.396269 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:00 crc kubenswrapper[4802]: I1206 03:41:00.396294 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:00 crc kubenswrapper[4802]: I1206 03:41:00.396311 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:00Z","lastTransitionTime":"2025-12-06T03:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:00 crc kubenswrapper[4802]: I1206 03:41:00.449358 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:41:00 crc kubenswrapper[4802]: I1206 03:41:00.449442 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:41:00 crc kubenswrapper[4802]: I1206 03:41:00.449389 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:41:00 crc kubenswrapper[4802]: I1206 03:41:00.449389 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:41:00 crc kubenswrapper[4802]: E1206 03:41:00.449561 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5ltx2" podUID="8881599e-71e5-4578-b3f9-21f4fa2f57aa" Dec 06 03:41:00 crc kubenswrapper[4802]: E1206 03:41:00.449686 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:41:00 crc kubenswrapper[4802]: E1206 03:41:00.449829 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:41:00 crc kubenswrapper[4802]: E1206 03:41:00.449911 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:41:00 crc kubenswrapper[4802]: I1206 03:41:00.498691 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:00 crc kubenswrapper[4802]: I1206 03:41:00.498790 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:00 crc kubenswrapper[4802]: I1206 03:41:00.498816 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:00 crc kubenswrapper[4802]: I1206 03:41:00.498849 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:00 crc kubenswrapper[4802]: I1206 03:41:00.498872 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:00Z","lastTransitionTime":"2025-12-06T03:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:00 crc kubenswrapper[4802]: I1206 03:41:00.601713 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:00 crc kubenswrapper[4802]: I1206 03:41:00.601795 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:00 crc kubenswrapper[4802]: I1206 03:41:00.601813 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:00 crc kubenswrapper[4802]: I1206 03:41:00.601850 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:00 crc kubenswrapper[4802]: I1206 03:41:00.601892 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:00Z","lastTransitionTime":"2025-12-06T03:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:00 crc kubenswrapper[4802]: I1206 03:41:00.705400 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:00 crc kubenswrapper[4802]: I1206 03:41:00.705469 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:00 crc kubenswrapper[4802]: I1206 03:41:00.705490 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:00 crc kubenswrapper[4802]: I1206 03:41:00.705516 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:00 crc kubenswrapper[4802]: I1206 03:41:00.705538 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:00Z","lastTransitionTime":"2025-12-06T03:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:00 crc kubenswrapper[4802]: I1206 03:41:00.808286 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:00 crc kubenswrapper[4802]: I1206 03:41:00.808346 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:00 crc kubenswrapper[4802]: I1206 03:41:00.808368 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:00 crc kubenswrapper[4802]: I1206 03:41:00.808397 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:00 crc kubenswrapper[4802]: I1206 03:41:00.808421 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:00Z","lastTransitionTime":"2025-12-06T03:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:00 crc kubenswrapper[4802]: I1206 03:41:00.911674 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:00 crc kubenswrapper[4802]: I1206 03:41:00.911742 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:00 crc kubenswrapper[4802]: I1206 03:41:00.911793 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:00 crc kubenswrapper[4802]: I1206 03:41:00.911826 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:00 crc kubenswrapper[4802]: I1206 03:41:00.911849 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:00Z","lastTransitionTime":"2025-12-06T03:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:01 crc kubenswrapper[4802]: I1206 03:41:01.015309 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:01 crc kubenswrapper[4802]: I1206 03:41:01.015415 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:01 crc kubenswrapper[4802]: I1206 03:41:01.015436 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:01 crc kubenswrapper[4802]: I1206 03:41:01.015466 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:01 crc kubenswrapper[4802]: I1206 03:41:01.015489 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:01Z","lastTransitionTime":"2025-12-06T03:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:01 crc kubenswrapper[4802]: I1206 03:41:01.118523 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:01 crc kubenswrapper[4802]: I1206 03:41:01.118580 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:01 crc kubenswrapper[4802]: I1206 03:41:01.118599 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:01 crc kubenswrapper[4802]: I1206 03:41:01.118624 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:01 crc kubenswrapper[4802]: I1206 03:41:01.118667 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:01Z","lastTransitionTime":"2025-12-06T03:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:01 crc kubenswrapper[4802]: I1206 03:41:01.221922 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:01 crc kubenswrapper[4802]: I1206 03:41:01.222292 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:01 crc kubenswrapper[4802]: I1206 03:41:01.222469 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:01 crc kubenswrapper[4802]: I1206 03:41:01.222648 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:01 crc kubenswrapper[4802]: I1206 03:41:01.222933 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:01Z","lastTransitionTime":"2025-12-06T03:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:01 crc kubenswrapper[4802]: I1206 03:41:01.326445 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:01 crc kubenswrapper[4802]: I1206 03:41:01.326505 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:01 crc kubenswrapper[4802]: I1206 03:41:01.326522 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:01 crc kubenswrapper[4802]: I1206 03:41:01.326544 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:01 crc kubenswrapper[4802]: I1206 03:41:01.326561 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:01Z","lastTransitionTime":"2025-12-06T03:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:01 crc kubenswrapper[4802]: I1206 03:41:01.429796 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:01 crc kubenswrapper[4802]: I1206 03:41:01.429856 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:01 crc kubenswrapper[4802]: I1206 03:41:01.429876 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:01 crc kubenswrapper[4802]: I1206 03:41:01.429901 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:01 crc kubenswrapper[4802]: I1206 03:41:01.429921 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:01Z","lastTransitionTime":"2025-12-06T03:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:01 crc kubenswrapper[4802]: I1206 03:41:01.533082 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:01 crc kubenswrapper[4802]: I1206 03:41:01.533146 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:01 crc kubenswrapper[4802]: I1206 03:41:01.533164 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:01 crc kubenswrapper[4802]: I1206 03:41:01.533188 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:01 crc kubenswrapper[4802]: I1206 03:41:01.533205 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:01Z","lastTransitionTime":"2025-12-06T03:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:01 crc kubenswrapper[4802]: I1206 03:41:01.636675 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:01 crc kubenswrapper[4802]: I1206 03:41:01.636812 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:01 crc kubenswrapper[4802]: I1206 03:41:01.636838 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:01 crc kubenswrapper[4802]: I1206 03:41:01.636871 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:01 crc kubenswrapper[4802]: I1206 03:41:01.636896 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:01Z","lastTransitionTime":"2025-12-06T03:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:01 crc kubenswrapper[4802]: I1206 03:41:01.740324 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:01 crc kubenswrapper[4802]: I1206 03:41:01.740416 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:01 crc kubenswrapper[4802]: I1206 03:41:01.740433 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:01 crc kubenswrapper[4802]: I1206 03:41:01.740455 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:01 crc kubenswrapper[4802]: I1206 03:41:01.740473 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:01Z","lastTransitionTime":"2025-12-06T03:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:01 crc kubenswrapper[4802]: I1206 03:41:01.842916 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:01 crc kubenswrapper[4802]: I1206 03:41:01.843007 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:01 crc kubenswrapper[4802]: I1206 03:41:01.843029 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:01 crc kubenswrapper[4802]: I1206 03:41:01.843059 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:01 crc kubenswrapper[4802]: I1206 03:41:01.843080 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:01Z","lastTransitionTime":"2025-12-06T03:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:01 crc kubenswrapper[4802]: I1206 03:41:01.945738 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:01 crc kubenswrapper[4802]: I1206 03:41:01.945829 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:01 crc kubenswrapper[4802]: I1206 03:41:01.945869 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:01 crc kubenswrapper[4802]: I1206 03:41:01.945904 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:01 crc kubenswrapper[4802]: I1206 03:41:01.945922 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:01Z","lastTransitionTime":"2025-12-06T03:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.041017 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.048508 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.048598 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.048622 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.048651 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.048673 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:02Z","lastTransitionTime":"2025-12-06T03:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.056949 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.067438 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtw2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3a3bac2-3923-4507-a4de-1556b3ad473e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fed5ee4189d54456a3216f4fdb1ecf003fa47e29dbfe4b2b2f301edfb897d6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtw2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:02Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.102700 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7730de4f033561ad77046e028cd1a3f0bfe2775387cbdfa35f47ea79909ec4cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a158aebd701ef46225d3e37109ae28ab58d107087bf06650c4b6d0055a3bbba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2357f37497fde64937eca7b38693893af745ab278653ebc79406e122a6d4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8941bd25107333b264d3f081a811fdfeea23ad92193fc61367665fd2cd1209c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f9cd4dc75e63f09f60b48b6da284643910bb32ba51f2a808cb4d9803f30042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6796f65ee9c5bd0c18f205697135d17677c5daf3ba378cc992cba795e0188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3870b61bce487c69724f668c38fc02190ff97b5e3b1434d66ed624412d18319b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf845e8893e51078e46991489d9a27c1194eab62a9d0a24eb3336d396e12cdfb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:40:38Z\\\",\\\"message\\\":\\\"Map:map[10.217.4.110:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {f9232b32-e89f-4c8e-acc4-c6801b70dcb0}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:NB_Global Row:map[] Rows:[] Columns:[] Mutations:[{Column:nb_cfg Mutator:+= Value:1}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {6011affd-30a6-4be6-872d-e4cf1ca780cf}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 03:40:38.703973 6236 ovnkube.go:599] Stopped ovnkube\\\\nI1206 03:40:38.704014 6236 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1206 03:40:38.703921 6236 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-controller]} name:Service_openshift-machine-config-operator/machine-config-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.16:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {3f1b9878-e751-4e46-a226-ce007d2c4aa7}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1206 03:40:38.704093 6236 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3870b61bce487c69724f668c38fc02190ff97b5e3b1434d66ed624412d18319b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:40:54Z\\\",\\\"message\\\":\\\"/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 03:40:54.580644 6463 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 03:40:54.580730 6463 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 03:40:54.580797 6463 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 03:40:54.580836 6463 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 03:40:54.580845 6463 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 03:40:54.580895 6463 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1206 03:40:54.580943 6463 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1206 03:40:54.580950 6463 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1206 03:40:54.581001 6463 factory.go:656] Stopping watch factory\\\\nI1206 03:40:54.581036 6463 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 03:40:54.581050 6463 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 03:40:54.581063 6463 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 03:40:54.581074 6463 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 03:40:54.581085 6463 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1206 03:40:54.581102 6463 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3cb210bfc97a439a7ff98c02c59f8ad16bc400dc214a616a44b43a8b01e2782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pvm6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:02Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.123502 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:02Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.140275 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1721e985-40bd-4b2d-a0e8-d4e365094ba6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a895e0cc04880c3287ba725ac5484835235f1e066a94ce98fcf4b6133c5cbd2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a21a97385f856ec996105b597d0ae81c56ed28b625ef3f2905e1af6ec8f6a45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://646e01a58908ba2e703aae5eef53e1b981b0f0c3d1d3202487f2a8b94ee3f55d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31752b84d86bda30792b29e4c6ed963753cc576c5de784395e8e6dcbebaced47\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"ver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:40:26.813730 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814130 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:40:26.814139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:40:26.814143 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:40:26.814145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 03:40:26.821403 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821442 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821482 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821496 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 03:40:26.821509 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821514 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 03:40:26.821612 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 03:40:26.821619 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1206 03:40:26.822049 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16a92d501c9e626fe91e4a6793e426d10eb292cfbab766db8b8cc72e5dae94ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:02Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.152113 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.152175 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.152200 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.152231 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.152255 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:02Z","lastTransitionTime":"2025-12-06T03:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.155068 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-b2knv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a3ae46f-03b5-4999-9d58-c38ee222f68a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76665d4c127445009fd66078827b54607c9ef812c8011d1c0b4cde1c0516853a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk4hj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-b2knv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:02Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.173593 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22rrq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03112a5d-5882-4fac-9ae5-13525ff82fe0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0205b089b12c42688203bb7ec0f27dc5f9a577dabdb446b811f184046f85dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trlt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22rrq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:02Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.188359 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ngspk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34808ab2-d21b-40b6-8fe3-e18bf9e5fc46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52509d21f5b4bfbd458162cba176ff82c627c7c4671204addf9d76f7392fd760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qkf8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9685a4d5d4f40ff4f59b6da6193785f340890453d1e422d3421a621dc7bd02f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qkf8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ngspk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:02Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.203286 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5ltx2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8881599e-71e5-4578-b3f9-21f4fa2f57aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbm6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbm6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5ltx2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:02Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.222181 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:02Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.235323 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6df38316-e0d3-4018-8d27-3620eba3a68d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57381e9589a2dc04a4d2733c37985f42594182128a374b7be0f471ea69138b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bf5e41701c7a2d45adc4740ceb4ac1785db17bc7cb15daed439f7556812d1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpxxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:02Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.245585 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kj7hf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd46dfa5-7d4a-49ed-8199-523f10cebdf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5eee8a0cf96ceb3785dac9df31bf8d31715d1c88fbea17c5e582b0a4554bb63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmkcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kj7hf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:02Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.255049 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.255155 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.255182 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.255220 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.255244 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:02Z","lastTransitionTime":"2025-12-06T03:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.259480 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2466610962332c6ca120fe055e4e8a3fc623088119dc1a412e06b745176f89e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:02Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.272819 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2093b6-c4b5-4db8-be16-fbc859885e8d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0d6e582161f0d9126126e03b27211be22f6b5f797154928c92fa13a8118f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03004a29aefe7fe2452fb9f6663c0807afccb055d7ddbf6b907f305e6a19c81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1801ac64db4ba91c2e8a249e3651d3e50ab62e9f2c87312ce9bf7f38518a82cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6e454507cadbb982a566a3692995aaa3c733ff1cd7fb4870476602190e1eeba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:02Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.285890 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:02Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.302285 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fad9b63d84f41fccf696cd313d7ca7f72b3d3251d6c1b7341573a3fadf1e378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:02Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.314595 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb3f278fff47c900c56d8a563d5ed75a2ae84e2034b0f45aad4083790b9e7946\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b10ca16ac26904ec5e6245344b71d836888cae9b069cae6a7a4831d609752db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:02Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.357434 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.357503 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.357521 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.357545 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.357564 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:02Z","lastTransitionTime":"2025-12-06T03:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.449688 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:41:02 crc kubenswrapper[4802]: E1206 03:41:02.449904 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5ltx2" podUID="8881599e-71e5-4578-b3f9-21f4fa2f57aa" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.449979 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.450001 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:41:02 crc kubenswrapper[4802]: E1206 03:41:02.450184 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.450255 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:41:02 crc kubenswrapper[4802]: E1206 03:41:02.450542 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:41:02 crc kubenswrapper[4802]: E1206 03:41:02.450320 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.459987 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.460036 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.460059 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.460090 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.460113 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:02Z","lastTransitionTime":"2025-12-06T03:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.563506 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.563577 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.563602 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.563631 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.563651 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:02Z","lastTransitionTime":"2025-12-06T03:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.667475 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.667556 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.667581 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.667611 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.667634 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:02Z","lastTransitionTime":"2025-12-06T03:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.771041 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.771116 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.771149 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.771192 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.771211 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:02Z","lastTransitionTime":"2025-12-06T03:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.874121 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.874259 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.874345 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.874383 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.874408 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:02Z","lastTransitionTime":"2025-12-06T03:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.978023 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.978073 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.978089 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.978113 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:02 crc kubenswrapper[4802]: I1206 03:41:02.978131 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:02Z","lastTransitionTime":"2025-12-06T03:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:03 crc kubenswrapper[4802]: I1206 03:41:03.081528 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:03 crc kubenswrapper[4802]: I1206 03:41:03.082011 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:03 crc kubenswrapper[4802]: I1206 03:41:03.082233 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:03 crc kubenswrapper[4802]: I1206 03:41:03.082451 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:03 crc kubenswrapper[4802]: I1206 03:41:03.082612 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:03Z","lastTransitionTime":"2025-12-06T03:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:03 crc kubenswrapper[4802]: I1206 03:41:03.185813 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:03 crc kubenswrapper[4802]: I1206 03:41:03.185866 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:03 crc kubenswrapper[4802]: I1206 03:41:03.185884 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:03 crc kubenswrapper[4802]: I1206 03:41:03.185906 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:03 crc kubenswrapper[4802]: I1206 03:41:03.185923 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:03Z","lastTransitionTime":"2025-12-06T03:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:03 crc kubenswrapper[4802]: I1206 03:41:03.289476 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:03 crc kubenswrapper[4802]: I1206 03:41:03.289555 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:03 crc kubenswrapper[4802]: I1206 03:41:03.289578 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:03 crc kubenswrapper[4802]: I1206 03:41:03.289613 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:03 crc kubenswrapper[4802]: I1206 03:41:03.289638 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:03Z","lastTransitionTime":"2025-12-06T03:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:03 crc kubenswrapper[4802]: I1206 03:41:03.392843 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:03 crc kubenswrapper[4802]: I1206 03:41:03.392934 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:03 crc kubenswrapper[4802]: I1206 03:41:03.392962 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:03 crc kubenswrapper[4802]: I1206 03:41:03.392987 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:03 crc kubenswrapper[4802]: I1206 03:41:03.393008 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:03Z","lastTransitionTime":"2025-12-06T03:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:03 crc kubenswrapper[4802]: I1206 03:41:03.496407 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:03 crc kubenswrapper[4802]: I1206 03:41:03.496485 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:03 crc kubenswrapper[4802]: I1206 03:41:03.496498 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:03 crc kubenswrapper[4802]: I1206 03:41:03.496520 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:03 crc kubenswrapper[4802]: I1206 03:41:03.496557 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:03Z","lastTransitionTime":"2025-12-06T03:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:03 crc kubenswrapper[4802]: I1206 03:41:03.599441 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:03 crc kubenswrapper[4802]: I1206 03:41:03.599491 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:03 crc kubenswrapper[4802]: I1206 03:41:03.599526 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:03 crc kubenswrapper[4802]: I1206 03:41:03.599546 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:03 crc kubenswrapper[4802]: I1206 03:41:03.599555 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:03Z","lastTransitionTime":"2025-12-06T03:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:03 crc kubenswrapper[4802]: I1206 03:41:03.702032 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:03 crc kubenswrapper[4802]: I1206 03:41:03.702075 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:03 crc kubenswrapper[4802]: I1206 03:41:03.702092 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:03 crc kubenswrapper[4802]: I1206 03:41:03.702113 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:03 crc kubenswrapper[4802]: I1206 03:41:03.702128 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:03Z","lastTransitionTime":"2025-12-06T03:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:03 crc kubenswrapper[4802]: I1206 03:41:03.805258 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:03 crc kubenswrapper[4802]: I1206 03:41:03.805328 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:03 crc kubenswrapper[4802]: I1206 03:41:03.805354 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:03 crc kubenswrapper[4802]: I1206 03:41:03.805403 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:03 crc kubenswrapper[4802]: I1206 03:41:03.805421 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:03Z","lastTransitionTime":"2025-12-06T03:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:03 crc kubenswrapper[4802]: I1206 03:41:03.908368 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:03 crc kubenswrapper[4802]: I1206 03:41:03.908454 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:03 crc kubenswrapper[4802]: I1206 03:41:03.908479 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:03 crc kubenswrapper[4802]: I1206 03:41:03.908509 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:03 crc kubenswrapper[4802]: I1206 03:41:03.908534 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:03Z","lastTransitionTime":"2025-12-06T03:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:04 crc kubenswrapper[4802]: I1206 03:41:04.011711 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:04 crc kubenswrapper[4802]: I1206 03:41:04.012003 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:04 crc kubenswrapper[4802]: I1206 03:41:04.012029 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:04 crc kubenswrapper[4802]: I1206 03:41:04.012054 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:04 crc kubenswrapper[4802]: I1206 03:41:04.012072 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:04Z","lastTransitionTime":"2025-12-06T03:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:04 crc kubenswrapper[4802]: I1206 03:41:04.114526 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:04 crc kubenswrapper[4802]: I1206 03:41:04.114588 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:04 crc kubenswrapper[4802]: I1206 03:41:04.114606 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:04 crc kubenswrapper[4802]: I1206 03:41:04.114630 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:04 crc kubenswrapper[4802]: I1206 03:41:04.114653 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:04Z","lastTransitionTime":"2025-12-06T03:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:04 crc kubenswrapper[4802]: I1206 03:41:04.218465 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:04 crc kubenswrapper[4802]: I1206 03:41:04.218548 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:04 crc kubenswrapper[4802]: I1206 03:41:04.218566 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:04 crc kubenswrapper[4802]: I1206 03:41:04.218632 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:04 crc kubenswrapper[4802]: I1206 03:41:04.218650 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:04Z","lastTransitionTime":"2025-12-06T03:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:04 crc kubenswrapper[4802]: I1206 03:41:04.322294 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:04 crc kubenswrapper[4802]: I1206 03:41:04.322384 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:04 crc kubenswrapper[4802]: I1206 03:41:04.322402 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:04 crc kubenswrapper[4802]: I1206 03:41:04.322458 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:04 crc kubenswrapper[4802]: I1206 03:41:04.322476 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:04Z","lastTransitionTime":"2025-12-06T03:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:04 crc kubenswrapper[4802]: I1206 03:41:04.426068 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:04 crc kubenswrapper[4802]: I1206 03:41:04.426130 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:04 crc kubenswrapper[4802]: I1206 03:41:04.426148 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:04 crc kubenswrapper[4802]: I1206 03:41:04.426172 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:04 crc kubenswrapper[4802]: I1206 03:41:04.426187 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:04Z","lastTransitionTime":"2025-12-06T03:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:04 crc kubenswrapper[4802]: I1206 03:41:04.449724 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:41:04 crc kubenswrapper[4802]: I1206 03:41:04.449811 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:41:04 crc kubenswrapper[4802]: I1206 03:41:04.449829 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:41:04 crc kubenswrapper[4802]: I1206 03:41:04.449783 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:41:04 crc kubenswrapper[4802]: E1206 03:41:04.449965 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:41:04 crc kubenswrapper[4802]: E1206 03:41:04.450150 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:41:04 crc kubenswrapper[4802]: E1206 03:41:04.450223 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5ltx2" podUID="8881599e-71e5-4578-b3f9-21f4fa2f57aa" Dec 06 03:41:04 crc kubenswrapper[4802]: E1206 03:41:04.450311 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:41:04 crc kubenswrapper[4802]: I1206 03:41:04.528385 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:04 crc kubenswrapper[4802]: I1206 03:41:04.528421 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:04 crc kubenswrapper[4802]: I1206 03:41:04.528433 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:04 crc kubenswrapper[4802]: I1206 03:41:04.528450 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:04 crc kubenswrapper[4802]: I1206 03:41:04.528461 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:04Z","lastTransitionTime":"2025-12-06T03:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:04 crc kubenswrapper[4802]: I1206 03:41:04.630974 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:04 crc kubenswrapper[4802]: I1206 03:41:04.631035 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:04 crc kubenswrapper[4802]: I1206 03:41:04.631055 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:04 crc kubenswrapper[4802]: I1206 03:41:04.631078 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:04 crc kubenswrapper[4802]: I1206 03:41:04.631094 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:04Z","lastTransitionTime":"2025-12-06T03:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:04 crc kubenswrapper[4802]: I1206 03:41:04.734141 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:04 crc kubenswrapper[4802]: I1206 03:41:04.734211 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:04 crc kubenswrapper[4802]: I1206 03:41:04.734228 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:04 crc kubenswrapper[4802]: I1206 03:41:04.734253 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:04 crc kubenswrapper[4802]: I1206 03:41:04.734276 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:04Z","lastTransitionTime":"2025-12-06T03:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:04 crc kubenswrapper[4802]: I1206 03:41:04.837720 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:04 crc kubenswrapper[4802]: I1206 03:41:04.837843 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:04 crc kubenswrapper[4802]: I1206 03:41:04.837869 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:04 crc kubenswrapper[4802]: I1206 03:41:04.837963 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:04 crc kubenswrapper[4802]: I1206 03:41:04.837981 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:04Z","lastTransitionTime":"2025-12-06T03:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:04 crc kubenswrapper[4802]: I1206 03:41:04.941649 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:04 crc kubenswrapper[4802]: I1206 03:41:04.941693 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:04 crc kubenswrapper[4802]: I1206 03:41:04.941705 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:04 crc kubenswrapper[4802]: I1206 03:41:04.941721 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:04 crc kubenswrapper[4802]: I1206 03:41:04.941732 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:04Z","lastTransitionTime":"2025-12-06T03:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:05 crc kubenswrapper[4802]: I1206 03:41:05.044940 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:05 crc kubenswrapper[4802]: I1206 03:41:05.045002 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:05 crc kubenswrapper[4802]: I1206 03:41:05.045024 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:05 crc kubenswrapper[4802]: I1206 03:41:05.045055 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:05 crc kubenswrapper[4802]: I1206 03:41:05.045080 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:05Z","lastTransitionTime":"2025-12-06T03:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:05 crc kubenswrapper[4802]: I1206 03:41:05.147969 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:05 crc kubenswrapper[4802]: I1206 03:41:05.148051 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:05 crc kubenswrapper[4802]: I1206 03:41:05.148070 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:05 crc kubenswrapper[4802]: I1206 03:41:05.148099 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:05 crc kubenswrapper[4802]: I1206 03:41:05.148117 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:05Z","lastTransitionTime":"2025-12-06T03:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:05 crc kubenswrapper[4802]: I1206 03:41:05.251991 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:05 crc kubenswrapper[4802]: I1206 03:41:05.252062 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:05 crc kubenswrapper[4802]: I1206 03:41:05.252082 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:05 crc kubenswrapper[4802]: I1206 03:41:05.252108 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:05 crc kubenswrapper[4802]: I1206 03:41:05.252129 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:05Z","lastTransitionTime":"2025-12-06T03:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:05 crc kubenswrapper[4802]: I1206 03:41:05.355672 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:05 crc kubenswrapper[4802]: I1206 03:41:05.355741 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:05 crc kubenswrapper[4802]: I1206 03:41:05.355799 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:05 crc kubenswrapper[4802]: I1206 03:41:05.355831 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:05 crc kubenswrapper[4802]: I1206 03:41:05.355857 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:05Z","lastTransitionTime":"2025-12-06T03:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:05 crc kubenswrapper[4802]: I1206 03:41:05.458858 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:05 crc kubenswrapper[4802]: I1206 03:41:05.458928 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:05 crc kubenswrapper[4802]: I1206 03:41:05.458948 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:05 crc kubenswrapper[4802]: I1206 03:41:05.458973 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:05 crc kubenswrapper[4802]: I1206 03:41:05.458992 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:05Z","lastTransitionTime":"2025-12-06T03:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:05 crc kubenswrapper[4802]: I1206 03:41:05.562739 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:05 crc kubenswrapper[4802]: I1206 03:41:05.562832 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:05 crc kubenswrapper[4802]: I1206 03:41:05.562850 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:05 crc kubenswrapper[4802]: I1206 03:41:05.562879 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:05 crc kubenswrapper[4802]: I1206 03:41:05.562905 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:05Z","lastTransitionTime":"2025-12-06T03:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:05 crc kubenswrapper[4802]: I1206 03:41:05.666042 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:05 crc kubenswrapper[4802]: I1206 03:41:05.666105 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:05 crc kubenswrapper[4802]: I1206 03:41:05.666123 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:05 crc kubenswrapper[4802]: I1206 03:41:05.666148 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:05 crc kubenswrapper[4802]: I1206 03:41:05.666167 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:05Z","lastTransitionTime":"2025-12-06T03:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:05 crc kubenswrapper[4802]: I1206 03:41:05.768492 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:05 crc kubenswrapper[4802]: I1206 03:41:05.768592 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:05 crc kubenswrapper[4802]: I1206 03:41:05.768612 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:05 crc kubenswrapper[4802]: I1206 03:41:05.768639 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:05 crc kubenswrapper[4802]: I1206 03:41:05.768658 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:05Z","lastTransitionTime":"2025-12-06T03:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:05 crc kubenswrapper[4802]: I1206 03:41:05.871735 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:05 crc kubenswrapper[4802]: I1206 03:41:05.871929 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:05 crc kubenswrapper[4802]: I1206 03:41:05.871958 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:05 crc kubenswrapper[4802]: I1206 03:41:05.871989 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:05 crc kubenswrapper[4802]: I1206 03:41:05.872012 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:05Z","lastTransitionTime":"2025-12-06T03:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:05 crc kubenswrapper[4802]: I1206 03:41:05.975602 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:05 crc kubenswrapper[4802]: I1206 03:41:05.975670 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:05 crc kubenswrapper[4802]: I1206 03:41:05.975687 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:05 crc kubenswrapper[4802]: I1206 03:41:05.975712 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:05 crc kubenswrapper[4802]: I1206 03:41:05.975732 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:05Z","lastTransitionTime":"2025-12-06T03:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:06 crc kubenswrapper[4802]: I1206 03:41:06.078934 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:06 crc kubenswrapper[4802]: I1206 03:41:06.079018 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:06 crc kubenswrapper[4802]: I1206 03:41:06.079044 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:06 crc kubenswrapper[4802]: I1206 03:41:06.079074 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:06 crc kubenswrapper[4802]: I1206 03:41:06.079095 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:06Z","lastTransitionTime":"2025-12-06T03:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:06 crc kubenswrapper[4802]: I1206 03:41:06.182391 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:06 crc kubenswrapper[4802]: I1206 03:41:06.182469 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:06 crc kubenswrapper[4802]: I1206 03:41:06.182488 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:06 crc kubenswrapper[4802]: I1206 03:41:06.182514 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:06 crc kubenswrapper[4802]: I1206 03:41:06.182537 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:06Z","lastTransitionTime":"2025-12-06T03:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:06 crc kubenswrapper[4802]: I1206 03:41:06.285438 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:06 crc kubenswrapper[4802]: I1206 03:41:06.285491 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:06 crc kubenswrapper[4802]: I1206 03:41:06.285509 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:06 crc kubenswrapper[4802]: I1206 03:41:06.285533 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:06 crc kubenswrapper[4802]: I1206 03:41:06.285549 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:06Z","lastTransitionTime":"2025-12-06T03:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:06 crc kubenswrapper[4802]: I1206 03:41:06.388471 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:06 crc kubenswrapper[4802]: I1206 03:41:06.388583 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:06 crc kubenswrapper[4802]: I1206 03:41:06.388609 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:06 crc kubenswrapper[4802]: I1206 03:41:06.388640 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:06 crc kubenswrapper[4802]: I1206 03:41:06.388659 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:06Z","lastTransitionTime":"2025-12-06T03:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:06 crc kubenswrapper[4802]: I1206 03:41:06.449819 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:41:06 crc kubenswrapper[4802]: I1206 03:41:06.449851 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:41:06 crc kubenswrapper[4802]: I1206 03:41:06.449903 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:41:06 crc kubenswrapper[4802]: E1206 03:41:06.449960 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:41:06 crc kubenswrapper[4802]: I1206 03:41:06.449918 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:41:06 crc kubenswrapper[4802]: E1206 03:41:06.450102 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5ltx2" podUID="8881599e-71e5-4578-b3f9-21f4fa2f57aa" Dec 06 03:41:06 crc kubenswrapper[4802]: E1206 03:41:06.450297 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:41:06 crc kubenswrapper[4802]: E1206 03:41:06.450516 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:41:06 crc kubenswrapper[4802]: I1206 03:41:06.492285 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:06 crc kubenswrapper[4802]: I1206 03:41:06.492344 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:06 crc kubenswrapper[4802]: I1206 03:41:06.492361 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:06 crc kubenswrapper[4802]: I1206 03:41:06.492384 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:06 crc kubenswrapper[4802]: I1206 03:41:06.492402 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:06Z","lastTransitionTime":"2025-12-06T03:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:06 crc kubenswrapper[4802]: I1206 03:41:06.595450 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:06 crc kubenswrapper[4802]: I1206 03:41:06.595520 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:06 crc kubenswrapper[4802]: I1206 03:41:06.595547 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:06 crc kubenswrapper[4802]: I1206 03:41:06.595592 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:06 crc kubenswrapper[4802]: I1206 03:41:06.595618 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:06Z","lastTransitionTime":"2025-12-06T03:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:06 crc kubenswrapper[4802]: I1206 03:41:06.698307 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:06 crc kubenswrapper[4802]: I1206 03:41:06.698379 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:06 crc kubenswrapper[4802]: I1206 03:41:06.698400 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:06 crc kubenswrapper[4802]: I1206 03:41:06.698426 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:06 crc kubenswrapper[4802]: I1206 03:41:06.698452 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:06Z","lastTransitionTime":"2025-12-06T03:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:06 crc kubenswrapper[4802]: I1206 03:41:06.801211 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:06 crc kubenswrapper[4802]: I1206 03:41:06.801298 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:06 crc kubenswrapper[4802]: I1206 03:41:06.801334 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:06 crc kubenswrapper[4802]: I1206 03:41:06.801364 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:06 crc kubenswrapper[4802]: I1206 03:41:06.801387 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:06Z","lastTransitionTime":"2025-12-06T03:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:06 crc kubenswrapper[4802]: I1206 03:41:06.904108 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:06 crc kubenswrapper[4802]: I1206 03:41:06.904205 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:06 crc kubenswrapper[4802]: I1206 03:41:06.904238 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:06 crc kubenswrapper[4802]: I1206 03:41:06.904269 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:06 crc kubenswrapper[4802]: I1206 03:41:06.904287 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:06Z","lastTransitionTime":"2025-12-06T03:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.007203 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.007257 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.007274 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.007299 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.007319 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:07Z","lastTransitionTime":"2025-12-06T03:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.110669 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.110727 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.110747 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.110826 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.110851 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:07Z","lastTransitionTime":"2025-12-06T03:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.213193 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.213266 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.213290 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.213320 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.213344 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:07Z","lastTransitionTime":"2025-12-06T03:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.316494 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.316569 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.316592 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.316620 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.316640 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:07Z","lastTransitionTime":"2025-12-06T03:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.419424 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.419491 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.419515 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.419548 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.419571 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:07Z","lastTransitionTime":"2025-12-06T03:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.467385 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2093b6-c4b5-4db8-be16-fbc859885e8d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0d6e582161f0d9126126e03b27211be22f6b5f797154928c92fa13a8118f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03004a29aefe7fe2452fb9f6663c0807afccb055d7ddbf6b907f305e6a19c81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1801ac64db4ba91c2e8a249e3651d3e50ab62e9f2c87312ce9bf7f38518a82cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6e454507cadbb982a566a3692995aaa3c733ff1cd7fb4870476602190e1eeba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:07Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.485436 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:07Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.504389 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fad9b63d84f41fccf696cd313d7ca7f72b3d3251d6c1b7341573a3fadf1e378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:07Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.521090 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb3f278fff47c900c56d8a563d5ed75a2ae84e2034b0f45aad4083790b9e7946\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b10ca16ac26904ec5e6245344b71d836888cae9b069cae6a7a4831d609752db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:07Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.522431 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.522495 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.522516 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.522539 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.522559 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:07Z","lastTransitionTime":"2025-12-06T03:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.541605 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:07Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.561130 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtw2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3a3bac2-3923-4507-a4de-1556b3ad473e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fed5ee4189d54456a3216f4fdb1ecf003fa47e29dbfe4b2b2f301edfb897d6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtw2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:07Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.578450 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7730de4f033561ad77046e028cd1a3f0bfe2775387cbdfa35f47ea79909ec4cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a158aebd701ef46225d3e37109ae28ab58d107087bf06650c4b6d0055a3bbba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2357f37497fde64937eca7b38693893af745ab278653ebc79406e122a6d4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8941bd25107333b264d3f081a811fdfeea23ad92193fc61367665fd2cd1209c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f9cd4dc75e63f09f60b48b6da284643910bb32ba51f2a808cb4d9803f30042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6796f65ee9c5bd0c18f205697135d17677c5daf3ba378cc992cba795e0188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3870b61bce487c69724f668c38fc02190ff97b5e3b1434d66ed624412d18319b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf845e8893e51078e46991489d9a27c1194eab62a9d0a24eb3336d396e12cdfb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:40:38Z\\\",\\\"message\\\":\\\"Map:map[10.217.4.110:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {f9232b32-e89f-4c8e-acc4-c6801b70dcb0}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:NB_Global Row:map[] Rows:[] Columns:[] Mutations:[{Column:nb_cfg Mutator:+= Value:1}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {6011affd-30a6-4be6-872d-e4cf1ca780cf}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 03:40:38.703973 6236 ovnkube.go:599] Stopped ovnkube\\\\nI1206 03:40:38.704014 6236 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1206 03:40:38.703921 6236 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-controller]} name:Service_openshift-machine-config-operator/machine-config-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.16:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {3f1b9878-e751-4e46-a226-ce007d2c4aa7}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1206 03:40:38.704093 6236 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3870b61bce487c69724f668c38fc02190ff97b5e3b1434d66ed624412d18319b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:40:54Z\\\",\\\"message\\\":\\\"/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 03:40:54.580644 6463 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 03:40:54.580730 6463 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 03:40:54.580797 6463 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 03:40:54.580836 6463 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 03:40:54.580845 6463 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 03:40:54.580895 6463 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1206 03:40:54.580943 6463 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1206 03:40:54.580950 6463 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1206 03:40:54.581001 6463 factory.go:656] Stopping watch factory\\\\nI1206 03:40:54.581036 6463 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 03:40:54.581050 6463 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 03:40:54.581063 6463 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 03:40:54.581074 6463 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 03:40:54.581085 6463 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1206 03:40:54.581102 6463 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3cb210bfc97a439a7ff98c02c59f8ad16bc400dc214a616a44b43a8b01e2782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pvm6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:07Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.590991 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1721e985-40bd-4b2d-a0e8-d4e365094ba6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a895e0cc04880c3287ba725ac5484835235f1e066a94ce98fcf4b6133c5cbd2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a21a97385f856ec996105b597d0ae81c56ed28b625ef3f2905e1af6ec8f6a45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://646e01a58908ba2e703aae5eef53e1b981b0f0c3d1d3202487f2a8b94ee3f55d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31752b84d86bda30792b29e4c6ed963753cc576c5de784395e8e6dcbebaced47\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"ver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:40:26.813730 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814130 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:40:26.814139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:40:26.814143 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:40:26.814145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 03:40:26.821403 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821442 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821482 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821496 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 03:40:26.821509 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821514 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 03:40:26.821612 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 03:40:26.821619 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1206 03:40:26.822049 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16a92d501c9e626fe91e4a6793e426d10eb292cfbab766db8b8cc72e5dae94ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:07Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.604331 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd230730-a9a1-47ed-b625-25e32231f271\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce321001135d0a2b878316f8003bf1bc54cba9a36f6a654ba376e1e7c0b4da71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af83814e6b2c5ed803157308ab895551e0826fd02f1ecc8b2751145c09d359da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9e6b48b5701dd330ef91712cdcd160afe9831a3881038bfb1e0397a9c8a3aa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2d359853c738e3e0e264f1a6a919b17afc71a362821b3ea170cda6695e3593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2d359853c738e3e0e264f1a6a919b17afc71a362821b3ea170cda6695e3593f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:07Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.614282 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-b2knv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a3ae46f-03b5-4999-9d58-c38ee222f68a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76665d4c127445009fd66078827b54607c9ef812c8011d1c0b4cde1c0516853a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk4hj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-b2knv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:07Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.625523 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.625546 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.625554 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.625567 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.625576 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:07Z","lastTransitionTime":"2025-12-06T03:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.629827 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22rrq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03112a5d-5882-4fac-9ae5-13525ff82fe0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0205b089b12c42688203bb7ec0f27dc5f9a577dabdb446b811f184046f85dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trlt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22rrq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:07Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.642451 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ngspk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34808ab2-d21b-40b6-8fe3-e18bf9e5fc46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52509d21f5b4bfbd458162cba176ff82c627c7c4671204addf9d76f7392fd760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qkf8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9685a4d5d4f40ff4f59b6da6193785f340890453d1e422d3421a621dc7bd02f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qkf8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ngspk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:07Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.653595 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5ltx2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8881599e-71e5-4578-b3f9-21f4fa2f57aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbm6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbm6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5ltx2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:07Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.669900 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2466610962332c6ca120fe055e4e8a3fc623088119dc1a412e06b745176f89e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:07Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.683357 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:07Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.695254 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6df38316-e0d3-4018-8d27-3620eba3a68d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57381e9589a2dc04a4d2733c37985f42594182128a374b7be0f471ea69138b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bf5e41701c7a2d45adc4740ceb4ac1785db17bc7cb15daed439f7556812d1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpxxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:07Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.708243 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kj7hf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd46dfa5-7d4a-49ed-8199-523f10cebdf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5eee8a0cf96ceb3785dac9df31bf8d31715d1c88fbea17c5e582b0a4554bb63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmkcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kj7hf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:07Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.728957 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.728996 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.729011 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.729036 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.729055 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:07Z","lastTransitionTime":"2025-12-06T03:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.831939 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.832254 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.832452 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.832620 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.832801 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:07Z","lastTransitionTime":"2025-12-06T03:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.934995 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.935081 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.935105 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.935132 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:07 crc kubenswrapper[4802]: I1206 03:41:07.935154 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:07Z","lastTransitionTime":"2025-12-06T03:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:08 crc kubenswrapper[4802]: I1206 03:41:08.038170 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:08 crc kubenswrapper[4802]: I1206 03:41:08.038526 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:08 crc kubenswrapper[4802]: I1206 03:41:08.038723 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:08 crc kubenswrapper[4802]: I1206 03:41:08.039016 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:08 crc kubenswrapper[4802]: I1206 03:41:08.039228 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:08Z","lastTransitionTime":"2025-12-06T03:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:08 crc kubenswrapper[4802]: I1206 03:41:08.142158 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:08 crc kubenswrapper[4802]: I1206 03:41:08.142211 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:08 crc kubenswrapper[4802]: I1206 03:41:08.142234 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:08 crc kubenswrapper[4802]: I1206 03:41:08.142265 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:08 crc kubenswrapper[4802]: I1206 03:41:08.142286 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:08Z","lastTransitionTime":"2025-12-06T03:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:08 crc kubenswrapper[4802]: I1206 03:41:08.244616 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:08 crc kubenswrapper[4802]: I1206 03:41:08.244679 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:08 crc kubenswrapper[4802]: I1206 03:41:08.244701 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:08 crc kubenswrapper[4802]: I1206 03:41:08.244726 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:08 crc kubenswrapper[4802]: I1206 03:41:08.244746 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:08Z","lastTransitionTime":"2025-12-06T03:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:08 crc kubenswrapper[4802]: I1206 03:41:08.348084 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:08 crc kubenswrapper[4802]: I1206 03:41:08.348166 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:08 crc kubenswrapper[4802]: I1206 03:41:08.348193 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:08 crc kubenswrapper[4802]: I1206 03:41:08.348227 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:08 crc kubenswrapper[4802]: I1206 03:41:08.348252 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:08Z","lastTransitionTime":"2025-12-06T03:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:08 crc kubenswrapper[4802]: I1206 03:41:08.449412 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:41:08 crc kubenswrapper[4802]: I1206 03:41:08.449494 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:41:08 crc kubenswrapper[4802]: I1206 03:41:08.449526 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:41:08 crc kubenswrapper[4802]: I1206 03:41:08.449441 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:41:08 crc kubenswrapper[4802]: E1206 03:41:08.449631 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5ltx2" podUID="8881599e-71e5-4578-b3f9-21f4fa2f57aa" Dec 06 03:41:08 crc kubenswrapper[4802]: E1206 03:41:08.450033 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:41:08 crc kubenswrapper[4802]: E1206 03:41:08.450140 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:41:08 crc kubenswrapper[4802]: E1206 03:41:08.450535 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:41:08 crc kubenswrapper[4802]: I1206 03:41:08.451478 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:08 crc kubenswrapper[4802]: I1206 03:41:08.451537 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:08 crc kubenswrapper[4802]: I1206 03:41:08.451557 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:08 crc kubenswrapper[4802]: I1206 03:41:08.451579 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:08 crc kubenswrapper[4802]: I1206 03:41:08.451598 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:08Z","lastTransitionTime":"2025-12-06T03:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:08 crc kubenswrapper[4802]: I1206 03:41:08.554256 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:08 crc kubenswrapper[4802]: I1206 03:41:08.554314 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:08 crc kubenswrapper[4802]: I1206 03:41:08.554326 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:08 crc kubenswrapper[4802]: I1206 03:41:08.554345 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:08 crc kubenswrapper[4802]: I1206 03:41:08.554358 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:08Z","lastTransitionTime":"2025-12-06T03:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:08 crc kubenswrapper[4802]: I1206 03:41:08.657987 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:08 crc kubenswrapper[4802]: I1206 03:41:08.658071 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:08 crc kubenswrapper[4802]: I1206 03:41:08.658095 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:08 crc kubenswrapper[4802]: I1206 03:41:08.658119 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:08 crc kubenswrapper[4802]: I1206 03:41:08.658140 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:08Z","lastTransitionTime":"2025-12-06T03:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:08 crc kubenswrapper[4802]: I1206 03:41:08.761597 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:08 crc kubenswrapper[4802]: I1206 03:41:08.761657 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:08 crc kubenswrapper[4802]: I1206 03:41:08.761675 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:08 crc kubenswrapper[4802]: I1206 03:41:08.761700 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:08 crc kubenswrapper[4802]: I1206 03:41:08.761717 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:08Z","lastTransitionTime":"2025-12-06T03:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:08 crc kubenswrapper[4802]: I1206 03:41:08.864913 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:08 crc kubenswrapper[4802]: I1206 03:41:08.864983 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:08 crc kubenswrapper[4802]: I1206 03:41:08.865010 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:08 crc kubenswrapper[4802]: I1206 03:41:08.865041 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:08 crc kubenswrapper[4802]: I1206 03:41:08.865068 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:08Z","lastTransitionTime":"2025-12-06T03:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:08 crc kubenswrapper[4802]: I1206 03:41:08.970389 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:08 crc kubenswrapper[4802]: I1206 03:41:08.970457 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:08 crc kubenswrapper[4802]: I1206 03:41:08.970481 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:08 crc kubenswrapper[4802]: I1206 03:41:08.970528 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:08 crc kubenswrapper[4802]: I1206 03:41:08.970552 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:08Z","lastTransitionTime":"2025-12-06T03:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.073714 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.073822 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.073841 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.073865 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.073882 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:09Z","lastTransitionTime":"2025-12-06T03:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.178448 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.178524 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.178548 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.178577 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.178601 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:09Z","lastTransitionTime":"2025-12-06T03:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.281559 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.281623 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.281641 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.281664 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.281680 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:09Z","lastTransitionTime":"2025-12-06T03:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.384544 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.384595 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.384618 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.384651 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.384676 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:09Z","lastTransitionTime":"2025-12-06T03:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.487577 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.487663 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.487708 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.487740 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.487805 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:09Z","lastTransitionTime":"2025-12-06T03:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.590879 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.590942 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.590961 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.590986 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.591004 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:09Z","lastTransitionTime":"2025-12-06T03:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.693829 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.693911 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.693936 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.693972 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.693996 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:09Z","lastTransitionTime":"2025-12-06T03:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.797088 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.797148 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.797171 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.797201 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.797226 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:09Z","lastTransitionTime":"2025-12-06T03:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.818382 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.818444 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.818465 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.818491 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.818508 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:09Z","lastTransitionTime":"2025-12-06T03:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:09 crc kubenswrapper[4802]: E1206 03:41:09.839028 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0e01ca3a-758f-44bc-b4a2-30c1bc04fda0\\\",\\\"systemUUID\\\":\\\"c7c815e4-d2a4-4244-b035-f988fc95e215\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:09Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.844840 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.844902 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.844920 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.844946 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.844962 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:09Z","lastTransitionTime":"2025-12-06T03:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:09 crc kubenswrapper[4802]: E1206 03:41:09.865255 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0e01ca3a-758f-44bc-b4a2-30c1bc04fda0\\\",\\\"systemUUID\\\":\\\"c7c815e4-d2a4-4244-b035-f988fc95e215\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:09Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.871312 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.871577 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.871806 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.872024 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.872216 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:09Z","lastTransitionTime":"2025-12-06T03:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:09 crc kubenswrapper[4802]: E1206 03:41:09.892450 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0e01ca3a-758f-44bc-b4a2-30c1bc04fda0\\\",\\\"systemUUID\\\":\\\"c7c815e4-d2a4-4244-b035-f988fc95e215\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:09Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.897400 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.897441 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.897452 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.897471 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.897481 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:09Z","lastTransitionTime":"2025-12-06T03:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:09 crc kubenswrapper[4802]: E1206 03:41:09.916868 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0e01ca3a-758f-44bc-b4a2-30c1bc04fda0\\\",\\\"systemUUID\\\":\\\"c7c815e4-d2a4-4244-b035-f988fc95e215\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:09Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.921907 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.921946 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.921960 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.921975 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.921987 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:09Z","lastTransitionTime":"2025-12-06T03:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:09 crc kubenswrapper[4802]: E1206 03:41:09.941978 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0e01ca3a-758f-44bc-b4a2-30c1bc04fda0\\\",\\\"systemUUID\\\":\\\"c7c815e4-d2a4-4244-b035-f988fc95e215\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:09Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:09 crc kubenswrapper[4802]: E1206 03:41:09.942499 4802 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.944195 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.944254 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.944272 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.944295 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:09 crc kubenswrapper[4802]: I1206 03:41:09.944311 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:09Z","lastTransitionTime":"2025-12-06T03:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.047419 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.047505 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.047533 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.047565 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.047589 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:10Z","lastTransitionTime":"2025-12-06T03:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.150853 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.150945 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.150970 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.150998 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.151016 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:10Z","lastTransitionTime":"2025-12-06T03:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.254424 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.254494 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.254514 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.254541 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.254563 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:10Z","lastTransitionTime":"2025-12-06T03:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.356729 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.356791 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.356817 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.356834 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.356847 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:10Z","lastTransitionTime":"2025-12-06T03:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.449048 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.449217 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.449340 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:41:10 crc kubenswrapper[4802]: E1206 03:41:10.449346 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.449382 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:41:10 crc kubenswrapper[4802]: E1206 03:41:10.449971 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:41:10 crc kubenswrapper[4802]: E1206 03:41:10.450195 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5ltx2" podUID="8881599e-71e5-4578-b3f9-21f4fa2f57aa" Dec 06 03:41:10 crc kubenswrapper[4802]: E1206 03:41:10.450313 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.450646 4802 scope.go:117] "RemoveContainer" containerID="3870b61bce487c69724f668c38fc02190ff97b5e3b1434d66ed624412d18319b" Dec 06 03:41:10 crc kubenswrapper[4802]: E1206 03:41:10.451216 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pvm6q_openshift-ovn-kubernetes(eafce4e0-e7fb-4877-b0ab-3283829ba1ef)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" podUID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.460296 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.460498 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.460585 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.460696 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.460815 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:10Z","lastTransitionTime":"2025-12-06T03:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.469888 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd230730-a9a1-47ed-b625-25e32231f271\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce321001135d0a2b878316f8003bf1bc54cba9a36f6a654ba376e1e7c0b4da71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af83814e6b2c5ed803157308ab895551e0826fd02f1ecc8b2751145c09d359da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9e6b48b5701dd330ef91712cdcd160afe9831a3881038bfb1e0397a9c8a3aa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2d359853c738e3e0e264f1a6a919b17afc71a362821b3ea170cda6695e3593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2d359853c738e3e0e264f1a6a919b17afc71a362821b3ea170cda6695e3593f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:10Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.486512 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-b2knv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a3ae46f-03b5-4999-9d58-c38ee222f68a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76665d4c127445009fd66078827b54607c9ef812c8011d1c0b4cde1c0516853a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk4hj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-b2knv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:10Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.510163 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22rrq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03112a5d-5882-4fac-9ae5-13525ff82fe0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0205b089b12c42688203bb7ec0f27dc5f9a577dabdb446b811f184046f85dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trlt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22rrq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:10Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.525665 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ngspk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34808ab2-d21b-40b6-8fe3-e18bf9e5fc46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52509d21f5b4bfbd458162cba176ff82c627c7c4671204addf9d76f7392fd760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qkf8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9685a4d5d4f40ff4f59b6da6193785f340890453d1e422d3421a621dc7bd02f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qkf8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ngspk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:10Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.541952 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5ltx2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8881599e-71e5-4578-b3f9-21f4fa2f57aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbm6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbm6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5ltx2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:10Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.562211 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1721e985-40bd-4b2d-a0e8-d4e365094ba6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a895e0cc04880c3287ba725ac5484835235f1e066a94ce98fcf4b6133c5cbd2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a21a97385f856ec996105b597d0ae81c56ed28b625ef3f2905e1af6ec8f6a45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://646e01a58908ba2e703aae5eef53e1b981b0f0c3d1d3202487f2a8b94ee3f55d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31752b84d86bda30792b29e4c6ed963753cc576c5de784395e8e6dcbebaced47\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"ver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:40:26.813730 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814130 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:40:26.814139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:40:26.814143 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:40:26.814145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 03:40:26.821403 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821442 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821482 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821496 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 03:40:26.821509 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821514 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 03:40:26.821612 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 03:40:26.821619 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1206 03:40:26.822049 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16a92d501c9e626fe91e4a6793e426d10eb292cfbab766db8b8cc72e5dae94ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:10Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.564356 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.564424 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.564445 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.564472 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.564493 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:10Z","lastTransitionTime":"2025-12-06T03:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.580036 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6df38316-e0d3-4018-8d27-3620eba3a68d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57381e9589a2dc04a4d2733c37985f42594182128a374b7be0f471ea69138b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bf5e41701c7a2d45adc4740ceb4ac1785db17bc7cb15daed439f7556812d1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpxxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:10Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.595556 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kj7hf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd46dfa5-7d4a-49ed-8199-523f10cebdf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5eee8a0cf96ceb3785dac9df31bf8d31715d1c88fbea17c5e582b0a4554bb63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmkcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kj7hf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:10Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.612547 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2466610962332c6ca120fe055e4e8a3fc623088119dc1a412e06b745176f89e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:10Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.628408 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:10Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.649130 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:10Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.667365 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.667434 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.667458 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.667488 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.667512 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:10Z","lastTransitionTime":"2025-12-06T03:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.671680 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fad9b63d84f41fccf696cd313d7ca7f72b3d3251d6c1b7341573a3fadf1e378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:10Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.692948 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb3f278fff47c900c56d8a563d5ed75a2ae84e2034b0f45aad4083790b9e7946\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b10ca16ac26904ec5e6245344b71d836888cae9b069cae6a7a4831d609752db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:10Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.714287 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2093b6-c4b5-4db8-be16-fbc859885e8d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0d6e582161f0d9126126e03b27211be22f6b5f797154928c92fa13a8118f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03004a29aefe7fe2452fb9f6663c0807afccb055d7ddbf6b907f305e6a19c81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1801ac64db4ba91c2e8a249e3651d3e50ab62e9f2c87312ce9bf7f38518a82cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6e454507cadbb982a566a3692995aaa3c733ff1cd7fb4870476602190e1eeba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:10Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.739127 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7730de4f033561ad77046e028cd1a3f0bfe2775387cbdfa35f47ea79909ec4cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a158aebd701ef46225d3e37109ae28ab58d107087bf06650c4b6d0055a3bbba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2357f37497fde64937eca7b38693893af745ab278653ebc79406e122a6d4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8941bd25107333b264d3f081a811fdfeea23ad92193fc61367665fd2cd1209c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f9cd4dc75e63f09f60b48b6da284643910bb32ba51f2a808cb4d9803f30042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6796f65ee9c5bd0c18f205697135d17677c5daf3ba378cc992cba795e0188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3870b61bce487c69724f668c38fc02190ff97b5e3b1434d66ed624412d18319b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3870b61bce487c69724f668c38fc02190ff97b5e3b1434d66ed624412d18319b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:40:54Z\\\",\\\"message\\\":\\\"/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 03:40:54.580644 6463 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 03:40:54.580730 6463 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 03:40:54.580797 6463 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 03:40:54.580836 6463 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 03:40:54.580845 6463 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 03:40:54.580895 6463 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1206 03:40:54.580943 6463 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1206 03:40:54.580950 6463 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1206 03:40:54.581001 6463 factory.go:656] Stopping watch factory\\\\nI1206 03:40:54.581036 6463 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 03:40:54.581050 6463 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 03:40:54.581063 6463 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 03:40:54.581074 6463 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 03:40:54.581085 6463 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1206 03:40:54.581102 6463 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:53Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pvm6q_openshift-ovn-kubernetes(eafce4e0-e7fb-4877-b0ab-3283829ba1ef)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3cb210bfc97a439a7ff98c02c59f8ad16bc400dc214a616a44b43a8b01e2782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pvm6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:10Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.753733 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:10Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.770823 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.770880 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.770896 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.770917 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.770931 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:10Z","lastTransitionTime":"2025-12-06T03:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.771704 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtw2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3a3bac2-3923-4507-a4de-1556b3ad473e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fed5ee4189d54456a3216f4fdb1ecf003fa47e29dbfe4b2b2f301edfb897d6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtw2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:10Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.873905 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.873961 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.873979 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.874003 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.874022 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:10Z","lastTransitionTime":"2025-12-06T03:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.976270 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.976312 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.976323 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.976341 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:10 crc kubenswrapper[4802]: I1206 03:41:10.976355 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:10Z","lastTransitionTime":"2025-12-06T03:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:11 crc kubenswrapper[4802]: I1206 03:41:11.078561 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:11 crc kubenswrapper[4802]: I1206 03:41:11.078597 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:11 crc kubenswrapper[4802]: I1206 03:41:11.078609 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:11 crc kubenswrapper[4802]: I1206 03:41:11.078625 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:11 crc kubenswrapper[4802]: I1206 03:41:11.078639 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:11Z","lastTransitionTime":"2025-12-06T03:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:11 crc kubenswrapper[4802]: I1206 03:41:11.181077 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:11 crc kubenswrapper[4802]: I1206 03:41:11.181116 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:11 crc kubenswrapper[4802]: I1206 03:41:11.181127 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:11 crc kubenswrapper[4802]: I1206 03:41:11.181142 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:11 crc kubenswrapper[4802]: I1206 03:41:11.181154 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:11Z","lastTransitionTime":"2025-12-06T03:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:11 crc kubenswrapper[4802]: I1206 03:41:11.283541 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:11 crc kubenswrapper[4802]: I1206 03:41:11.283603 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:11 crc kubenswrapper[4802]: I1206 03:41:11.283621 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:11 crc kubenswrapper[4802]: I1206 03:41:11.283648 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:11 crc kubenswrapper[4802]: I1206 03:41:11.283665 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:11Z","lastTransitionTime":"2025-12-06T03:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:11 crc kubenswrapper[4802]: I1206 03:41:11.386017 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:11 crc kubenswrapper[4802]: I1206 03:41:11.386085 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:11 crc kubenswrapper[4802]: I1206 03:41:11.386132 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:11 crc kubenswrapper[4802]: I1206 03:41:11.386151 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:11 crc kubenswrapper[4802]: I1206 03:41:11.386167 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:11Z","lastTransitionTime":"2025-12-06T03:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:11 crc kubenswrapper[4802]: I1206 03:41:11.488156 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:11 crc kubenswrapper[4802]: I1206 03:41:11.488197 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:11 crc kubenswrapper[4802]: I1206 03:41:11.488214 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:11 crc kubenswrapper[4802]: I1206 03:41:11.488233 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:11 crc kubenswrapper[4802]: I1206 03:41:11.488245 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:11Z","lastTransitionTime":"2025-12-06T03:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:11 crc kubenswrapper[4802]: I1206 03:41:11.590419 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:11 crc kubenswrapper[4802]: I1206 03:41:11.590459 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:11 crc kubenswrapper[4802]: I1206 03:41:11.590469 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:11 crc kubenswrapper[4802]: I1206 03:41:11.590510 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:11 crc kubenswrapper[4802]: I1206 03:41:11.590524 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:11Z","lastTransitionTime":"2025-12-06T03:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:11 crc kubenswrapper[4802]: I1206 03:41:11.693493 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:11 crc kubenswrapper[4802]: I1206 03:41:11.693529 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:11 crc kubenswrapper[4802]: I1206 03:41:11.693540 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:11 crc kubenswrapper[4802]: I1206 03:41:11.693555 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:11 crc kubenswrapper[4802]: I1206 03:41:11.693566 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:11Z","lastTransitionTime":"2025-12-06T03:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:11 crc kubenswrapper[4802]: I1206 03:41:11.796286 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:11 crc kubenswrapper[4802]: I1206 03:41:11.796335 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:11 crc kubenswrapper[4802]: I1206 03:41:11.796349 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:11 crc kubenswrapper[4802]: I1206 03:41:11.796369 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:11 crc kubenswrapper[4802]: I1206 03:41:11.796380 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:11Z","lastTransitionTime":"2025-12-06T03:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:11 crc kubenswrapper[4802]: I1206 03:41:11.898782 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:11 crc kubenswrapper[4802]: I1206 03:41:11.898821 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:11 crc kubenswrapper[4802]: I1206 03:41:11.898830 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:11 crc kubenswrapper[4802]: I1206 03:41:11.898843 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:11 crc kubenswrapper[4802]: I1206 03:41:11.898853 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:11Z","lastTransitionTime":"2025-12-06T03:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:12 crc kubenswrapper[4802]: I1206 03:41:12.001155 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:12 crc kubenswrapper[4802]: I1206 03:41:12.001229 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:12 crc kubenswrapper[4802]: I1206 03:41:12.001250 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:12 crc kubenswrapper[4802]: I1206 03:41:12.001276 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:12 crc kubenswrapper[4802]: I1206 03:41:12.001292 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:12Z","lastTransitionTime":"2025-12-06T03:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:12 crc kubenswrapper[4802]: I1206 03:41:12.104102 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:12 crc kubenswrapper[4802]: I1206 03:41:12.104142 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:12 crc kubenswrapper[4802]: I1206 03:41:12.104158 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:12 crc kubenswrapper[4802]: I1206 03:41:12.104192 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:12 crc kubenswrapper[4802]: I1206 03:41:12.104210 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:12Z","lastTransitionTime":"2025-12-06T03:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:12 crc kubenswrapper[4802]: I1206 03:41:12.207618 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:12 crc kubenswrapper[4802]: I1206 03:41:12.207704 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:12 crc kubenswrapper[4802]: I1206 03:41:12.207724 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:12 crc kubenswrapper[4802]: I1206 03:41:12.207779 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:12 crc kubenswrapper[4802]: I1206 03:41:12.207803 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:12Z","lastTransitionTime":"2025-12-06T03:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:12 crc kubenswrapper[4802]: I1206 03:41:12.310226 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:12 crc kubenswrapper[4802]: I1206 03:41:12.310279 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:12 crc kubenswrapper[4802]: I1206 03:41:12.310306 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:12 crc kubenswrapper[4802]: I1206 03:41:12.310329 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:12 crc kubenswrapper[4802]: I1206 03:41:12.310345 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:12Z","lastTransitionTime":"2025-12-06T03:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:12 crc kubenswrapper[4802]: I1206 03:41:12.412545 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:12 crc kubenswrapper[4802]: I1206 03:41:12.412596 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:12 crc kubenswrapper[4802]: I1206 03:41:12.412614 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:12 crc kubenswrapper[4802]: I1206 03:41:12.412637 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:12 crc kubenswrapper[4802]: I1206 03:41:12.412653 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:12Z","lastTransitionTime":"2025-12-06T03:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:12 crc kubenswrapper[4802]: I1206 03:41:12.449894 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:41:12 crc kubenswrapper[4802]: I1206 03:41:12.449918 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:41:12 crc kubenswrapper[4802]: I1206 03:41:12.449926 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:41:12 crc kubenswrapper[4802]: E1206 03:41:12.450022 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:41:12 crc kubenswrapper[4802]: I1206 03:41:12.450102 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:41:12 crc kubenswrapper[4802]: E1206 03:41:12.450244 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5ltx2" podUID="8881599e-71e5-4578-b3f9-21f4fa2f57aa" Dec 06 03:41:12 crc kubenswrapper[4802]: E1206 03:41:12.450286 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:41:12 crc kubenswrapper[4802]: E1206 03:41:12.450430 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:41:12 crc kubenswrapper[4802]: I1206 03:41:12.515268 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:12 crc kubenswrapper[4802]: I1206 03:41:12.515311 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:12 crc kubenswrapper[4802]: I1206 03:41:12.515323 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:12 crc kubenswrapper[4802]: I1206 03:41:12.515342 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:12 crc kubenswrapper[4802]: I1206 03:41:12.515357 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:12Z","lastTransitionTime":"2025-12-06T03:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:12 crc kubenswrapper[4802]: I1206 03:41:12.617305 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:12 crc kubenswrapper[4802]: I1206 03:41:12.617343 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:12 crc kubenswrapper[4802]: I1206 03:41:12.617351 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:12 crc kubenswrapper[4802]: I1206 03:41:12.617365 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:12 crc kubenswrapper[4802]: I1206 03:41:12.617374 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:12Z","lastTransitionTime":"2025-12-06T03:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:12 crc kubenswrapper[4802]: I1206 03:41:12.719853 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:12 crc kubenswrapper[4802]: I1206 03:41:12.719912 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:12 crc kubenswrapper[4802]: I1206 03:41:12.719931 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:12 crc kubenswrapper[4802]: I1206 03:41:12.719956 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:12 crc kubenswrapper[4802]: I1206 03:41:12.719975 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:12Z","lastTransitionTime":"2025-12-06T03:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:12 crc kubenswrapper[4802]: I1206 03:41:12.822649 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:12 crc kubenswrapper[4802]: I1206 03:41:12.822716 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:12 crc kubenswrapper[4802]: I1206 03:41:12.822733 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:12 crc kubenswrapper[4802]: I1206 03:41:12.822783 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:12 crc kubenswrapper[4802]: I1206 03:41:12.822805 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:12Z","lastTransitionTime":"2025-12-06T03:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:12 crc kubenswrapper[4802]: I1206 03:41:12.897421 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8881599e-71e5-4578-b3f9-21f4fa2f57aa-metrics-certs\") pod \"network-metrics-daemon-5ltx2\" (UID: \"8881599e-71e5-4578-b3f9-21f4fa2f57aa\") " pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:41:12 crc kubenswrapper[4802]: E1206 03:41:12.897637 4802 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 03:41:12 crc kubenswrapper[4802]: E1206 03:41:12.897737 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8881599e-71e5-4578-b3f9-21f4fa2f57aa-metrics-certs podName:8881599e-71e5-4578-b3f9-21f4fa2f57aa nodeName:}" failed. No retries permitted until 2025-12-06 03:41:44.897713558 +0000 UTC m=+97.769622750 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8881599e-71e5-4578-b3f9-21f4fa2f57aa-metrics-certs") pod "network-metrics-daemon-5ltx2" (UID: "8881599e-71e5-4578-b3f9-21f4fa2f57aa") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 03:41:12 crc kubenswrapper[4802]: I1206 03:41:12.924984 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:12 crc kubenswrapper[4802]: I1206 03:41:12.925043 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:12 crc kubenswrapper[4802]: I1206 03:41:12.925062 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:12 crc kubenswrapper[4802]: I1206 03:41:12.925084 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:12 crc kubenswrapper[4802]: I1206 03:41:12.925101 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:12Z","lastTransitionTime":"2025-12-06T03:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:13 crc kubenswrapper[4802]: I1206 03:41:13.028288 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:13 crc kubenswrapper[4802]: I1206 03:41:13.028333 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:13 crc kubenswrapper[4802]: I1206 03:41:13.028343 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:13 crc kubenswrapper[4802]: I1206 03:41:13.028356 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:13 crc kubenswrapper[4802]: I1206 03:41:13.028366 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:13Z","lastTransitionTime":"2025-12-06T03:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:13 crc kubenswrapper[4802]: I1206 03:41:13.131135 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:13 crc kubenswrapper[4802]: I1206 03:41:13.131190 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:13 crc kubenswrapper[4802]: I1206 03:41:13.131207 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:13 crc kubenswrapper[4802]: I1206 03:41:13.131229 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:13 crc kubenswrapper[4802]: I1206 03:41:13.131247 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:13Z","lastTransitionTime":"2025-12-06T03:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:13 crc kubenswrapper[4802]: I1206 03:41:13.233006 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:13 crc kubenswrapper[4802]: I1206 03:41:13.233033 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:13 crc kubenswrapper[4802]: I1206 03:41:13.233043 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:13 crc kubenswrapper[4802]: I1206 03:41:13.233056 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:13 crc kubenswrapper[4802]: I1206 03:41:13.233067 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:13Z","lastTransitionTime":"2025-12-06T03:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:13 crc kubenswrapper[4802]: I1206 03:41:13.335674 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:13 crc kubenswrapper[4802]: I1206 03:41:13.335727 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:13 crc kubenswrapper[4802]: I1206 03:41:13.335737 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:13 crc kubenswrapper[4802]: I1206 03:41:13.335763 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:13 crc kubenswrapper[4802]: I1206 03:41:13.335772 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:13Z","lastTransitionTime":"2025-12-06T03:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:13 crc kubenswrapper[4802]: I1206 03:41:13.438253 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:13 crc kubenswrapper[4802]: I1206 03:41:13.438290 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:13 crc kubenswrapper[4802]: I1206 03:41:13.438301 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:13 crc kubenswrapper[4802]: I1206 03:41:13.438316 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:13 crc kubenswrapper[4802]: I1206 03:41:13.438327 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:13Z","lastTransitionTime":"2025-12-06T03:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:13 crc kubenswrapper[4802]: I1206 03:41:13.540805 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:13 crc kubenswrapper[4802]: I1206 03:41:13.540848 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:13 crc kubenswrapper[4802]: I1206 03:41:13.540861 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:13 crc kubenswrapper[4802]: I1206 03:41:13.540878 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:13 crc kubenswrapper[4802]: I1206 03:41:13.540890 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:13Z","lastTransitionTime":"2025-12-06T03:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:13 crc kubenswrapper[4802]: I1206 03:41:13.642985 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:13 crc kubenswrapper[4802]: I1206 03:41:13.643024 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:13 crc kubenswrapper[4802]: I1206 03:41:13.643034 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:13 crc kubenswrapper[4802]: I1206 03:41:13.643050 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:13 crc kubenswrapper[4802]: I1206 03:41:13.643060 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:13Z","lastTransitionTime":"2025-12-06T03:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:13 crc kubenswrapper[4802]: I1206 03:41:13.745224 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:13 crc kubenswrapper[4802]: I1206 03:41:13.745281 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:13 crc kubenswrapper[4802]: I1206 03:41:13.745299 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:13 crc kubenswrapper[4802]: I1206 03:41:13.745323 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:13 crc kubenswrapper[4802]: I1206 03:41:13.745340 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:13Z","lastTransitionTime":"2025-12-06T03:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:13 crc kubenswrapper[4802]: I1206 03:41:13.847399 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:13 crc kubenswrapper[4802]: I1206 03:41:13.847436 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:13 crc kubenswrapper[4802]: I1206 03:41:13.847448 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:13 crc kubenswrapper[4802]: I1206 03:41:13.847464 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:13 crc kubenswrapper[4802]: I1206 03:41:13.847476 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:13Z","lastTransitionTime":"2025-12-06T03:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:13 crc kubenswrapper[4802]: I1206 03:41:13.949832 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:13 crc kubenswrapper[4802]: I1206 03:41:13.949866 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:13 crc kubenswrapper[4802]: I1206 03:41:13.949877 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:13 crc kubenswrapper[4802]: I1206 03:41:13.949894 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:13 crc kubenswrapper[4802]: I1206 03:41:13.949905 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:13Z","lastTransitionTime":"2025-12-06T03:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.052302 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.052335 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.052344 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.052360 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.052369 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:14Z","lastTransitionTime":"2025-12-06T03:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.154635 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.154678 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.154689 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.154704 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.154714 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:14Z","lastTransitionTime":"2025-12-06T03:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.257051 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.257096 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.257107 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.257125 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.257137 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:14Z","lastTransitionTime":"2025-12-06T03:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.359871 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.359908 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.359919 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.359934 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.359945 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:14Z","lastTransitionTime":"2025-12-06T03:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.449516 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.449559 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.449614 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.449582 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:41:14 crc kubenswrapper[4802]: E1206 03:41:14.449710 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:41:14 crc kubenswrapper[4802]: E1206 03:41:14.449797 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5ltx2" podUID="8881599e-71e5-4578-b3f9-21f4fa2f57aa" Dec 06 03:41:14 crc kubenswrapper[4802]: E1206 03:41:14.449862 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:41:14 crc kubenswrapper[4802]: E1206 03:41:14.449900 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.462025 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.462043 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.462052 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.462066 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.462077 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:14Z","lastTransitionTime":"2025-12-06T03:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.565385 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.565447 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.565465 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.565491 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.565508 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:14Z","lastTransitionTime":"2025-12-06T03:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.667569 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.667629 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.667645 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.667673 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.667693 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:14Z","lastTransitionTime":"2025-12-06T03:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.769798 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.769834 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.769843 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.769858 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.769867 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:14Z","lastTransitionTime":"2025-12-06T03:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.872444 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.872495 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.872508 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.872526 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.872538 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:14Z","lastTransitionTime":"2025-12-06T03:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.921987 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-22rrq_03112a5d-5882-4fac-9ae5-13525ff82fe0/kube-multus/0.log" Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.922044 4802 generic.go:334] "Generic (PLEG): container finished" podID="03112a5d-5882-4fac-9ae5-13525ff82fe0" containerID="c0205b089b12c42688203bb7ec0f27dc5f9a577dabdb446b811f184046f85dc3" exitCode=1 Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.922080 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-22rrq" event={"ID":"03112a5d-5882-4fac-9ae5-13525ff82fe0","Type":"ContainerDied","Data":"c0205b089b12c42688203bb7ec0f27dc5f9a577dabdb446b811f184046f85dc3"} Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.922490 4802 scope.go:117] "RemoveContainer" containerID="c0205b089b12c42688203bb7ec0f27dc5f9a577dabdb446b811f184046f85dc3" Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.942045 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2466610962332c6ca120fe055e4e8a3fc623088119dc1a412e06b745176f89e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:14Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.961146 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:14Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.970588 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6df38316-e0d3-4018-8d27-3620eba3a68d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57381e9589a2dc04a4d2733c37985f42594182128a374b7be0f471ea69138b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bf5e41701c7a2d45adc4740ceb4ac1785db17bc7cb15daed439f7556812d1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpxxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:14Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.974986 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.975016 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.975026 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.975041 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.975052 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:14Z","lastTransitionTime":"2025-12-06T03:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.981076 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kj7hf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd46dfa5-7d4a-49ed-8199-523f10cebdf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5eee8a0cf96ceb3785dac9df31bf8d31715d1c88fbea17c5e582b0a4554bb63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmkcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kj7hf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:14Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:14 crc kubenswrapper[4802]: I1206 03:41:14.992620 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb3f278fff47c900c56d8a563d5ed75a2ae84e2034b0f45aad4083790b9e7946\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b10ca16ac26904ec5e6245344b71d836888cae9b069cae6a7a4831d609752db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:14Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.003602 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2093b6-c4b5-4db8-be16-fbc859885e8d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0d6e582161f0d9126126e03b27211be22f6b5f797154928c92fa13a8118f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03004a29aefe7fe2452fb9f6663c0807afccb055d7ddbf6b907f305e6a19c81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1801ac64db4ba91c2e8a249e3651d3e50ab62e9f2c87312ce9bf7f38518a82cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6e454507cadbb982a566a3692995aaa3c733ff1cd7fb4870476602190e1eeba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:15Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.014281 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:15Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.025436 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fad9b63d84f41fccf696cd313d7ca7f72b3d3251d6c1b7341573a3fadf1e378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:15Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.036474 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:15Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.050903 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtw2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3a3bac2-3923-4507-a4de-1556b3ad473e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fed5ee4189d54456a3216f4fdb1ecf003fa47e29dbfe4b2b2f301edfb897d6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtw2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:15Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.068010 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7730de4f033561ad77046e028cd1a3f0bfe2775387cbdfa35f47ea79909ec4cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a158aebd701ef46225d3e37109ae28ab58d107087bf06650c4b6d0055a3bbba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2357f37497fde64937eca7b38693893af745ab278653ebc79406e122a6d4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8941bd25107333b264d3f081a811fdfeea23ad92193fc61367665fd2cd1209c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f9cd4dc75e63f09f60b48b6da284643910bb32ba51f2a808cb4d9803f30042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6796f65ee9c5bd0c18f205697135d17677c5daf3ba378cc992cba795e0188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3870b61bce487c69724f668c38fc02190ff97b5e3b1434d66ed624412d18319b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3870b61bce487c69724f668c38fc02190ff97b5e3b1434d66ed624412d18319b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:40:54Z\\\",\\\"message\\\":\\\"/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 03:40:54.580644 6463 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 03:40:54.580730 6463 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 03:40:54.580797 6463 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 03:40:54.580836 6463 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 03:40:54.580845 6463 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 03:40:54.580895 6463 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1206 03:40:54.580943 6463 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1206 03:40:54.580950 6463 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1206 03:40:54.581001 6463 factory.go:656] Stopping watch factory\\\\nI1206 03:40:54.581036 6463 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 03:40:54.581050 6463 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 03:40:54.581063 6463 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 03:40:54.581074 6463 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 03:40:54.581085 6463 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1206 03:40:54.581102 6463 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:53Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pvm6q_openshift-ovn-kubernetes(eafce4e0-e7fb-4877-b0ab-3283829ba1ef)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3cb210bfc97a439a7ff98c02c59f8ad16bc400dc214a616a44b43a8b01e2782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pvm6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:15Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.076864 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.076897 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.076906 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.076918 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.076929 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:15Z","lastTransitionTime":"2025-12-06T03:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.081646 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22rrq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03112a5d-5882-4fac-9ae5-13525ff82fe0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0205b089b12c42688203bb7ec0f27dc5f9a577dabdb446b811f184046f85dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0205b089b12c42688203bb7ec0f27dc5f9a577dabdb446b811f184046f85dc3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:41:14Z\\\",\\\"message\\\":\\\"2025-12-06T03:40:28+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_384cd081-7034-46cf-bb48-841728494c60\\\\n2025-12-06T03:40:28+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_384cd081-7034-46cf-bb48-841728494c60 to /host/opt/cni/bin/\\\\n2025-12-06T03:40:29Z [verbose] multus-daemon started\\\\n2025-12-06T03:40:29Z [verbose] Readiness Indicator file check\\\\n2025-12-06T03:41:14Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trlt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22rrq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:15Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.097577 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ngspk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34808ab2-d21b-40b6-8fe3-e18bf9e5fc46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52509d21f5b4bfbd458162cba176ff82c627c7c4671204addf9d76f7392fd760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qkf8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9685a4d5d4f40ff4f59b6da6193785f340890453d1e422d3421a621dc7bd02f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qkf8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ngspk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:15Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.106574 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5ltx2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8881599e-71e5-4578-b3f9-21f4fa2f57aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbm6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbm6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5ltx2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:15Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.120490 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1721e985-40bd-4b2d-a0e8-d4e365094ba6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a895e0cc04880c3287ba725ac5484835235f1e066a94ce98fcf4b6133c5cbd2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a21a97385f856ec996105b597d0ae81c56ed28b625ef3f2905e1af6ec8f6a45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://646e01a58908ba2e703aae5eef53e1b981b0f0c3d1d3202487f2a8b94ee3f55d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31752b84d86bda30792b29e4c6ed963753cc576c5de784395e8e6dcbebaced47\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"ver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:40:26.813730 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814130 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:40:26.814139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:40:26.814143 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:40:26.814145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 03:40:26.821403 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821442 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821482 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821496 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 03:40:26.821509 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821514 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 03:40:26.821612 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 03:40:26.821619 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1206 03:40:26.822049 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16a92d501c9e626fe91e4a6793e426d10eb292cfbab766db8b8cc72e5dae94ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:15Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.132367 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd230730-a9a1-47ed-b625-25e32231f271\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce321001135d0a2b878316f8003bf1bc54cba9a36f6a654ba376e1e7c0b4da71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af83814e6b2c5ed803157308ab895551e0826fd02f1ecc8b2751145c09d359da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9e6b48b5701dd330ef91712cdcd160afe9831a3881038bfb1e0397a9c8a3aa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2d359853c738e3e0e264f1a6a919b17afc71a362821b3ea170cda6695e3593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2d359853c738e3e0e264f1a6a919b17afc71a362821b3ea170cda6695e3593f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:15Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.178611 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-b2knv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a3ae46f-03b5-4999-9d58-c38ee222f68a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76665d4c127445009fd66078827b54607c9ef812c8011d1c0b4cde1c0516853a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk4hj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-b2knv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:15Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.179689 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.179730 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.179739 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.179774 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.179786 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:15Z","lastTransitionTime":"2025-12-06T03:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.281881 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.281912 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.281922 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.281936 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.281944 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:15Z","lastTransitionTime":"2025-12-06T03:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.384246 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.384288 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.384298 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.384312 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.384321 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:15Z","lastTransitionTime":"2025-12-06T03:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.486442 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.486495 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.486511 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.486532 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.486547 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:15Z","lastTransitionTime":"2025-12-06T03:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.588797 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.588844 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.588854 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.588869 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.588880 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:15Z","lastTransitionTime":"2025-12-06T03:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.691125 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.691366 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.691440 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.691509 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.691584 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:15Z","lastTransitionTime":"2025-12-06T03:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.794570 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.794779 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.794876 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.794959 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.795030 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:15Z","lastTransitionTime":"2025-12-06T03:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.897278 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.897312 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.897320 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.897333 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.897341 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:15Z","lastTransitionTime":"2025-12-06T03:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.927025 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-22rrq_03112a5d-5882-4fac-9ae5-13525ff82fe0/kube-multus/0.log" Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.927108 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-22rrq" event={"ID":"03112a5d-5882-4fac-9ae5-13525ff82fe0","Type":"ContainerStarted","Data":"831dab65557627cbce0d055459052dae5b594f33b66858367ccd034ff6811924"} Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.947713 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:15Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.958840 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6df38316-e0d3-4018-8d27-3620eba3a68d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57381e9589a2dc04a4d2733c37985f42594182128a374b7be0f471ea69138b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bf5e41701c7a2d45adc4740ceb4ac1785db17bc7cb15daed439f7556812d1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpxxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:15Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.968555 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kj7hf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd46dfa5-7d4a-49ed-8199-523f10cebdf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5eee8a0cf96ceb3785dac9df31bf8d31715d1c88fbea17c5e582b0a4554bb63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmkcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kj7hf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:15Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.982895 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2466610962332c6ca120fe055e4e8a3fc623088119dc1a412e06b745176f89e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:15Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.996290 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2093b6-c4b5-4db8-be16-fbc859885e8d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0d6e582161f0d9126126e03b27211be22f6b5f797154928c92fa13a8118f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03004a29aefe7fe2452fb9f6663c0807afccb055d7ddbf6b907f305e6a19c81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1801ac64db4ba91c2e8a249e3651d3e50ab62e9f2c87312ce9bf7f38518a82cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6e454507cadbb982a566a3692995aaa3c733ff1cd7fb4870476602190e1eeba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:15Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.999841 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:15 crc kubenswrapper[4802]: I1206 03:41:15.999910 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:15.999927 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:15.999954 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:15.999972 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:15Z","lastTransitionTime":"2025-12-06T03:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.013838 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:16Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.028004 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fad9b63d84f41fccf696cd313d7ca7f72b3d3251d6c1b7341573a3fadf1e378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:16Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.047999 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb3f278fff47c900c56d8a563d5ed75a2ae84e2034b0f45aad4083790b9e7946\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b10ca16ac26904ec5e6245344b71d836888cae9b069cae6a7a4831d609752db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:16Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.069005 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtw2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3a3bac2-3923-4507-a4de-1556b3ad473e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fed5ee4189d54456a3216f4fdb1ecf003fa47e29dbfe4b2b2f301edfb897d6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtw2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:16Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.091804 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7730de4f033561ad77046e028cd1a3f0bfe2775387cbdfa35f47ea79909ec4cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a158aebd701ef46225d3e37109ae28ab58d107087bf06650c4b6d0055a3bbba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2357f37497fde64937eca7b38693893af745ab278653ebc79406e122a6d4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8941bd25107333b264d3f081a811fdfeea23ad92193fc61367665fd2cd1209c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f9cd4dc75e63f09f60b48b6da284643910bb32ba51f2a808cb4d9803f30042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6796f65ee9c5bd0c18f205697135d17677c5daf3ba378cc992cba795e0188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3870b61bce487c69724f668c38fc02190ff97b5e3b1434d66ed624412d18319b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3870b61bce487c69724f668c38fc02190ff97b5e3b1434d66ed624412d18319b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:40:54Z\\\",\\\"message\\\":\\\"/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 03:40:54.580644 6463 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 03:40:54.580730 6463 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 03:40:54.580797 6463 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 03:40:54.580836 6463 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 03:40:54.580845 6463 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 03:40:54.580895 6463 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1206 03:40:54.580943 6463 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1206 03:40:54.580950 6463 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1206 03:40:54.581001 6463 factory.go:656] Stopping watch factory\\\\nI1206 03:40:54.581036 6463 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 03:40:54.581050 6463 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 03:40:54.581063 6463 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 03:40:54.581074 6463 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 03:40:54.581085 6463 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1206 03:40:54.581102 6463 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:53Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pvm6q_openshift-ovn-kubernetes(eafce4e0-e7fb-4877-b0ab-3283829ba1ef)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3cb210bfc97a439a7ff98c02c59f8ad16bc400dc214a616a44b43a8b01e2782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pvm6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:16Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.102690 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.102786 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.102802 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.102819 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.102831 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:16Z","lastTransitionTime":"2025-12-06T03:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.106535 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:16Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.127593 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1721e985-40bd-4b2d-a0e8-d4e365094ba6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a895e0cc04880c3287ba725ac5484835235f1e066a94ce98fcf4b6133c5cbd2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a21a97385f856ec996105b597d0ae81c56ed28b625ef3f2905e1af6ec8f6a45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://646e01a58908ba2e703aae5eef53e1b981b0f0c3d1d3202487f2a8b94ee3f55d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31752b84d86bda30792b29e4c6ed963753cc576c5de784395e8e6dcbebaced47\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"ver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:40:26.813730 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814130 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:40:26.814139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:40:26.814143 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:40:26.814145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 03:40:26.821403 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821442 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821482 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821496 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 03:40:26.821509 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821514 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 03:40:26.821612 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 03:40:26.821619 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1206 03:40:26.822049 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16a92d501c9e626fe91e4a6793e426d10eb292cfbab766db8b8cc72e5dae94ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:16Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.143234 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd230730-a9a1-47ed-b625-25e32231f271\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce321001135d0a2b878316f8003bf1bc54cba9a36f6a654ba376e1e7c0b4da71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af83814e6b2c5ed803157308ab895551e0826fd02f1ecc8b2751145c09d359da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9e6b48b5701dd330ef91712cdcd160afe9831a3881038bfb1e0397a9c8a3aa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2d359853c738e3e0e264f1a6a919b17afc71a362821b3ea170cda6695e3593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2d359853c738e3e0e264f1a6a919b17afc71a362821b3ea170cda6695e3593f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:16Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.156086 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-b2knv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a3ae46f-03b5-4999-9d58-c38ee222f68a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76665d4c127445009fd66078827b54607c9ef812c8011d1c0b4cde1c0516853a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk4hj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-b2knv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:16Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.167770 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22rrq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03112a5d-5882-4fac-9ae5-13525ff82fe0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://831dab65557627cbce0d055459052dae5b594f33b66858367ccd034ff6811924\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0205b089b12c42688203bb7ec0f27dc5f9a577dabdb446b811f184046f85dc3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:41:14Z\\\",\\\"message\\\":\\\"2025-12-06T03:40:28+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_384cd081-7034-46cf-bb48-841728494c60\\\\n2025-12-06T03:40:28+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_384cd081-7034-46cf-bb48-841728494c60 to /host/opt/cni/bin/\\\\n2025-12-06T03:40:29Z [verbose] multus-daemon started\\\\n2025-12-06T03:40:29Z [verbose] Readiness Indicator file check\\\\n2025-12-06T03:41:14Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:41:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trlt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22rrq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:16Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.176741 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ngspk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34808ab2-d21b-40b6-8fe3-e18bf9e5fc46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52509d21f5b4bfbd458162cba176ff82c627c7c4671204addf9d76f7392fd760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qkf8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9685a4d5d4f40ff4f59b6da6193785f340890453d1e422d3421a621dc7bd02f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qkf8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ngspk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:16Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.187138 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5ltx2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8881599e-71e5-4578-b3f9-21f4fa2f57aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbm6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbm6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5ltx2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:16Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.205317 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.205381 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.205398 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.205425 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.205442 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:16Z","lastTransitionTime":"2025-12-06T03:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.308317 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.308394 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.308413 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.308441 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.308458 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:16Z","lastTransitionTime":"2025-12-06T03:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.411184 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.411229 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.411245 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.411268 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.411285 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:16Z","lastTransitionTime":"2025-12-06T03:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.449153 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:41:16 crc kubenswrapper[4802]: E1206 03:41:16.449306 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.449545 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:41:16 crc kubenswrapper[4802]: E1206 03:41:16.449650 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.449932 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.450072 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:41:16 crc kubenswrapper[4802]: E1206 03:41:16.450162 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:41:16 crc kubenswrapper[4802]: E1206 03:41:16.450278 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5ltx2" podUID="8881599e-71e5-4578-b3f9-21f4fa2f57aa" Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.514927 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.515003 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.515028 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.515056 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.515077 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:16Z","lastTransitionTime":"2025-12-06T03:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.617958 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.617991 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.617998 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.618013 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.618024 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:16Z","lastTransitionTime":"2025-12-06T03:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.720507 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.720546 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.720554 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.720567 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.720577 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:16Z","lastTransitionTime":"2025-12-06T03:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.823533 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.823833 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.823861 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.823885 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.823904 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:16Z","lastTransitionTime":"2025-12-06T03:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.926214 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.926255 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.926268 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.926286 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:16 crc kubenswrapper[4802]: I1206 03:41:16.926298 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:16Z","lastTransitionTime":"2025-12-06T03:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.029158 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.029203 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.029215 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.029231 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.029243 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:17Z","lastTransitionTime":"2025-12-06T03:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.131572 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.131619 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.131630 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.131646 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.131657 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:17Z","lastTransitionTime":"2025-12-06T03:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.234389 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.234441 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.234459 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.234482 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.234499 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:17Z","lastTransitionTime":"2025-12-06T03:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.337838 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.337903 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.337913 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.337938 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.337955 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:17Z","lastTransitionTime":"2025-12-06T03:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.441994 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.442391 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.442414 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.442436 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.442451 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:17Z","lastTransitionTime":"2025-12-06T03:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.464035 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kj7hf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd46dfa5-7d4a-49ed-8199-523f10cebdf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5eee8a0cf96ceb3785dac9df31bf8d31715d1c88fbea17c5e582b0a4554bb63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmkcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kj7hf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:17Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.481543 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2466610962332c6ca120fe055e4e8a3fc623088119dc1a412e06b745176f89e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:17Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.501549 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:17Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.517201 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6df38316-e0d3-4018-8d27-3620eba3a68d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57381e9589a2dc04a4d2733c37985f42594182128a374b7be0f471ea69138b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bf5e41701c7a2d45adc4740ceb4ac1785db17bc7cb15daed439f7556812d1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpxxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:17Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.531831 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fad9b63d84f41fccf696cd313d7ca7f72b3d3251d6c1b7341573a3fadf1e378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:17Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.544293 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.544329 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.544339 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.544357 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.544367 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:17Z","lastTransitionTime":"2025-12-06T03:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.548577 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb3f278fff47c900c56d8a563d5ed75a2ae84e2034b0f45aad4083790b9e7946\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b10ca16ac26904ec5e6245344b71d836888cae9b069cae6a7a4831d609752db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:17Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.562364 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2093b6-c4b5-4db8-be16-fbc859885e8d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0d6e582161f0d9126126e03b27211be22f6b5f797154928c92fa13a8118f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03004a29aefe7fe2452fb9f6663c0807afccb055d7ddbf6b907f305e6a19c81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1801ac64db4ba91c2e8a249e3651d3e50ab62e9f2c87312ce9bf7f38518a82cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6e454507cadbb982a566a3692995aaa3c733ff1cd7fb4870476602190e1eeba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:17Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.580073 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:17Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.595154 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:17Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.616486 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtw2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3a3bac2-3923-4507-a4de-1556b3ad473e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fed5ee4189d54456a3216f4fdb1ecf003fa47e29dbfe4b2b2f301edfb897d6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtw2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:17Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.635323 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7730de4f033561ad77046e028cd1a3f0bfe2775387cbdfa35f47ea79909ec4cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a158aebd701ef46225d3e37109ae28ab58d107087bf06650c4b6d0055a3bbba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2357f37497fde64937eca7b38693893af745ab278653ebc79406e122a6d4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8941bd25107333b264d3f081a811fdfeea23ad92193fc61367665fd2cd1209c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f9cd4dc75e63f09f60b48b6da284643910bb32ba51f2a808cb4d9803f30042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6796f65ee9c5bd0c18f205697135d17677c5daf3ba378cc992cba795e0188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3870b61bce487c69724f668c38fc02190ff97b5e3b1434d66ed624412d18319b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3870b61bce487c69724f668c38fc02190ff97b5e3b1434d66ed624412d18319b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:40:54Z\\\",\\\"message\\\":\\\"/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 03:40:54.580644 6463 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 03:40:54.580730 6463 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 03:40:54.580797 6463 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 03:40:54.580836 6463 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 03:40:54.580845 6463 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 03:40:54.580895 6463 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1206 03:40:54.580943 6463 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1206 03:40:54.580950 6463 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1206 03:40:54.581001 6463 factory.go:656] Stopping watch factory\\\\nI1206 03:40:54.581036 6463 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 03:40:54.581050 6463 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 03:40:54.581063 6463 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 03:40:54.581074 6463 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 03:40:54.581085 6463 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1206 03:40:54.581102 6463 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:53Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pvm6q_openshift-ovn-kubernetes(eafce4e0-e7fb-4877-b0ab-3283829ba1ef)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3cb210bfc97a439a7ff98c02c59f8ad16bc400dc214a616a44b43a8b01e2782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pvm6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:17Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.647698 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.647839 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.647862 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.647894 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.647871 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-b2knv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a3ae46f-03b5-4999-9d58-c38ee222f68a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76665d4c127445009fd66078827b54607c9ef812c8011d1c0b4cde1c0516853a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk4hj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-b2knv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:17Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.647914 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:17Z","lastTransitionTime":"2025-12-06T03:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.663545 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22rrq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03112a5d-5882-4fac-9ae5-13525ff82fe0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://831dab65557627cbce0d055459052dae5b594f33b66858367ccd034ff6811924\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0205b089b12c42688203bb7ec0f27dc5f9a577dabdb446b811f184046f85dc3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:41:14Z\\\",\\\"message\\\":\\\"2025-12-06T03:40:28+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_384cd081-7034-46cf-bb48-841728494c60\\\\n2025-12-06T03:40:28+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_384cd081-7034-46cf-bb48-841728494c60 to /host/opt/cni/bin/\\\\n2025-12-06T03:40:29Z [verbose] multus-daemon started\\\\n2025-12-06T03:40:29Z [verbose] Readiness Indicator file check\\\\n2025-12-06T03:41:14Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:41:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trlt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22rrq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:17Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.677593 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ngspk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34808ab2-d21b-40b6-8fe3-e18bf9e5fc46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52509d21f5b4bfbd458162cba176ff82c627c7c4671204addf9d76f7392fd760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qkf8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9685a4d5d4f40ff4f59b6da6193785f340890453d1e422d3421a621dc7bd02f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qkf8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ngspk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:17Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.687344 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5ltx2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8881599e-71e5-4578-b3f9-21f4fa2f57aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbm6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbm6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5ltx2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:17Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.698937 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1721e985-40bd-4b2d-a0e8-d4e365094ba6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a895e0cc04880c3287ba725ac5484835235f1e066a94ce98fcf4b6133c5cbd2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a21a97385f856ec996105b597d0ae81c56ed28b625ef3f2905e1af6ec8f6a45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://646e01a58908ba2e703aae5eef53e1b981b0f0c3d1d3202487f2a8b94ee3f55d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31752b84d86bda30792b29e4c6ed963753cc576c5de784395e8e6dcbebaced47\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"ver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:40:26.813730 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814130 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:40:26.814139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:40:26.814143 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:40:26.814145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 03:40:26.821403 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821442 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821482 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821496 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 03:40:26.821509 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821514 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 03:40:26.821612 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 03:40:26.821619 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1206 03:40:26.822049 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16a92d501c9e626fe91e4a6793e426d10eb292cfbab766db8b8cc72e5dae94ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:17Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.709705 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd230730-a9a1-47ed-b625-25e32231f271\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce321001135d0a2b878316f8003bf1bc54cba9a36f6a654ba376e1e7c0b4da71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af83814e6b2c5ed803157308ab895551e0826fd02f1ecc8b2751145c09d359da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9e6b48b5701dd330ef91712cdcd160afe9831a3881038bfb1e0397a9c8a3aa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2d359853c738e3e0e264f1a6a919b17afc71a362821b3ea170cda6695e3593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2d359853c738e3e0e264f1a6a919b17afc71a362821b3ea170cda6695e3593f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:17Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.750612 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.750660 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.750676 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.750700 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.750717 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:17Z","lastTransitionTime":"2025-12-06T03:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.853323 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.853401 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.853425 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.853452 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.853472 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:17Z","lastTransitionTime":"2025-12-06T03:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.955866 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.955911 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.955920 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.955935 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:17 crc kubenswrapper[4802]: I1206 03:41:17.955948 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:17Z","lastTransitionTime":"2025-12-06T03:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:18 crc kubenswrapper[4802]: I1206 03:41:18.060166 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:18 crc kubenswrapper[4802]: I1206 03:41:18.060255 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:18 crc kubenswrapper[4802]: I1206 03:41:18.060274 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:18 crc kubenswrapper[4802]: I1206 03:41:18.060298 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:18 crc kubenswrapper[4802]: I1206 03:41:18.060327 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:18Z","lastTransitionTime":"2025-12-06T03:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:18 crc kubenswrapper[4802]: I1206 03:41:18.163699 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:18 crc kubenswrapper[4802]: I1206 03:41:18.163842 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:18 crc kubenswrapper[4802]: I1206 03:41:18.163868 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:18 crc kubenswrapper[4802]: I1206 03:41:18.163899 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:18 crc kubenswrapper[4802]: I1206 03:41:18.163922 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:18Z","lastTransitionTime":"2025-12-06T03:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:18 crc kubenswrapper[4802]: I1206 03:41:18.269804 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:18 crc kubenswrapper[4802]: I1206 03:41:18.269851 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:18 crc kubenswrapper[4802]: I1206 03:41:18.269864 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:18 crc kubenswrapper[4802]: I1206 03:41:18.269882 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:18 crc kubenswrapper[4802]: I1206 03:41:18.269893 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:18Z","lastTransitionTime":"2025-12-06T03:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:18 crc kubenswrapper[4802]: I1206 03:41:18.372300 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:18 crc kubenswrapper[4802]: I1206 03:41:18.372350 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:18 crc kubenswrapper[4802]: I1206 03:41:18.372368 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:18 crc kubenswrapper[4802]: I1206 03:41:18.372391 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:18 crc kubenswrapper[4802]: I1206 03:41:18.372409 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:18Z","lastTransitionTime":"2025-12-06T03:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:18 crc kubenswrapper[4802]: I1206 03:41:18.449268 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:41:18 crc kubenswrapper[4802]: I1206 03:41:18.449353 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:41:18 crc kubenswrapper[4802]: I1206 03:41:18.449353 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:41:18 crc kubenswrapper[4802]: I1206 03:41:18.449362 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:41:18 crc kubenswrapper[4802]: E1206 03:41:18.449474 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5ltx2" podUID="8881599e-71e5-4578-b3f9-21f4fa2f57aa" Dec 06 03:41:18 crc kubenswrapper[4802]: E1206 03:41:18.449576 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:41:18 crc kubenswrapper[4802]: E1206 03:41:18.449617 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:41:18 crc kubenswrapper[4802]: E1206 03:41:18.449714 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:41:18 crc kubenswrapper[4802]: I1206 03:41:18.475379 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:18 crc kubenswrapper[4802]: I1206 03:41:18.475445 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:18 crc kubenswrapper[4802]: I1206 03:41:18.475467 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:18 crc kubenswrapper[4802]: I1206 03:41:18.475497 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:18 crc kubenswrapper[4802]: I1206 03:41:18.475518 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:18Z","lastTransitionTime":"2025-12-06T03:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:18 crc kubenswrapper[4802]: I1206 03:41:18.578318 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:18 crc kubenswrapper[4802]: I1206 03:41:18.578377 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:18 crc kubenswrapper[4802]: I1206 03:41:18.578390 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:18 crc kubenswrapper[4802]: I1206 03:41:18.578413 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:18 crc kubenswrapper[4802]: I1206 03:41:18.578424 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:18Z","lastTransitionTime":"2025-12-06T03:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:18 crc kubenswrapper[4802]: I1206 03:41:18.680900 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:18 crc kubenswrapper[4802]: I1206 03:41:18.680944 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:18 crc kubenswrapper[4802]: I1206 03:41:18.680952 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:18 crc kubenswrapper[4802]: I1206 03:41:18.680968 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:18 crc kubenswrapper[4802]: I1206 03:41:18.680977 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:18Z","lastTransitionTime":"2025-12-06T03:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:18 crc kubenswrapper[4802]: I1206 03:41:18.783740 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:18 crc kubenswrapper[4802]: I1206 03:41:18.783792 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:18 crc kubenswrapper[4802]: I1206 03:41:18.783803 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:18 crc kubenswrapper[4802]: I1206 03:41:18.783820 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:18 crc kubenswrapper[4802]: I1206 03:41:18.783833 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:18Z","lastTransitionTime":"2025-12-06T03:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:18 crc kubenswrapper[4802]: I1206 03:41:18.885839 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:18 crc kubenswrapper[4802]: I1206 03:41:18.885873 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:18 crc kubenswrapper[4802]: I1206 03:41:18.885884 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:18 crc kubenswrapper[4802]: I1206 03:41:18.885900 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:18 crc kubenswrapper[4802]: I1206 03:41:18.885911 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:18Z","lastTransitionTime":"2025-12-06T03:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:18 crc kubenswrapper[4802]: I1206 03:41:18.987611 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:18 crc kubenswrapper[4802]: I1206 03:41:18.987654 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:18 crc kubenswrapper[4802]: I1206 03:41:18.987666 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:18 crc kubenswrapper[4802]: I1206 03:41:18.987682 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:18 crc kubenswrapper[4802]: I1206 03:41:18.987693 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:18Z","lastTransitionTime":"2025-12-06T03:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:19 crc kubenswrapper[4802]: I1206 03:41:19.089887 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:19 crc kubenswrapper[4802]: I1206 03:41:19.089967 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:19 crc kubenswrapper[4802]: I1206 03:41:19.089981 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:19 crc kubenswrapper[4802]: I1206 03:41:19.090005 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:19 crc kubenswrapper[4802]: I1206 03:41:19.090021 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:19Z","lastTransitionTime":"2025-12-06T03:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:19 crc kubenswrapper[4802]: I1206 03:41:19.192223 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:19 crc kubenswrapper[4802]: I1206 03:41:19.192254 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:19 crc kubenswrapper[4802]: I1206 03:41:19.192264 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:19 crc kubenswrapper[4802]: I1206 03:41:19.192279 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:19 crc kubenswrapper[4802]: I1206 03:41:19.192288 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:19Z","lastTransitionTime":"2025-12-06T03:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:19 crc kubenswrapper[4802]: I1206 03:41:19.294633 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:19 crc kubenswrapper[4802]: I1206 03:41:19.294692 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:19 crc kubenswrapper[4802]: I1206 03:41:19.294711 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:19 crc kubenswrapper[4802]: I1206 03:41:19.294736 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:19 crc kubenswrapper[4802]: I1206 03:41:19.294792 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:19Z","lastTransitionTime":"2025-12-06T03:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:19 crc kubenswrapper[4802]: I1206 03:41:19.397705 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:19 crc kubenswrapper[4802]: I1206 03:41:19.397766 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:19 crc kubenswrapper[4802]: I1206 03:41:19.397776 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:19 crc kubenswrapper[4802]: I1206 03:41:19.397792 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:19 crc kubenswrapper[4802]: I1206 03:41:19.397801 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:19Z","lastTransitionTime":"2025-12-06T03:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:19 crc kubenswrapper[4802]: I1206 03:41:19.500585 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:19 crc kubenswrapper[4802]: I1206 03:41:19.500656 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:19 crc kubenswrapper[4802]: I1206 03:41:19.500679 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:19 crc kubenswrapper[4802]: I1206 03:41:19.500712 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:19 crc kubenswrapper[4802]: I1206 03:41:19.500735 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:19Z","lastTransitionTime":"2025-12-06T03:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:19 crc kubenswrapper[4802]: I1206 03:41:19.603744 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:19 crc kubenswrapper[4802]: I1206 03:41:19.603814 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:19 crc kubenswrapper[4802]: I1206 03:41:19.603827 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:19 crc kubenswrapper[4802]: I1206 03:41:19.603846 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:19 crc kubenswrapper[4802]: I1206 03:41:19.603863 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:19Z","lastTransitionTime":"2025-12-06T03:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:19 crc kubenswrapper[4802]: I1206 03:41:19.706002 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:19 crc kubenswrapper[4802]: I1206 03:41:19.706092 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:19 crc kubenswrapper[4802]: I1206 03:41:19.706117 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:19 crc kubenswrapper[4802]: I1206 03:41:19.706155 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:19 crc kubenswrapper[4802]: I1206 03:41:19.706182 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:19Z","lastTransitionTime":"2025-12-06T03:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:19 crc kubenswrapper[4802]: I1206 03:41:19.809427 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:19 crc kubenswrapper[4802]: I1206 03:41:19.809483 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:19 crc kubenswrapper[4802]: I1206 03:41:19.809499 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:19 crc kubenswrapper[4802]: I1206 03:41:19.809520 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:19 crc kubenswrapper[4802]: I1206 03:41:19.809537 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:19Z","lastTransitionTime":"2025-12-06T03:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:19 crc kubenswrapper[4802]: I1206 03:41:19.911301 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:19 crc kubenswrapper[4802]: I1206 03:41:19.911342 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:19 crc kubenswrapper[4802]: I1206 03:41:19.911357 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:19 crc kubenswrapper[4802]: I1206 03:41:19.911374 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:19 crc kubenswrapper[4802]: I1206 03:41:19.911386 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:19Z","lastTransitionTime":"2025-12-06T03:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.014611 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.014661 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.014677 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.014703 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.014720 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:20Z","lastTransitionTime":"2025-12-06T03:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.096164 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.096229 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.096248 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.096274 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.096296 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:20Z","lastTransitionTime":"2025-12-06T03:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:20 crc kubenswrapper[4802]: E1206 03:41:20.117617 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0e01ca3a-758f-44bc-b4a2-30c1bc04fda0\\\",\\\"systemUUID\\\":\\\"c7c815e4-d2a4-4244-b035-f988fc95e215\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:20Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.122471 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.122516 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.122576 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.122603 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.122621 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:20Z","lastTransitionTime":"2025-12-06T03:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:20 crc kubenswrapper[4802]: E1206 03:41:20.138227 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0e01ca3a-758f-44bc-b4a2-30c1bc04fda0\\\",\\\"systemUUID\\\":\\\"c7c815e4-d2a4-4244-b035-f988fc95e215\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:20Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.142503 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.142594 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.142620 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.142647 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.142665 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:20Z","lastTransitionTime":"2025-12-06T03:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:20 crc kubenswrapper[4802]: E1206 03:41:20.157784 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0e01ca3a-758f-44bc-b4a2-30c1bc04fda0\\\",\\\"systemUUID\\\":\\\"c7c815e4-d2a4-4244-b035-f988fc95e215\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:20Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.162302 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.162363 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.162387 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.162434 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.162460 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:20Z","lastTransitionTime":"2025-12-06T03:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:20 crc kubenswrapper[4802]: E1206 03:41:20.186838 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0e01ca3a-758f-44bc-b4a2-30c1bc04fda0\\\",\\\"systemUUID\\\":\\\"c7c815e4-d2a4-4244-b035-f988fc95e215\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:20Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.192171 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.192237 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.192258 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.192286 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.192309 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:20Z","lastTransitionTime":"2025-12-06T03:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:20 crc kubenswrapper[4802]: E1206 03:41:20.212420 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0e01ca3a-758f-44bc-b4a2-30c1bc04fda0\\\",\\\"systemUUID\\\":\\\"c7c815e4-d2a4-4244-b035-f988fc95e215\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:20Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:20 crc kubenswrapper[4802]: E1206 03:41:20.212578 4802 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.214252 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.214280 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.214290 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.214305 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.214316 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:20Z","lastTransitionTime":"2025-12-06T03:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.316960 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.317006 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.317014 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.317027 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.317035 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:20Z","lastTransitionTime":"2025-12-06T03:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.420018 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.420062 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.420074 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.420093 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.420106 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:20Z","lastTransitionTime":"2025-12-06T03:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.449539 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.449656 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.449711 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:41:20 crc kubenswrapper[4802]: E1206 03:41:20.450051 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.450288 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:41:20 crc kubenswrapper[4802]: E1206 03:41:20.450365 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:41:20 crc kubenswrapper[4802]: E1206 03:41:20.450455 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:41:20 crc kubenswrapper[4802]: E1206 03:41:20.450626 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5ltx2" podUID="8881599e-71e5-4578-b3f9-21f4fa2f57aa" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.464887 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.522679 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.522710 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.522721 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.522736 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.522762 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:20Z","lastTransitionTime":"2025-12-06T03:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.626206 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.626265 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.626286 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.626312 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.626332 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:20Z","lastTransitionTime":"2025-12-06T03:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.728884 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.728957 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.728982 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.729010 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.729032 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:20Z","lastTransitionTime":"2025-12-06T03:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.832374 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.832446 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.832463 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.832487 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.832505 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:20Z","lastTransitionTime":"2025-12-06T03:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.935143 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.935219 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.935243 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.935273 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:20 crc kubenswrapper[4802]: I1206 03:41:20.935296 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:20Z","lastTransitionTime":"2025-12-06T03:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:21 crc kubenswrapper[4802]: I1206 03:41:21.038832 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:21 crc kubenswrapper[4802]: I1206 03:41:21.038966 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:21 crc kubenswrapper[4802]: I1206 03:41:21.038998 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:21 crc kubenswrapper[4802]: I1206 03:41:21.039090 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:21 crc kubenswrapper[4802]: I1206 03:41:21.039168 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:21Z","lastTransitionTime":"2025-12-06T03:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:21 crc kubenswrapper[4802]: I1206 03:41:21.142537 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:21 crc kubenswrapper[4802]: I1206 03:41:21.142647 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:21 crc kubenswrapper[4802]: I1206 03:41:21.142674 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:21 crc kubenswrapper[4802]: I1206 03:41:21.142707 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:21 crc kubenswrapper[4802]: I1206 03:41:21.142733 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:21Z","lastTransitionTime":"2025-12-06T03:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:21 crc kubenswrapper[4802]: I1206 03:41:21.247232 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:21 crc kubenswrapper[4802]: I1206 03:41:21.247283 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:21 crc kubenswrapper[4802]: I1206 03:41:21.247300 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:21 crc kubenswrapper[4802]: I1206 03:41:21.247323 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:21 crc kubenswrapper[4802]: I1206 03:41:21.247340 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:21Z","lastTransitionTime":"2025-12-06T03:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:21 crc kubenswrapper[4802]: I1206 03:41:21.351036 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:21 crc kubenswrapper[4802]: I1206 03:41:21.351104 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:21 crc kubenswrapper[4802]: I1206 03:41:21.351122 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:21 crc kubenswrapper[4802]: I1206 03:41:21.351149 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:21 crc kubenswrapper[4802]: I1206 03:41:21.351168 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:21Z","lastTransitionTime":"2025-12-06T03:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:21 crc kubenswrapper[4802]: I1206 03:41:21.451141 4802 scope.go:117] "RemoveContainer" containerID="3870b61bce487c69724f668c38fc02190ff97b5e3b1434d66ed624412d18319b" Dec 06 03:41:21 crc kubenswrapper[4802]: I1206 03:41:21.454079 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:21 crc kubenswrapper[4802]: I1206 03:41:21.454133 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:21 crc kubenswrapper[4802]: I1206 03:41:21.454150 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:21 crc kubenswrapper[4802]: I1206 03:41:21.454178 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:21 crc kubenswrapper[4802]: I1206 03:41:21.454200 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:21Z","lastTransitionTime":"2025-12-06T03:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:21 crc kubenswrapper[4802]: I1206 03:41:21.557136 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:21 crc kubenswrapper[4802]: I1206 03:41:21.557224 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:21 crc kubenswrapper[4802]: I1206 03:41:21.557250 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:21 crc kubenswrapper[4802]: I1206 03:41:21.557283 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:21 crc kubenswrapper[4802]: I1206 03:41:21.557308 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:21Z","lastTransitionTime":"2025-12-06T03:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:21 crc kubenswrapper[4802]: I1206 03:41:21.660894 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:21 crc kubenswrapper[4802]: I1206 03:41:21.660959 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:21 crc kubenswrapper[4802]: I1206 03:41:21.660976 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:21 crc kubenswrapper[4802]: I1206 03:41:21.661005 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:21 crc kubenswrapper[4802]: I1206 03:41:21.661028 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:21Z","lastTransitionTime":"2025-12-06T03:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:21 crc kubenswrapper[4802]: I1206 03:41:21.764133 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:21 crc kubenswrapper[4802]: I1206 03:41:21.764209 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:21 crc kubenswrapper[4802]: I1206 03:41:21.764232 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:21 crc kubenswrapper[4802]: I1206 03:41:21.764262 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:21 crc kubenswrapper[4802]: I1206 03:41:21.764296 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:21Z","lastTransitionTime":"2025-12-06T03:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:21 crc kubenswrapper[4802]: I1206 03:41:21.867328 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:21 crc kubenswrapper[4802]: I1206 03:41:21.867389 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:21 crc kubenswrapper[4802]: I1206 03:41:21.867407 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:21 crc kubenswrapper[4802]: I1206 03:41:21.867432 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:21 crc kubenswrapper[4802]: I1206 03:41:21.867452 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:21Z","lastTransitionTime":"2025-12-06T03:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:21 crc kubenswrapper[4802]: I1206 03:41:21.970647 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:21 crc kubenswrapper[4802]: I1206 03:41:21.970740 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:21 crc kubenswrapper[4802]: I1206 03:41:21.970802 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:21 crc kubenswrapper[4802]: I1206 03:41:21.970839 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:21 crc kubenswrapper[4802]: I1206 03:41:21.970864 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:21Z","lastTransitionTime":"2025-12-06T03:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.074141 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.074222 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.074243 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.074274 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.074302 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:22Z","lastTransitionTime":"2025-12-06T03:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.177248 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.177330 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.177357 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.177390 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.177413 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:22Z","lastTransitionTime":"2025-12-06T03:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.280666 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.280813 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.280839 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.280873 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.280897 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:22Z","lastTransitionTime":"2025-12-06T03:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.384361 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.384417 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.384433 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.384457 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.384474 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:22Z","lastTransitionTime":"2025-12-06T03:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.449178 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.449218 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.449218 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.449283 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:41:22 crc kubenswrapper[4802]: E1206 03:41:22.449322 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:41:22 crc kubenswrapper[4802]: E1206 03:41:22.449420 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:41:22 crc kubenswrapper[4802]: E1206 03:41:22.449551 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:41:22 crc kubenswrapper[4802]: E1206 03:41:22.449732 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5ltx2" podUID="8881599e-71e5-4578-b3f9-21f4fa2f57aa" Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.487006 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.487046 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.487057 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.487074 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.487084 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:22Z","lastTransitionTime":"2025-12-06T03:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.589326 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.589403 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.589430 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.589467 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.589492 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:22Z","lastTransitionTime":"2025-12-06T03:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.692273 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.692329 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.692347 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.692371 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.692389 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:22Z","lastTransitionTime":"2025-12-06T03:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.794700 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.794779 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.794795 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.794816 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.794826 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:22Z","lastTransitionTime":"2025-12-06T03:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.897026 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.897091 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.897110 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.897141 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.897160 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:22Z","lastTransitionTime":"2025-12-06T03:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.952493 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pvm6q_eafce4e0-e7fb-4877-b0ab-3283829ba1ef/ovnkube-controller/2.log" Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.956961 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" event={"ID":"eafce4e0-e7fb-4877-b0ab-3283829ba1ef","Type":"ContainerStarted","Data":"bf554aaba0e773f6a6c1f29bb1819b7ae25b2b2db255c4bd0a16c512f314e9cf"} Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.957645 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.976615 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd230730-a9a1-47ed-b625-25e32231f271\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce321001135d0a2b878316f8003bf1bc54cba9a36f6a654ba376e1e7c0b4da71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af83814e6b2c5ed803157308ab895551e0826fd02f1ecc8b2751145c09d359da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9e6b48b5701dd330ef91712cdcd160afe9831a3881038bfb1e0397a9c8a3aa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2d359853c738e3e0e264f1a6a919b17afc71a362821b3ea170cda6695e3593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2d359853c738e3e0e264f1a6a919b17afc71a362821b3ea170cda6695e3593f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:22Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.991574 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-b2knv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a3ae46f-03b5-4999-9d58-c38ee222f68a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76665d4c127445009fd66078827b54607c9ef812c8011d1c0b4cde1c0516853a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk4hj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-b2knv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:22Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.999610 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.999673 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.999690 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.999714 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:22 crc kubenswrapper[4802]: I1206 03:41:22.999732 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:22Z","lastTransitionTime":"2025-12-06T03:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.013087 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22rrq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03112a5d-5882-4fac-9ae5-13525ff82fe0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://831dab65557627cbce0d055459052dae5b594f33b66858367ccd034ff6811924\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0205b089b12c42688203bb7ec0f27dc5f9a577dabdb446b811f184046f85dc3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:41:14Z\\\",\\\"message\\\":\\\"2025-12-06T03:40:28+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_384cd081-7034-46cf-bb48-841728494c60\\\\n2025-12-06T03:40:28+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_384cd081-7034-46cf-bb48-841728494c60 to /host/opt/cni/bin/\\\\n2025-12-06T03:40:29Z [verbose] multus-daemon started\\\\n2025-12-06T03:40:29Z [verbose] Readiness Indicator file check\\\\n2025-12-06T03:41:14Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:41:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trlt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22rrq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:23Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.033933 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ngspk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34808ab2-d21b-40b6-8fe3-e18bf9e5fc46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52509d21f5b4bfbd458162cba176ff82c627c7c4671204addf9d76f7392fd760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qkf8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9685a4d5d4f40ff4f59b6da6193785f340890453d1e422d3421a621dc7bd02f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qkf8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ngspk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:23Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.048303 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5ltx2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8881599e-71e5-4578-b3f9-21f4fa2f57aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbm6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbm6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5ltx2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:23Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.069401 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1721e985-40bd-4b2d-a0e8-d4e365094ba6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a895e0cc04880c3287ba725ac5484835235f1e066a94ce98fcf4b6133c5cbd2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a21a97385f856ec996105b597d0ae81c56ed28b625ef3f2905e1af6ec8f6a45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://646e01a58908ba2e703aae5eef53e1b981b0f0c3d1d3202487f2a8b94ee3f55d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31752b84d86bda30792b29e4c6ed963753cc576c5de784395e8e6dcbebaced47\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"ver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:40:26.813730 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814130 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:40:26.814139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:40:26.814143 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:40:26.814145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 03:40:26.821403 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821442 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821482 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821496 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 03:40:26.821509 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821514 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 03:40:26.821612 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 03:40:26.821619 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1206 03:40:26.822049 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16a92d501c9e626fe91e4a6793e426d10eb292cfbab766db8b8cc72e5dae94ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:23Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.087262 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6df38316-e0d3-4018-8d27-3620eba3a68d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57381e9589a2dc04a4d2733c37985f42594182128a374b7be0f471ea69138b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bf5e41701c7a2d45adc4740ceb4ac1785db17bc7cb15daed439f7556812d1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpxxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:23Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.100716 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kj7hf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd46dfa5-7d4a-49ed-8199-523f10cebdf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5eee8a0cf96ceb3785dac9df31bf8d31715d1c88fbea17c5e582b0a4554bb63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmkcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kj7hf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:23Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.103115 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.103242 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.103328 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.103463 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.103567 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:23Z","lastTransitionTime":"2025-12-06T03:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.118936 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2466610962332c6ca120fe055e4e8a3fc623088119dc1a412e06b745176f89e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:23Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.132222 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:23Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.147380 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:23Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.166608 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fad9b63d84f41fccf696cd313d7ca7f72b3d3251d6c1b7341573a3fadf1e378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:23Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.206038 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.206441 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.206501 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.206584 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.206661 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:23Z","lastTransitionTime":"2025-12-06T03:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.208201 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb3f278fff47c900c56d8a563d5ed75a2ae84e2034b0f45aad4083790b9e7946\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b10ca16ac26904ec5e6245344b71d836888cae9b069cae6a7a4831d609752db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:23Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.224940 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9de75acd-1ba9-4322-ae49-791724b4b6a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://990943a2a48d79458297f4e5cf4178bbd6e9eafbcadc154a803846c589474e85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18a08a15442f04e37a5aa71afbcc293e80b876477d60cbe71d562c75d141a9ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18a08a15442f04e37a5aa71afbcc293e80b876477d60cbe71d562c75d141a9ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:23Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.245625 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2093b6-c4b5-4db8-be16-fbc859885e8d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0d6e582161f0d9126126e03b27211be22f6b5f797154928c92fa13a8118f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03004a29aefe7fe2452fb9f6663c0807afccb055d7ddbf6b907f305e6a19c81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1801ac64db4ba91c2e8a249e3651d3e50ab62e9f2c87312ce9bf7f38518a82cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6e454507cadbb982a566a3692995aaa3c733ff1cd7fb4870476602190e1eeba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:23Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.277043 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7730de4f033561ad77046e028cd1a3f0bfe2775387cbdfa35f47ea79909ec4cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a158aebd701ef46225d3e37109ae28ab58d107087bf06650c4b6d0055a3bbba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2357f37497fde64937eca7b38693893af745ab278653ebc79406e122a6d4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8941bd25107333b264d3f081a811fdfeea23ad92193fc61367665fd2cd1209c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f9cd4dc75e63f09f60b48b6da284643910bb32ba51f2a808cb4d9803f30042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6796f65ee9c5bd0c18f205697135d17677c5daf3ba378cc992cba795e0188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf554aaba0e773f6a6c1f29bb1819b7ae25b2b2db255c4bd0a16c512f314e9cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3870b61bce487c69724f668c38fc02190ff97b5e3b1434d66ed624412d18319b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:40:54Z\\\",\\\"message\\\":\\\"/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 03:40:54.580644 6463 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 03:40:54.580730 6463 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 03:40:54.580797 6463 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 03:40:54.580836 6463 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 03:40:54.580845 6463 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 03:40:54.580895 6463 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1206 03:40:54.580943 6463 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1206 03:40:54.580950 6463 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1206 03:40:54.581001 6463 factory.go:656] Stopping watch factory\\\\nI1206 03:40:54.581036 6463 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 03:40:54.581050 6463 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 03:40:54.581063 6463 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 03:40:54.581074 6463 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 03:40:54.581085 6463 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1206 03:40:54.581102 6463 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:53Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:41:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3cb210bfc97a439a7ff98c02c59f8ad16bc400dc214a616a44b43a8b01e2782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pvm6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:23Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.296590 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:23Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.309561 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.309615 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.309639 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.309669 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.309691 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:23Z","lastTransitionTime":"2025-12-06T03:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.318695 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtw2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3a3bac2-3923-4507-a4de-1556b3ad473e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fed5ee4189d54456a3216f4fdb1ecf003fa47e29dbfe4b2b2f301edfb897d6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtw2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:23Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.412690 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.412725 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.412735 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.412769 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.412779 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:23Z","lastTransitionTime":"2025-12-06T03:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.516695 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.516809 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.516838 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.516869 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.516891 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:23Z","lastTransitionTime":"2025-12-06T03:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.619898 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.619987 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.620007 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.620032 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.620059 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:23Z","lastTransitionTime":"2025-12-06T03:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.723748 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.723843 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.723862 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.723888 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.723909 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:23Z","lastTransitionTime":"2025-12-06T03:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.827276 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.827406 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.827432 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.827466 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.827491 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:23Z","lastTransitionTime":"2025-12-06T03:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.930516 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.930557 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.930568 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.930584 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.930595 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:23Z","lastTransitionTime":"2025-12-06T03:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.963172 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pvm6q_eafce4e0-e7fb-4877-b0ab-3283829ba1ef/ovnkube-controller/3.log" Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.964052 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pvm6q_eafce4e0-e7fb-4877-b0ab-3283829ba1ef/ovnkube-controller/2.log" Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.968365 4802 generic.go:334] "Generic (PLEG): container finished" podID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerID="bf554aaba0e773f6a6c1f29bb1819b7ae25b2b2db255c4bd0a16c512f314e9cf" exitCode=1 Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.968465 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" event={"ID":"eafce4e0-e7fb-4877-b0ab-3283829ba1ef","Type":"ContainerDied","Data":"bf554aaba0e773f6a6c1f29bb1819b7ae25b2b2db255c4bd0a16c512f314e9cf"} Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.968571 4802 scope.go:117] "RemoveContainer" containerID="3870b61bce487c69724f668c38fc02190ff97b5e3b1434d66ed624412d18319b" Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.970266 4802 scope.go:117] "RemoveContainer" containerID="bf554aaba0e773f6a6c1f29bb1819b7ae25b2b2db255c4bd0a16c512f314e9cf" Dec 06 03:41:23 crc kubenswrapper[4802]: E1206 03:41:23.970929 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pvm6q_openshift-ovn-kubernetes(eafce4e0-e7fb-4877-b0ab-3283829ba1ef)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" podUID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" Dec 06 03:41:23 crc kubenswrapper[4802]: I1206 03:41:23.989161 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ngspk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34808ab2-d21b-40b6-8fe3-e18bf9e5fc46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52509d21f5b4bfbd458162cba176ff82c627c7c4671204addf9d76f7392fd760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qkf8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9685a4d5d4f40ff4f59b6da6193785f340890453d1e422d3421a621dc7bd02f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qkf8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ngspk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:23Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.007257 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5ltx2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8881599e-71e5-4578-b3f9-21f4fa2f57aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbm6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbm6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5ltx2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:24Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.028613 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1721e985-40bd-4b2d-a0e8-d4e365094ba6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a895e0cc04880c3287ba725ac5484835235f1e066a94ce98fcf4b6133c5cbd2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a21a97385f856ec996105b597d0ae81c56ed28b625ef3f2905e1af6ec8f6a45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://646e01a58908ba2e703aae5eef53e1b981b0f0c3d1d3202487f2a8b94ee3f55d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31752b84d86bda30792b29e4c6ed963753cc576c5de784395e8e6dcbebaced47\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"ver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:40:26.813730 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814130 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:40:26.814139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:40:26.814143 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:40:26.814145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 03:40:26.821403 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821442 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821482 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821496 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 03:40:26.821509 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821514 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 03:40:26.821612 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 03:40:26.821619 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1206 03:40:26.822049 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16a92d501c9e626fe91e4a6793e426d10eb292cfbab766db8b8cc72e5dae94ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:24Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.033941 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.034013 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.034038 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.034071 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.034096 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:24Z","lastTransitionTime":"2025-12-06T03:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.050021 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd230730-a9a1-47ed-b625-25e32231f271\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce321001135d0a2b878316f8003bf1bc54cba9a36f6a654ba376e1e7c0b4da71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af83814e6b2c5ed803157308ab895551e0826fd02f1ecc8b2751145c09d359da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9e6b48b5701dd330ef91712cdcd160afe9831a3881038bfb1e0397a9c8a3aa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2d359853c738e3e0e264f1a6a919b17afc71a362821b3ea170cda6695e3593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2d359853c738e3e0e264f1a6a919b17afc71a362821b3ea170cda6695e3593f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:24Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.067727 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-b2knv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a3ae46f-03b5-4999-9d58-c38ee222f68a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76665d4c127445009fd66078827b54607c9ef812c8011d1c0b4cde1c0516853a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk4hj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-b2knv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:24Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.084690 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22rrq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03112a5d-5882-4fac-9ae5-13525ff82fe0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://831dab65557627cbce0d055459052dae5b594f33b66858367ccd034ff6811924\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0205b089b12c42688203bb7ec0f27dc5f9a577dabdb446b811f184046f85dc3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:41:14Z\\\",\\\"message\\\":\\\"2025-12-06T03:40:28+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_384cd081-7034-46cf-bb48-841728494c60\\\\n2025-12-06T03:40:28+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_384cd081-7034-46cf-bb48-841728494c60 to /host/opt/cni/bin/\\\\n2025-12-06T03:40:29Z [verbose] multus-daemon started\\\\n2025-12-06T03:40:29Z [verbose] Readiness Indicator file check\\\\n2025-12-06T03:41:14Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:41:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trlt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22rrq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:24Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.096224 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2466610962332c6ca120fe055e4e8a3fc623088119dc1a412e06b745176f89e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:24Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.112107 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:24Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.128598 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6df38316-e0d3-4018-8d27-3620eba3a68d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57381e9589a2dc04a4d2733c37985f42594182128a374b7be0f471ea69138b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bf5e41701c7a2d45adc4740ceb4ac1785db17bc7cb15daed439f7556812d1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpxxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:24Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.136856 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.136908 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.136921 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.136939 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.136952 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:24Z","lastTransitionTime":"2025-12-06T03:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.142172 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kj7hf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd46dfa5-7d4a-49ed-8199-523f10cebdf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5eee8a0cf96ceb3785dac9df31bf8d31715d1c88fbea17c5e582b0a4554bb63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmkcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kj7hf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:24Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.155310 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9de75acd-1ba9-4322-ae49-791724b4b6a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://990943a2a48d79458297f4e5cf4178bbd6e9eafbcadc154a803846c589474e85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18a08a15442f04e37a5aa71afbcc293e80b876477d60cbe71d562c75d141a9ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18a08a15442f04e37a5aa71afbcc293e80b876477d60cbe71d562c75d141a9ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:24Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.175909 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2093b6-c4b5-4db8-be16-fbc859885e8d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0d6e582161f0d9126126e03b27211be22f6b5f797154928c92fa13a8118f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03004a29aefe7fe2452fb9f6663c0807afccb055d7ddbf6b907f305e6a19c81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1801ac64db4ba91c2e8a249e3651d3e50ab62e9f2c87312ce9bf7f38518a82cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6e454507cadbb982a566a3692995aaa3c733ff1cd7fb4870476602190e1eeba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:24Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.194596 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:24Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.210608 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fad9b63d84f41fccf696cd313d7ca7f72b3d3251d6c1b7341573a3fadf1e378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:24Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.232866 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb3f278fff47c900c56d8a563d5ed75a2ae84e2034b0f45aad4083790b9e7946\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b10ca16ac26904ec5e6245344b71d836888cae9b069cae6a7a4831d609752db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:24Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.240469 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.240506 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.240519 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.240537 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.240549 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:24Z","lastTransitionTime":"2025-12-06T03:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.250210 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:24Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.275641 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtw2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3a3bac2-3923-4507-a4de-1556b3ad473e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fed5ee4189d54456a3216f4fdb1ecf003fa47e29dbfe4b2b2f301edfb897d6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtw2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:24Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.307881 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7730de4f033561ad77046e028cd1a3f0bfe2775387cbdfa35f47ea79909ec4cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a158aebd701ef46225d3e37109ae28ab58d107087bf06650c4b6d0055a3bbba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2357f37497fde64937eca7b38693893af745ab278653ebc79406e122a6d4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8941bd25107333b264d3f081a811fdfeea23ad92193fc61367665fd2cd1209c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f9cd4dc75e63f09f60b48b6da284643910bb32ba51f2a808cb4d9803f30042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6796f65ee9c5bd0c18f205697135d17677c5daf3ba378cc992cba795e0188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf554aaba0e773f6a6c1f29bb1819b7ae25b2b2db255c4bd0a16c512f314e9cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3870b61bce487c69724f668c38fc02190ff97b5e3b1434d66ed624412d18319b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:40:54Z\\\",\\\"message\\\":\\\"/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 03:40:54.580644 6463 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 03:40:54.580730 6463 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 03:40:54.580797 6463 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 03:40:54.580836 6463 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 03:40:54.580845 6463 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 03:40:54.580895 6463 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1206 03:40:54.580943 6463 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1206 03:40:54.580950 6463 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1206 03:40:54.581001 6463 factory.go:656] Stopping watch factory\\\\nI1206 03:40:54.581036 6463 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 03:40:54.581050 6463 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 03:40:54.581063 6463 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 03:40:54.581074 6463 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 03:40:54.581085 6463 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1206 03:40:54.581102 6463 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:53Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf554aaba0e773f6a6c1f29bb1819b7ae25b2b2db255c4bd0a16c512f314e9cf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:41:23Z\\\",\\\"message\\\":\\\"42b7e8 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 03:41:22.975067 6823 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1206 03:41:22.975567 6823 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 03:41:22.975037 6823 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-kj7hf in node crc\\\\nI1206 03:41:22.975614 6823 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-kj7hf after 0 failed attempt(s)\\\\nI1206 03:41:22.975624 6823 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-kj7hf\\\\nF1206 03:41:22.974932 6823 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:41:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3cb210bfc97a439a7ff98c02c59f8ad16bc400dc214a616a44b43a8b01e2782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pvm6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:24Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.377407 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.377448 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.377457 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.377473 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.377481 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:24Z","lastTransitionTime":"2025-12-06T03:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.449587 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.449619 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.449627 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.449778 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:41:24 crc kubenswrapper[4802]: E1206 03:41:24.449778 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:41:24 crc kubenswrapper[4802]: E1206 03:41:24.450345 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5ltx2" podUID="8881599e-71e5-4578-b3f9-21f4fa2f57aa" Dec 06 03:41:24 crc kubenswrapper[4802]: E1206 03:41:24.450498 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:41:24 crc kubenswrapper[4802]: E1206 03:41:24.450579 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.480717 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.480784 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.480793 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.480810 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.480819 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:24Z","lastTransitionTime":"2025-12-06T03:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.584347 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.584634 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.584797 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.584924 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.585105 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:24Z","lastTransitionTime":"2025-12-06T03:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.688632 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.688700 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.688718 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.688781 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.688802 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:24Z","lastTransitionTime":"2025-12-06T03:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.792017 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.792097 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.792114 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.792144 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.792161 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:24Z","lastTransitionTime":"2025-12-06T03:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.895010 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.895118 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.895139 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.895163 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.895182 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:24Z","lastTransitionTime":"2025-12-06T03:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.974089 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pvm6q_eafce4e0-e7fb-4877-b0ab-3283829ba1ef/ovnkube-controller/3.log" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.979261 4802 scope.go:117] "RemoveContainer" containerID="bf554aaba0e773f6a6c1f29bb1819b7ae25b2b2db255c4bd0a16c512f314e9cf" Dec 06 03:41:24 crc kubenswrapper[4802]: E1206 03:41:24.979428 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pvm6q_openshift-ovn-kubernetes(eafce4e0-e7fb-4877-b0ab-3283829ba1ef)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" podUID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.997780 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.997859 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.997901 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.997920 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:24 crc kubenswrapper[4802]: I1206 03:41:24.997936 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:24Z","lastTransitionTime":"2025-12-06T03:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.000442 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2466610962332c6ca120fe055e4e8a3fc623088119dc1a412e06b745176f89e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:24Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.020198 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:25Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.041061 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6df38316-e0d3-4018-8d27-3620eba3a68d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57381e9589a2dc04a4d2733c37985f42594182128a374b7be0f471ea69138b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bf5e41701c7a2d45adc4740ceb4ac1785db17bc7cb15daed439f7556812d1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpxxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:25Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.055946 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kj7hf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd46dfa5-7d4a-49ed-8199-523f10cebdf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5eee8a0cf96ceb3785dac9df31bf8d31715d1c88fbea17c5e582b0a4554bb63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmkcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kj7hf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:25Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.076253 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb3f278fff47c900c56d8a563d5ed75a2ae84e2034b0f45aad4083790b9e7946\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b10ca16ac26904ec5e6245344b71d836888cae9b069cae6a7a4831d609752db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:25Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.090107 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9de75acd-1ba9-4322-ae49-791724b4b6a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://990943a2a48d79458297f4e5cf4178bbd6e9eafbcadc154a803846c589474e85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18a08a15442f04e37a5aa71afbcc293e80b876477d60cbe71d562c75d141a9ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18a08a15442f04e37a5aa71afbcc293e80b876477d60cbe71d562c75d141a9ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:25Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.100540 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.100594 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.100609 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.100629 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.100643 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:25Z","lastTransitionTime":"2025-12-06T03:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.116534 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2093b6-c4b5-4db8-be16-fbc859885e8d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0d6e582161f0d9126126e03b27211be22f6b5f797154928c92fa13a8118f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03004a29aefe7fe2452fb9f6663c0807afccb055d7ddbf6b907f305e6a19c81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1801ac64db4ba91c2e8a249e3651d3e50ab62e9f2c87312ce9bf7f38518a82cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6e454507cadbb982a566a3692995aaa3c733ff1cd7fb4870476602190e1eeba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:25Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.136448 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:25Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.157685 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fad9b63d84f41fccf696cd313d7ca7f72b3d3251d6c1b7341573a3fadf1e378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:25Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.175269 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:25Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.202335 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtw2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3a3bac2-3923-4507-a4de-1556b3ad473e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fed5ee4189d54456a3216f4fdb1ecf003fa47e29dbfe4b2b2f301edfb897d6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtw2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:25Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.202836 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.202890 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.202907 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.202930 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.202947 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:25Z","lastTransitionTime":"2025-12-06T03:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.228142 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7730de4f033561ad77046e028cd1a3f0bfe2775387cbdfa35f47ea79909ec4cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a158aebd701ef46225d3e37109ae28ab58d107087bf06650c4b6d0055a3bbba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2357f37497fde64937eca7b38693893af745ab278653ebc79406e122a6d4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8941bd25107333b264d3f081a811fdfeea23ad92193fc61367665fd2cd1209c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f9cd4dc75e63f09f60b48b6da284643910bb32ba51f2a808cb4d9803f30042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6796f65ee9c5bd0c18f205697135d17677c5daf3ba378cc992cba795e0188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf554aaba0e773f6a6c1f29bb1819b7ae25b2b2db255c4bd0a16c512f314e9cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf554aaba0e773f6a6c1f29bb1819b7ae25b2b2db255c4bd0a16c512f314e9cf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:41:23Z\\\",\\\"message\\\":\\\"42b7e8 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 03:41:22.975067 6823 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1206 03:41:22.975567 6823 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 03:41:22.975037 6823 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-kj7hf in node crc\\\\nI1206 03:41:22.975614 6823 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-kj7hf after 0 failed attempt(s)\\\\nI1206 03:41:22.975624 6823 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-kj7hf\\\\nF1206 03:41:22.974932 6823 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:41:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pvm6q_openshift-ovn-kubernetes(eafce4e0-e7fb-4877-b0ab-3283829ba1ef)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3cb210bfc97a439a7ff98c02c59f8ad16bc400dc214a616a44b43a8b01e2782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pvm6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:25Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.253017 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22rrq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03112a5d-5882-4fac-9ae5-13525ff82fe0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://831dab65557627cbce0d055459052dae5b594f33b66858367ccd034ff6811924\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0205b089b12c42688203bb7ec0f27dc5f9a577dabdb446b811f184046f85dc3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:41:14Z\\\",\\\"message\\\":\\\"2025-12-06T03:40:28+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_384cd081-7034-46cf-bb48-841728494c60\\\\n2025-12-06T03:40:28+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_384cd081-7034-46cf-bb48-841728494c60 to /host/opt/cni/bin/\\\\n2025-12-06T03:40:29Z [verbose] multus-daemon started\\\\n2025-12-06T03:40:29Z [verbose] Readiness Indicator file check\\\\n2025-12-06T03:41:14Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:41:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trlt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22rrq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:25Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.272346 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ngspk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34808ab2-d21b-40b6-8fe3-e18bf9e5fc46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52509d21f5b4bfbd458162cba176ff82c627c7c4671204addf9d76f7392fd760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qkf8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9685a4d5d4f40ff4f59b6da6193785f340890453d1e422d3421a621dc7bd02f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qkf8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ngspk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:25Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.289226 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5ltx2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8881599e-71e5-4578-b3f9-21f4fa2f57aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbm6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbm6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5ltx2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:25Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.305793 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.305856 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.305878 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.305907 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.305929 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:25Z","lastTransitionTime":"2025-12-06T03:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.312022 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1721e985-40bd-4b2d-a0e8-d4e365094ba6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a895e0cc04880c3287ba725ac5484835235f1e066a94ce98fcf4b6133c5cbd2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a21a97385f856ec996105b597d0ae81c56ed28b625ef3f2905e1af6ec8f6a45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://646e01a58908ba2e703aae5eef53e1b981b0f0c3d1d3202487f2a8b94ee3f55d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31752b84d86bda30792b29e4c6ed963753cc576c5de784395e8e6dcbebaced47\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"ver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:40:26.813730 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814130 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:40:26.814139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:40:26.814143 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:40:26.814145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 03:40:26.821403 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821442 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821482 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821496 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 03:40:26.821509 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821514 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 03:40:26.821612 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 03:40:26.821619 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1206 03:40:26.822049 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16a92d501c9e626fe91e4a6793e426d10eb292cfbab766db8b8cc72e5dae94ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:25Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.329838 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd230730-a9a1-47ed-b625-25e32231f271\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce321001135d0a2b878316f8003bf1bc54cba9a36f6a654ba376e1e7c0b4da71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af83814e6b2c5ed803157308ab895551e0826fd02f1ecc8b2751145c09d359da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9e6b48b5701dd330ef91712cdcd160afe9831a3881038bfb1e0397a9c8a3aa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2d359853c738e3e0e264f1a6a919b17afc71a362821b3ea170cda6695e3593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2d359853c738e3e0e264f1a6a919b17afc71a362821b3ea170cda6695e3593f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:25Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.341905 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-b2knv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a3ae46f-03b5-4999-9d58-c38ee222f68a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76665d4c127445009fd66078827b54607c9ef812c8011d1c0b4cde1c0516853a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk4hj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-b2knv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:25Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.409284 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.409364 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.409385 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.409411 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.409432 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:25Z","lastTransitionTime":"2025-12-06T03:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.512411 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.512474 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.512498 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.512532 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.512555 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:25Z","lastTransitionTime":"2025-12-06T03:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.615930 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.615999 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.616012 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.616031 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.616042 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:25Z","lastTransitionTime":"2025-12-06T03:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.719391 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.719455 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.719471 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.719498 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.719514 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:25Z","lastTransitionTime":"2025-12-06T03:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.822987 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.823037 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.823082 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.823103 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.823115 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:25Z","lastTransitionTime":"2025-12-06T03:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.927908 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.927989 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.928008 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.928032 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:25 crc kubenswrapper[4802]: I1206 03:41:25.928050 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:25Z","lastTransitionTime":"2025-12-06T03:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:26 crc kubenswrapper[4802]: I1206 03:41:26.030815 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:26 crc kubenswrapper[4802]: I1206 03:41:26.030880 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:26 crc kubenswrapper[4802]: I1206 03:41:26.030903 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:26 crc kubenswrapper[4802]: I1206 03:41:26.030931 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:26 crc kubenswrapper[4802]: I1206 03:41:26.030953 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:26Z","lastTransitionTime":"2025-12-06T03:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:26 crc kubenswrapper[4802]: I1206 03:41:26.134468 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:26 crc kubenswrapper[4802]: I1206 03:41:26.134543 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:26 crc kubenswrapper[4802]: I1206 03:41:26.134568 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:26 crc kubenswrapper[4802]: I1206 03:41:26.134602 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:26 crc kubenswrapper[4802]: I1206 03:41:26.134668 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:26Z","lastTransitionTime":"2025-12-06T03:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:26 crc kubenswrapper[4802]: I1206 03:41:26.237892 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:26 crc kubenswrapper[4802]: I1206 03:41:26.237928 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:26 crc kubenswrapper[4802]: I1206 03:41:26.237940 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:26 crc kubenswrapper[4802]: I1206 03:41:26.237956 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:26 crc kubenswrapper[4802]: I1206 03:41:26.237968 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:26Z","lastTransitionTime":"2025-12-06T03:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:26 crc kubenswrapper[4802]: I1206 03:41:26.341004 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:26 crc kubenswrapper[4802]: I1206 03:41:26.341055 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:26 crc kubenswrapper[4802]: I1206 03:41:26.341072 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:26 crc kubenswrapper[4802]: I1206 03:41:26.341099 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:26 crc kubenswrapper[4802]: I1206 03:41:26.341118 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:26Z","lastTransitionTime":"2025-12-06T03:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:26 crc kubenswrapper[4802]: I1206 03:41:26.444024 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:26 crc kubenswrapper[4802]: I1206 03:41:26.444303 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:26 crc kubenswrapper[4802]: I1206 03:41:26.444428 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:26 crc kubenswrapper[4802]: I1206 03:41:26.444548 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:26 crc kubenswrapper[4802]: I1206 03:41:26.444664 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:26Z","lastTransitionTime":"2025-12-06T03:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:26 crc kubenswrapper[4802]: I1206 03:41:26.449334 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:41:26 crc kubenswrapper[4802]: I1206 03:41:26.449494 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:41:26 crc kubenswrapper[4802]: I1206 03:41:26.449393 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:41:26 crc kubenswrapper[4802]: I1206 03:41:26.449345 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:41:26 crc kubenswrapper[4802]: E1206 03:41:26.449936 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5ltx2" podUID="8881599e-71e5-4578-b3f9-21f4fa2f57aa" Dec 06 03:41:26 crc kubenswrapper[4802]: E1206 03:41:26.450162 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:41:26 crc kubenswrapper[4802]: E1206 03:41:26.450344 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:41:26 crc kubenswrapper[4802]: E1206 03:41:26.450478 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:41:26 crc kubenswrapper[4802]: I1206 03:41:26.547377 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:26 crc kubenswrapper[4802]: I1206 03:41:26.547416 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:26 crc kubenswrapper[4802]: I1206 03:41:26.547425 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:26 crc kubenswrapper[4802]: I1206 03:41:26.547440 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:26 crc kubenswrapper[4802]: I1206 03:41:26.547451 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:26Z","lastTransitionTime":"2025-12-06T03:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:26 crc kubenswrapper[4802]: I1206 03:41:26.650079 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:26 crc kubenswrapper[4802]: I1206 03:41:26.650318 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:26 crc kubenswrapper[4802]: I1206 03:41:26.650384 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:26 crc kubenswrapper[4802]: I1206 03:41:26.650450 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:26 crc kubenswrapper[4802]: I1206 03:41:26.650507 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:26Z","lastTransitionTime":"2025-12-06T03:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:26 crc kubenswrapper[4802]: I1206 03:41:26.753174 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:26 crc kubenswrapper[4802]: I1206 03:41:26.753229 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:26 crc kubenswrapper[4802]: I1206 03:41:26.753246 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:26 crc kubenswrapper[4802]: I1206 03:41:26.753269 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:26 crc kubenswrapper[4802]: I1206 03:41:26.753291 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:26Z","lastTransitionTime":"2025-12-06T03:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:26 crc kubenswrapper[4802]: I1206 03:41:26.856134 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:26 crc kubenswrapper[4802]: I1206 03:41:26.856171 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:26 crc kubenswrapper[4802]: I1206 03:41:26.856182 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:26 crc kubenswrapper[4802]: I1206 03:41:26.856198 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:26 crc kubenswrapper[4802]: I1206 03:41:26.856209 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:26Z","lastTransitionTime":"2025-12-06T03:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:26 crc kubenswrapper[4802]: I1206 03:41:26.958809 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:26 crc kubenswrapper[4802]: I1206 03:41:26.958871 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:26 crc kubenswrapper[4802]: I1206 03:41:26.958893 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:26 crc kubenswrapper[4802]: I1206 03:41:26.958923 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:26 crc kubenswrapper[4802]: I1206 03:41:26.958946 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:26Z","lastTransitionTime":"2025-12-06T03:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.062076 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.062118 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.062128 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.062143 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.062153 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:27Z","lastTransitionTime":"2025-12-06T03:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.164868 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.164918 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.164934 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.164958 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.164977 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:27Z","lastTransitionTime":"2025-12-06T03:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.267711 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.267792 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.267810 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.267829 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.267842 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:27Z","lastTransitionTime":"2025-12-06T03:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.371420 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.371514 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.371539 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.371602 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.371627 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:27Z","lastTransitionTime":"2025-12-06T03:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.467681 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kj7hf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd46dfa5-7d4a-49ed-8199-523f10cebdf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5eee8a0cf96ceb3785dac9df31bf8d31715d1c88fbea17c5e582b0a4554bb63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmkcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kj7hf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:27Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.475003 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.475057 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.475078 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.475102 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.475119 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:27Z","lastTransitionTime":"2025-12-06T03:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.486953 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2466610962332c6ca120fe055e4e8a3fc623088119dc1a412e06b745176f89e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:27Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.508452 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:27Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.525796 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6df38316-e0d3-4018-8d27-3620eba3a68d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57381e9589a2dc04a4d2733c37985f42594182128a374b7be0f471ea69138b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bf5e41701c7a2d45adc4740ceb4ac1785db17bc7cb15daed439f7556812d1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpxxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:27Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.551535 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fad9b63d84f41fccf696cd313d7ca7f72b3d3251d6c1b7341573a3fadf1e378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:27Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.572575 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb3f278fff47c900c56d8a563d5ed75a2ae84e2034b0f45aad4083790b9e7946\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b10ca16ac26904ec5e6245344b71d836888cae9b069cae6a7a4831d609752db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:27Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.578029 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.578062 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.578072 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.578087 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.578099 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:27Z","lastTransitionTime":"2025-12-06T03:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.587739 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9de75acd-1ba9-4322-ae49-791724b4b6a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://990943a2a48d79458297f4e5cf4178bbd6e9eafbcadc154a803846c589474e85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18a08a15442f04e37a5aa71afbcc293e80b876477d60cbe71d562c75d141a9ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18a08a15442f04e37a5aa71afbcc293e80b876477d60cbe71d562c75d141a9ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:27Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.607850 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2093b6-c4b5-4db8-be16-fbc859885e8d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0d6e582161f0d9126126e03b27211be22f6b5f797154928c92fa13a8118f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03004a29aefe7fe2452fb9f6663c0807afccb055d7ddbf6b907f305e6a19c81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1801ac64db4ba91c2e8a249e3651d3e50ab62e9f2c87312ce9bf7f38518a82cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6e454507cadbb982a566a3692995aaa3c733ff1cd7fb4870476602190e1eeba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:27Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.625119 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:27Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.647465 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:27Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.666882 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtw2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3a3bac2-3923-4507-a4de-1556b3ad473e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fed5ee4189d54456a3216f4fdb1ecf003fa47e29dbfe4b2b2f301edfb897d6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtw2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:27Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.680976 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.681035 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.681053 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.681078 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.681095 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:27Z","lastTransitionTime":"2025-12-06T03:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.690957 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7730de4f033561ad77046e028cd1a3f0bfe2775387cbdfa35f47ea79909ec4cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a158aebd701ef46225d3e37109ae28ab58d107087bf06650c4b6d0055a3bbba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2357f37497fde64937eca7b38693893af745ab278653ebc79406e122a6d4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8941bd25107333b264d3f081a811fdfeea23ad92193fc61367665fd2cd1209c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f9cd4dc75e63f09f60b48b6da284643910bb32ba51f2a808cb4d9803f30042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6796f65ee9c5bd0c18f205697135d17677c5daf3ba378cc992cba795e0188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf554aaba0e773f6a6c1f29bb1819b7ae25b2b2db255c4bd0a16c512f314e9cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf554aaba0e773f6a6c1f29bb1819b7ae25b2b2db255c4bd0a16c512f314e9cf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:41:23Z\\\",\\\"message\\\":\\\"42b7e8 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 03:41:22.975067 6823 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1206 03:41:22.975567 6823 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 03:41:22.975037 6823 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-kj7hf in node crc\\\\nI1206 03:41:22.975614 6823 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-kj7hf after 0 failed attempt(s)\\\\nI1206 03:41:22.975624 6823 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-kj7hf\\\\nF1206 03:41:22.974932 6823 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:41:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pvm6q_openshift-ovn-kubernetes(eafce4e0-e7fb-4877-b0ab-3283829ba1ef)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3cb210bfc97a439a7ff98c02c59f8ad16bc400dc214a616a44b43a8b01e2782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pvm6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:27Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.705773 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-b2knv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a3ae46f-03b5-4999-9d58-c38ee222f68a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76665d4c127445009fd66078827b54607c9ef812c8011d1c0b4cde1c0516853a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk4hj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-b2knv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:27Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.727109 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22rrq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03112a5d-5882-4fac-9ae5-13525ff82fe0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://831dab65557627cbce0d055459052dae5b594f33b66858367ccd034ff6811924\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0205b089b12c42688203bb7ec0f27dc5f9a577dabdb446b811f184046f85dc3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:41:14Z\\\",\\\"message\\\":\\\"2025-12-06T03:40:28+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_384cd081-7034-46cf-bb48-841728494c60\\\\n2025-12-06T03:40:28+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_384cd081-7034-46cf-bb48-841728494c60 to /host/opt/cni/bin/\\\\n2025-12-06T03:40:29Z [verbose] multus-daemon started\\\\n2025-12-06T03:40:29Z [verbose] Readiness Indicator file check\\\\n2025-12-06T03:41:14Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:41:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trlt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22rrq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:27Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.740928 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ngspk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34808ab2-d21b-40b6-8fe3-e18bf9e5fc46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52509d21f5b4bfbd458162cba176ff82c627c7c4671204addf9d76f7392fd760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qkf8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9685a4d5d4f40ff4f59b6da6193785f340890453d1e422d3421a621dc7bd02f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qkf8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ngspk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:27Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.753035 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5ltx2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8881599e-71e5-4578-b3f9-21f4fa2f57aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbm6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbm6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5ltx2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:27Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.769830 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1721e985-40bd-4b2d-a0e8-d4e365094ba6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a895e0cc04880c3287ba725ac5484835235f1e066a94ce98fcf4b6133c5cbd2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a21a97385f856ec996105b597d0ae81c56ed28b625ef3f2905e1af6ec8f6a45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://646e01a58908ba2e703aae5eef53e1b981b0f0c3d1d3202487f2a8b94ee3f55d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31752b84d86bda30792b29e4c6ed963753cc576c5de784395e8e6dcbebaced47\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"ver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:40:26.813730 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814130 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:40:26.814139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:40:26.814143 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:40:26.814145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 03:40:26.821403 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821442 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821482 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821496 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 03:40:26.821509 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821514 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 03:40:26.821612 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 03:40:26.821619 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1206 03:40:26.822049 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16a92d501c9e626fe91e4a6793e426d10eb292cfbab766db8b8cc72e5dae94ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:27Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.781709 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd230730-a9a1-47ed-b625-25e32231f271\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce321001135d0a2b878316f8003bf1bc54cba9a36f6a654ba376e1e7c0b4da71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af83814e6b2c5ed803157308ab895551e0826fd02f1ecc8b2751145c09d359da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9e6b48b5701dd330ef91712cdcd160afe9831a3881038bfb1e0397a9c8a3aa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2d359853c738e3e0e264f1a6a919b17afc71a362821b3ea170cda6695e3593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2d359853c738e3e0e264f1a6a919b17afc71a362821b3ea170cda6695e3593f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:27Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.783477 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.783519 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.783536 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.783557 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.783574 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:27Z","lastTransitionTime":"2025-12-06T03:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.887521 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.887573 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.887589 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.887613 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.887628 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:27Z","lastTransitionTime":"2025-12-06T03:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.989443 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.989488 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.989499 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.989515 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:27 crc kubenswrapper[4802]: I1206 03:41:27.989527 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:27Z","lastTransitionTime":"2025-12-06T03:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:28 crc kubenswrapper[4802]: I1206 03:41:28.091718 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:28 crc kubenswrapper[4802]: I1206 03:41:28.091777 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:28 crc kubenswrapper[4802]: I1206 03:41:28.091786 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:28 crc kubenswrapper[4802]: I1206 03:41:28.091800 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:28 crc kubenswrapper[4802]: I1206 03:41:28.091809 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:28Z","lastTransitionTime":"2025-12-06T03:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:28 crc kubenswrapper[4802]: I1206 03:41:28.193936 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:28 crc kubenswrapper[4802]: I1206 03:41:28.193975 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:28 crc kubenswrapper[4802]: I1206 03:41:28.193987 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:28 crc kubenswrapper[4802]: I1206 03:41:28.194003 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:28 crc kubenswrapper[4802]: I1206 03:41:28.194014 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:28Z","lastTransitionTime":"2025-12-06T03:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:28 crc kubenswrapper[4802]: I1206 03:41:28.298959 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:28 crc kubenswrapper[4802]: I1206 03:41:28.299038 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:28 crc kubenswrapper[4802]: I1206 03:41:28.299061 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:28 crc kubenswrapper[4802]: I1206 03:41:28.299092 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:28 crc kubenswrapper[4802]: I1206 03:41:28.299115 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:28Z","lastTransitionTime":"2025-12-06T03:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:28 crc kubenswrapper[4802]: I1206 03:41:28.402169 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:28 crc kubenswrapper[4802]: I1206 03:41:28.402242 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:28 crc kubenswrapper[4802]: I1206 03:41:28.402260 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:28 crc kubenswrapper[4802]: I1206 03:41:28.402288 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:28 crc kubenswrapper[4802]: I1206 03:41:28.402306 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:28Z","lastTransitionTime":"2025-12-06T03:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:28 crc kubenswrapper[4802]: I1206 03:41:28.449912 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:41:28 crc kubenswrapper[4802]: I1206 03:41:28.449939 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:41:28 crc kubenswrapper[4802]: E1206 03:41:28.450046 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5ltx2" podUID="8881599e-71e5-4578-b3f9-21f4fa2f57aa" Dec 06 03:41:28 crc kubenswrapper[4802]: I1206 03:41:28.450091 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:41:28 crc kubenswrapper[4802]: I1206 03:41:28.450123 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:41:28 crc kubenswrapper[4802]: E1206 03:41:28.450294 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:41:28 crc kubenswrapper[4802]: E1206 03:41:28.450361 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:41:28 crc kubenswrapper[4802]: E1206 03:41:28.450535 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:41:28 crc kubenswrapper[4802]: I1206 03:41:28.505943 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:28 crc kubenswrapper[4802]: I1206 03:41:28.506021 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:28 crc kubenswrapper[4802]: I1206 03:41:28.506045 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:28 crc kubenswrapper[4802]: I1206 03:41:28.506075 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:28 crc kubenswrapper[4802]: I1206 03:41:28.506098 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:28Z","lastTransitionTime":"2025-12-06T03:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:28 crc kubenswrapper[4802]: I1206 03:41:28.610019 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:28 crc kubenswrapper[4802]: I1206 03:41:28.610082 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:28 crc kubenswrapper[4802]: I1206 03:41:28.610104 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:28 crc kubenswrapper[4802]: I1206 03:41:28.610136 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:28 crc kubenswrapper[4802]: I1206 03:41:28.610161 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:28Z","lastTransitionTime":"2025-12-06T03:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:28 crc kubenswrapper[4802]: I1206 03:41:28.712705 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:28 crc kubenswrapper[4802]: I1206 03:41:28.712800 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:28 crc kubenswrapper[4802]: I1206 03:41:28.712823 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:28 crc kubenswrapper[4802]: I1206 03:41:28.712854 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:28 crc kubenswrapper[4802]: I1206 03:41:28.712875 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:28Z","lastTransitionTime":"2025-12-06T03:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:28 crc kubenswrapper[4802]: I1206 03:41:28.815714 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:28 crc kubenswrapper[4802]: I1206 03:41:28.815825 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:28 crc kubenswrapper[4802]: I1206 03:41:28.815853 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:28 crc kubenswrapper[4802]: I1206 03:41:28.815885 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:28 crc kubenswrapper[4802]: I1206 03:41:28.815907 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:28Z","lastTransitionTime":"2025-12-06T03:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:28 crc kubenswrapper[4802]: I1206 03:41:28.919675 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:28 crc kubenswrapper[4802]: I1206 03:41:28.919744 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:28 crc kubenswrapper[4802]: I1206 03:41:28.919809 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:28 crc kubenswrapper[4802]: I1206 03:41:28.919839 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:28 crc kubenswrapper[4802]: I1206 03:41:28.919863 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:28Z","lastTransitionTime":"2025-12-06T03:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:29 crc kubenswrapper[4802]: I1206 03:41:29.022243 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:29 crc kubenswrapper[4802]: I1206 03:41:29.022304 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:29 crc kubenswrapper[4802]: I1206 03:41:29.022323 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:29 crc kubenswrapper[4802]: I1206 03:41:29.022345 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:29 crc kubenswrapper[4802]: I1206 03:41:29.022363 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:29Z","lastTransitionTime":"2025-12-06T03:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:29 crc kubenswrapper[4802]: I1206 03:41:29.125173 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:29 crc kubenswrapper[4802]: I1206 03:41:29.125235 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:29 crc kubenswrapper[4802]: I1206 03:41:29.125251 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:29 crc kubenswrapper[4802]: I1206 03:41:29.125274 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:29 crc kubenswrapper[4802]: I1206 03:41:29.125291 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:29Z","lastTransitionTime":"2025-12-06T03:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:29 crc kubenswrapper[4802]: I1206 03:41:29.228310 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:29 crc kubenswrapper[4802]: I1206 03:41:29.228374 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:29 crc kubenswrapper[4802]: I1206 03:41:29.228390 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:29 crc kubenswrapper[4802]: I1206 03:41:29.228415 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:29 crc kubenswrapper[4802]: I1206 03:41:29.228432 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:29Z","lastTransitionTime":"2025-12-06T03:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:29 crc kubenswrapper[4802]: I1206 03:41:29.331400 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:29 crc kubenswrapper[4802]: I1206 03:41:29.331456 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:29 crc kubenswrapper[4802]: I1206 03:41:29.331473 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:29 crc kubenswrapper[4802]: I1206 03:41:29.331499 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:29 crc kubenswrapper[4802]: I1206 03:41:29.331515 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:29Z","lastTransitionTime":"2025-12-06T03:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:29 crc kubenswrapper[4802]: I1206 03:41:29.435660 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:29 crc kubenswrapper[4802]: I1206 03:41:29.435732 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:29 crc kubenswrapper[4802]: I1206 03:41:29.435781 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:29 crc kubenswrapper[4802]: I1206 03:41:29.435805 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:29 crc kubenswrapper[4802]: I1206 03:41:29.435823 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:29Z","lastTransitionTime":"2025-12-06T03:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:29 crc kubenswrapper[4802]: I1206 03:41:29.538596 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:29 crc kubenswrapper[4802]: I1206 03:41:29.538662 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:29 crc kubenswrapper[4802]: I1206 03:41:29.538680 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:29 crc kubenswrapper[4802]: I1206 03:41:29.538705 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:29 crc kubenswrapper[4802]: I1206 03:41:29.538721 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:29Z","lastTransitionTime":"2025-12-06T03:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:29 crc kubenswrapper[4802]: I1206 03:41:29.641704 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:29 crc kubenswrapper[4802]: I1206 03:41:29.641813 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:29 crc kubenswrapper[4802]: I1206 03:41:29.641830 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:29 crc kubenswrapper[4802]: I1206 03:41:29.641854 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:29 crc kubenswrapper[4802]: I1206 03:41:29.641870 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:29Z","lastTransitionTime":"2025-12-06T03:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:29 crc kubenswrapper[4802]: I1206 03:41:29.744730 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:29 crc kubenswrapper[4802]: I1206 03:41:29.744797 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:29 crc kubenswrapper[4802]: I1206 03:41:29.744810 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:29 crc kubenswrapper[4802]: I1206 03:41:29.744826 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:29 crc kubenswrapper[4802]: I1206 03:41:29.744837 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:29Z","lastTransitionTime":"2025-12-06T03:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:29 crc kubenswrapper[4802]: I1206 03:41:29.847796 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:29 crc kubenswrapper[4802]: I1206 03:41:29.848197 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:29 crc kubenswrapper[4802]: I1206 03:41:29.848217 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:29 crc kubenswrapper[4802]: I1206 03:41:29.848247 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:29 crc kubenswrapper[4802]: I1206 03:41:29.848266 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:29Z","lastTransitionTime":"2025-12-06T03:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:29 crc kubenswrapper[4802]: I1206 03:41:29.951000 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:29 crc kubenswrapper[4802]: I1206 03:41:29.951075 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:29 crc kubenswrapper[4802]: I1206 03:41:29.951092 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:29 crc kubenswrapper[4802]: I1206 03:41:29.951118 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:29 crc kubenswrapper[4802]: I1206 03:41:29.951135 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:29Z","lastTransitionTime":"2025-12-06T03:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.054236 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.054303 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.054327 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.054354 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.054371 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:30Z","lastTransitionTime":"2025-12-06T03:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.158276 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.158337 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.158366 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.158386 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.158399 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:30Z","lastTransitionTime":"2025-12-06T03:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.261535 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.261564 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.261571 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.261584 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.261595 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:30Z","lastTransitionTime":"2025-12-06T03:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.347421 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.347492 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.347510 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.347535 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.347554 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:30Z","lastTransitionTime":"2025-12-06T03:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:30 crc kubenswrapper[4802]: E1206 03:41:30.379221 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0e01ca3a-758f-44bc-b4a2-30c1bc04fda0\\\",\\\"systemUUID\\\":\\\"c7c815e4-d2a4-4244-b035-f988fc95e215\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:30Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.380572 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:41:30 crc kubenswrapper[4802]: E1206 03:41:30.380861 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 03:41:30 crc kubenswrapper[4802]: E1206 03:41:30.380905 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 03:41:30 crc kubenswrapper[4802]: E1206 03:41:30.380930 4802 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:41:30 crc kubenswrapper[4802]: E1206 03:41:30.381028 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 03:42:34.381000353 +0000 UTC m=+147.252909545 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.388794 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.388861 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.388884 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.388921 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.388944 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:30Z","lastTransitionTime":"2025-12-06T03:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:30 crc kubenswrapper[4802]: E1206 03:41:30.413184 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0e01ca3a-758f-44bc-b4a2-30c1bc04fda0\\\",\\\"systemUUID\\\":\\\"c7c815e4-d2a4-4244-b035-f988fc95e215\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:30Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.421115 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.421179 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.421201 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.421228 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.421247 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:30Z","lastTransitionTime":"2025-12-06T03:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:30 crc kubenswrapper[4802]: E1206 03:41:30.439382 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0e01ca3a-758f-44bc-b4a2-30c1bc04fda0\\\",\\\"systemUUID\\\":\\\"c7c815e4-d2a4-4244-b035-f988fc95e215\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:30Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.443664 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.443707 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.443725 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.443747 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.443803 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:30Z","lastTransitionTime":"2025-12-06T03:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.449792 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.449798 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.449798 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:41:30 crc kubenswrapper[4802]: E1206 03:41:30.450306 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.449796 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:41:30 crc kubenswrapper[4802]: E1206 03:41:30.450426 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:41:30 crc kubenswrapper[4802]: E1206 03:41:30.450573 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:41:30 crc kubenswrapper[4802]: E1206 03:41:30.450961 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5ltx2" podUID="8881599e-71e5-4578-b3f9-21f4fa2f57aa" Dec 06 03:41:30 crc kubenswrapper[4802]: E1206 03:41:30.459710 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0e01ca3a-758f-44bc-b4a2-30c1bc04fda0\\\",\\\"systemUUID\\\":\\\"c7c815e4-d2a4-4244-b035-f988fc95e215\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:30Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.464307 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.464374 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.464400 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.464429 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.464451 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:30Z","lastTransitionTime":"2025-12-06T03:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:30 crc kubenswrapper[4802]: E1206 03:41:30.478318 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0e01ca3a-758f-44bc-b4a2-30c1bc04fda0\\\",\\\"systemUUID\\\":\\\"c7c815e4-d2a4-4244-b035-f988fc95e215\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:30Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:30 crc kubenswrapper[4802]: E1206 03:41:30.478769 4802 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.480383 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.480449 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.480467 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.480493 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.480509 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:30Z","lastTransitionTime":"2025-12-06T03:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.480986 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:41:30 crc kubenswrapper[4802]: E1206 03:41:30.481143 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:42:34.481120359 +0000 UTC m=+147.353029501 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.481310 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:41:30 crc kubenswrapper[4802]: E1206 03:41:30.481425 4802 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.481429 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:41:30 crc kubenswrapper[4802]: E1206 03:41:30.481480 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 03:42:34.481469308 +0000 UTC m=+147.353378490 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.481645 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:41:30 crc kubenswrapper[4802]: E1206 03:41:30.481812 4802 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 03:41:30 crc kubenswrapper[4802]: E1206 03:41:30.481969 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 03:42:34.48195035 +0000 UTC m=+147.353859512 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 03:41:30 crc kubenswrapper[4802]: E1206 03:41:30.481868 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 03:41:30 crc kubenswrapper[4802]: E1206 03:41:30.482151 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 03:41:30 crc kubenswrapper[4802]: E1206 03:41:30.482234 4802 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:41:30 crc kubenswrapper[4802]: E1206 03:41:30.482347 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 03:42:34.482336341 +0000 UTC m=+147.354245503 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.583449 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.583492 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.583503 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.583523 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.583533 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:30Z","lastTransitionTime":"2025-12-06T03:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.686415 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.686538 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.686563 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.686590 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.686608 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:30Z","lastTransitionTime":"2025-12-06T03:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.789812 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.789881 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.789892 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.789915 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.789932 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:30Z","lastTransitionTime":"2025-12-06T03:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.892637 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.892680 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.892688 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.892705 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.892714 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:30Z","lastTransitionTime":"2025-12-06T03:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.994581 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.994635 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.994647 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.994669 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:30 crc kubenswrapper[4802]: I1206 03:41:30.994684 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:30Z","lastTransitionTime":"2025-12-06T03:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:31 crc kubenswrapper[4802]: I1206 03:41:31.097212 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:31 crc kubenswrapper[4802]: I1206 03:41:31.097277 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:31 crc kubenswrapper[4802]: I1206 03:41:31.097296 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:31 crc kubenswrapper[4802]: I1206 03:41:31.097321 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:31 crc kubenswrapper[4802]: I1206 03:41:31.097340 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:31Z","lastTransitionTime":"2025-12-06T03:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:31 crc kubenswrapper[4802]: I1206 03:41:31.199872 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:31 crc kubenswrapper[4802]: I1206 03:41:31.199935 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:31 crc kubenswrapper[4802]: I1206 03:41:31.199952 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:31 crc kubenswrapper[4802]: I1206 03:41:31.199976 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:31 crc kubenswrapper[4802]: I1206 03:41:31.199993 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:31Z","lastTransitionTime":"2025-12-06T03:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:31 crc kubenswrapper[4802]: I1206 03:41:31.302924 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:31 crc kubenswrapper[4802]: I1206 03:41:31.302997 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:31 crc kubenswrapper[4802]: I1206 03:41:31.303021 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:31 crc kubenswrapper[4802]: I1206 03:41:31.303049 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:31 crc kubenswrapper[4802]: I1206 03:41:31.303072 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:31Z","lastTransitionTime":"2025-12-06T03:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:31 crc kubenswrapper[4802]: I1206 03:41:31.406064 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:31 crc kubenswrapper[4802]: I1206 03:41:31.406118 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:31 crc kubenswrapper[4802]: I1206 03:41:31.406134 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:31 crc kubenswrapper[4802]: I1206 03:41:31.406156 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:31 crc kubenswrapper[4802]: I1206 03:41:31.406173 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:31Z","lastTransitionTime":"2025-12-06T03:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:31 crc kubenswrapper[4802]: I1206 03:41:31.508443 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:31 crc kubenswrapper[4802]: I1206 03:41:31.508532 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:31 crc kubenswrapper[4802]: I1206 03:41:31.508564 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:31 crc kubenswrapper[4802]: I1206 03:41:31.508600 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:31 crc kubenswrapper[4802]: I1206 03:41:31.508623 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:31Z","lastTransitionTime":"2025-12-06T03:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:31 crc kubenswrapper[4802]: I1206 03:41:31.612658 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:31 crc kubenswrapper[4802]: I1206 03:41:31.612743 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:31 crc kubenswrapper[4802]: I1206 03:41:31.612807 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:31 crc kubenswrapper[4802]: I1206 03:41:31.612836 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:31 crc kubenswrapper[4802]: I1206 03:41:31.612890 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:31Z","lastTransitionTime":"2025-12-06T03:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:31 crc kubenswrapper[4802]: I1206 03:41:31.715695 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:31 crc kubenswrapper[4802]: I1206 03:41:31.715746 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:31 crc kubenswrapper[4802]: I1206 03:41:31.715804 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:31 crc kubenswrapper[4802]: I1206 03:41:31.715829 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:31 crc kubenswrapper[4802]: I1206 03:41:31.715846 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:31Z","lastTransitionTime":"2025-12-06T03:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:31 crc kubenswrapper[4802]: I1206 03:41:31.819197 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:31 crc kubenswrapper[4802]: I1206 03:41:31.819284 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:31 crc kubenswrapper[4802]: I1206 03:41:31.819309 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:31 crc kubenswrapper[4802]: I1206 03:41:31.819339 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:31 crc kubenswrapper[4802]: I1206 03:41:31.819364 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:31Z","lastTransitionTime":"2025-12-06T03:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:31 crc kubenswrapper[4802]: I1206 03:41:31.923066 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:31 crc kubenswrapper[4802]: I1206 03:41:31.923125 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:31 crc kubenswrapper[4802]: I1206 03:41:31.923142 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:31 crc kubenswrapper[4802]: I1206 03:41:31.923168 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:31 crc kubenswrapper[4802]: I1206 03:41:31.923186 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:31Z","lastTransitionTime":"2025-12-06T03:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:32 crc kubenswrapper[4802]: I1206 03:41:32.025491 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:32 crc kubenswrapper[4802]: I1206 03:41:32.025778 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:32 crc kubenswrapper[4802]: I1206 03:41:32.025944 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:32 crc kubenswrapper[4802]: I1206 03:41:32.026102 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:32 crc kubenswrapper[4802]: I1206 03:41:32.026244 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:32Z","lastTransitionTime":"2025-12-06T03:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:32 crc kubenswrapper[4802]: I1206 03:41:32.129845 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:32 crc kubenswrapper[4802]: I1206 03:41:32.129900 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:32 crc kubenswrapper[4802]: I1206 03:41:32.129912 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:32 crc kubenswrapper[4802]: I1206 03:41:32.129929 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:32 crc kubenswrapper[4802]: I1206 03:41:32.129942 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:32Z","lastTransitionTime":"2025-12-06T03:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:32 crc kubenswrapper[4802]: I1206 03:41:32.232696 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:32 crc kubenswrapper[4802]: I1206 03:41:32.232811 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:32 crc kubenswrapper[4802]: I1206 03:41:32.232832 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:32 crc kubenswrapper[4802]: I1206 03:41:32.232859 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:32 crc kubenswrapper[4802]: I1206 03:41:32.232877 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:32Z","lastTransitionTime":"2025-12-06T03:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:32 crc kubenswrapper[4802]: I1206 03:41:32.335718 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:32 crc kubenswrapper[4802]: I1206 03:41:32.335821 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:32 crc kubenswrapper[4802]: I1206 03:41:32.335841 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:32 crc kubenswrapper[4802]: I1206 03:41:32.335866 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:32 crc kubenswrapper[4802]: I1206 03:41:32.335884 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:32Z","lastTransitionTime":"2025-12-06T03:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:32 crc kubenswrapper[4802]: I1206 03:41:32.438998 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:32 crc kubenswrapper[4802]: I1206 03:41:32.439066 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:32 crc kubenswrapper[4802]: I1206 03:41:32.439082 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:32 crc kubenswrapper[4802]: I1206 03:41:32.439106 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:32 crc kubenswrapper[4802]: I1206 03:41:32.439123 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:32Z","lastTransitionTime":"2025-12-06T03:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:32 crc kubenswrapper[4802]: I1206 03:41:32.449278 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:41:32 crc kubenswrapper[4802]: I1206 03:41:32.449361 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:41:32 crc kubenswrapper[4802]: I1206 03:41:32.449398 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:41:32 crc kubenswrapper[4802]: E1206 03:41:32.449451 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5ltx2" podUID="8881599e-71e5-4578-b3f9-21f4fa2f57aa" Dec 06 03:41:32 crc kubenswrapper[4802]: E1206 03:41:32.449552 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:41:32 crc kubenswrapper[4802]: I1206 03:41:32.449569 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:41:32 crc kubenswrapper[4802]: E1206 03:41:32.449656 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:41:32 crc kubenswrapper[4802]: E1206 03:41:32.449739 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:41:32 crc kubenswrapper[4802]: I1206 03:41:32.541414 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:32 crc kubenswrapper[4802]: I1206 03:41:32.541452 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:32 crc kubenswrapper[4802]: I1206 03:41:32.541478 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:32 crc kubenswrapper[4802]: I1206 03:41:32.541495 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:32 crc kubenswrapper[4802]: I1206 03:41:32.541506 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:32Z","lastTransitionTime":"2025-12-06T03:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:32 crc kubenswrapper[4802]: I1206 03:41:32.644224 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:32 crc kubenswrapper[4802]: I1206 03:41:32.644261 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:32 crc kubenswrapper[4802]: I1206 03:41:32.644272 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:32 crc kubenswrapper[4802]: I1206 03:41:32.644288 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:32 crc kubenswrapper[4802]: I1206 03:41:32.644299 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:32Z","lastTransitionTime":"2025-12-06T03:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:32 crc kubenswrapper[4802]: I1206 03:41:32.746285 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:32 crc kubenswrapper[4802]: I1206 03:41:32.746365 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:32 crc kubenswrapper[4802]: I1206 03:41:32.746388 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:32 crc kubenswrapper[4802]: I1206 03:41:32.746912 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:32 crc kubenswrapper[4802]: I1206 03:41:32.747169 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:32Z","lastTransitionTime":"2025-12-06T03:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:32 crc kubenswrapper[4802]: I1206 03:41:32.850225 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:32 crc kubenswrapper[4802]: I1206 03:41:32.850275 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:32 crc kubenswrapper[4802]: I1206 03:41:32.850291 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:32 crc kubenswrapper[4802]: I1206 03:41:32.850316 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:32 crc kubenswrapper[4802]: I1206 03:41:32.850335 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:32Z","lastTransitionTime":"2025-12-06T03:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:32 crc kubenswrapper[4802]: I1206 03:41:32.953031 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:32 crc kubenswrapper[4802]: I1206 03:41:32.953088 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:32 crc kubenswrapper[4802]: I1206 03:41:32.953107 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:32 crc kubenswrapper[4802]: I1206 03:41:32.953135 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:32 crc kubenswrapper[4802]: I1206 03:41:32.953154 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:32Z","lastTransitionTime":"2025-12-06T03:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:33 crc kubenswrapper[4802]: I1206 03:41:33.056147 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:33 crc kubenswrapper[4802]: I1206 03:41:33.056194 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:33 crc kubenswrapper[4802]: I1206 03:41:33.056213 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:33 crc kubenswrapper[4802]: I1206 03:41:33.056236 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:33 crc kubenswrapper[4802]: I1206 03:41:33.056253 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:33Z","lastTransitionTime":"2025-12-06T03:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:33 crc kubenswrapper[4802]: I1206 03:41:33.159660 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:33 crc kubenswrapper[4802]: I1206 03:41:33.159827 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:33 crc kubenswrapper[4802]: I1206 03:41:33.159858 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:33 crc kubenswrapper[4802]: I1206 03:41:33.159890 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:33 crc kubenswrapper[4802]: I1206 03:41:33.159913 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:33Z","lastTransitionTime":"2025-12-06T03:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:33 crc kubenswrapper[4802]: I1206 03:41:33.262452 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:33 crc kubenswrapper[4802]: I1206 03:41:33.262499 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:33 crc kubenswrapper[4802]: I1206 03:41:33.262515 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:33 crc kubenswrapper[4802]: I1206 03:41:33.262539 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:33 crc kubenswrapper[4802]: I1206 03:41:33.262555 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:33Z","lastTransitionTime":"2025-12-06T03:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:33 crc kubenswrapper[4802]: I1206 03:41:33.365411 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:33 crc kubenswrapper[4802]: I1206 03:41:33.365467 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:33 crc kubenswrapper[4802]: I1206 03:41:33.365478 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:33 crc kubenswrapper[4802]: I1206 03:41:33.365497 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:33 crc kubenswrapper[4802]: I1206 03:41:33.365508 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:33Z","lastTransitionTime":"2025-12-06T03:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:33 crc kubenswrapper[4802]: I1206 03:41:33.468914 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:33 crc kubenswrapper[4802]: I1206 03:41:33.468989 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:33 crc kubenswrapper[4802]: I1206 03:41:33.469011 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:33 crc kubenswrapper[4802]: I1206 03:41:33.469035 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:33 crc kubenswrapper[4802]: I1206 03:41:33.469052 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:33Z","lastTransitionTime":"2025-12-06T03:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:33 crc kubenswrapper[4802]: I1206 03:41:33.571744 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:33 crc kubenswrapper[4802]: I1206 03:41:33.571847 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:33 crc kubenswrapper[4802]: I1206 03:41:33.571864 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:33 crc kubenswrapper[4802]: I1206 03:41:33.571887 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:33 crc kubenswrapper[4802]: I1206 03:41:33.571905 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:33Z","lastTransitionTime":"2025-12-06T03:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:33 crc kubenswrapper[4802]: I1206 03:41:33.675353 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:33 crc kubenswrapper[4802]: I1206 03:41:33.675430 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:33 crc kubenswrapper[4802]: I1206 03:41:33.675453 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:33 crc kubenswrapper[4802]: I1206 03:41:33.675481 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:33 crc kubenswrapper[4802]: I1206 03:41:33.675502 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:33Z","lastTransitionTime":"2025-12-06T03:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:33 crc kubenswrapper[4802]: I1206 03:41:33.778846 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:33 crc kubenswrapper[4802]: I1206 03:41:33.778932 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:33 crc kubenswrapper[4802]: I1206 03:41:33.778959 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:33 crc kubenswrapper[4802]: I1206 03:41:33.778991 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:33 crc kubenswrapper[4802]: I1206 03:41:33.779015 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:33Z","lastTransitionTime":"2025-12-06T03:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:33 crc kubenswrapper[4802]: I1206 03:41:33.882740 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:33 crc kubenswrapper[4802]: I1206 03:41:33.882842 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:33 crc kubenswrapper[4802]: I1206 03:41:33.882865 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:33 crc kubenswrapper[4802]: I1206 03:41:33.882897 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:33 crc kubenswrapper[4802]: I1206 03:41:33.882918 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:33Z","lastTransitionTime":"2025-12-06T03:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:33 crc kubenswrapper[4802]: I1206 03:41:33.985301 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:33 crc kubenswrapper[4802]: I1206 03:41:33.985356 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:33 crc kubenswrapper[4802]: I1206 03:41:33.985365 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:33 crc kubenswrapper[4802]: I1206 03:41:33.985383 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:33 crc kubenswrapper[4802]: I1206 03:41:33.985392 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:33Z","lastTransitionTime":"2025-12-06T03:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:34 crc kubenswrapper[4802]: I1206 03:41:34.088509 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:34 crc kubenswrapper[4802]: I1206 03:41:34.088545 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:34 crc kubenswrapper[4802]: I1206 03:41:34.088557 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:34 crc kubenswrapper[4802]: I1206 03:41:34.088574 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:34 crc kubenswrapper[4802]: I1206 03:41:34.088587 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:34Z","lastTransitionTime":"2025-12-06T03:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:34 crc kubenswrapper[4802]: I1206 03:41:34.191445 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:34 crc kubenswrapper[4802]: I1206 03:41:34.191504 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:34 crc kubenswrapper[4802]: I1206 03:41:34.191519 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:34 crc kubenswrapper[4802]: I1206 03:41:34.191548 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:34 crc kubenswrapper[4802]: I1206 03:41:34.191565 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:34Z","lastTransitionTime":"2025-12-06T03:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:34 crc kubenswrapper[4802]: I1206 03:41:34.294393 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:34 crc kubenswrapper[4802]: I1206 03:41:34.294428 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:34 crc kubenswrapper[4802]: I1206 03:41:34.294438 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:34 crc kubenswrapper[4802]: I1206 03:41:34.294451 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:34 crc kubenswrapper[4802]: I1206 03:41:34.294465 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:34Z","lastTransitionTime":"2025-12-06T03:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:34 crc kubenswrapper[4802]: I1206 03:41:34.396070 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:34 crc kubenswrapper[4802]: I1206 03:41:34.396103 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:34 crc kubenswrapper[4802]: I1206 03:41:34.396111 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:34 crc kubenswrapper[4802]: I1206 03:41:34.396124 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:34 crc kubenswrapper[4802]: I1206 03:41:34.396133 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:34Z","lastTransitionTime":"2025-12-06T03:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:34 crc kubenswrapper[4802]: I1206 03:41:34.449159 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:41:34 crc kubenswrapper[4802]: I1206 03:41:34.449202 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:41:34 crc kubenswrapper[4802]: I1206 03:41:34.449244 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:41:34 crc kubenswrapper[4802]: I1206 03:41:34.449168 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:41:34 crc kubenswrapper[4802]: E1206 03:41:34.449339 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5ltx2" podUID="8881599e-71e5-4578-b3f9-21f4fa2f57aa" Dec 06 03:41:34 crc kubenswrapper[4802]: E1206 03:41:34.449407 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:41:34 crc kubenswrapper[4802]: E1206 03:41:34.449642 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:41:34 crc kubenswrapper[4802]: E1206 03:41:34.449706 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:41:34 crc kubenswrapper[4802]: I1206 03:41:34.498380 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:34 crc kubenswrapper[4802]: I1206 03:41:34.498477 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:34 crc kubenswrapper[4802]: I1206 03:41:34.498493 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:34 crc kubenswrapper[4802]: I1206 03:41:34.498518 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:34 crc kubenswrapper[4802]: I1206 03:41:34.498545 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:34Z","lastTransitionTime":"2025-12-06T03:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:34 crc kubenswrapper[4802]: I1206 03:41:34.601781 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:34 crc kubenswrapper[4802]: I1206 03:41:34.601842 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:34 crc kubenswrapper[4802]: I1206 03:41:34.601865 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:34 crc kubenswrapper[4802]: I1206 03:41:34.601892 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:34 crc kubenswrapper[4802]: I1206 03:41:34.601912 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:34Z","lastTransitionTime":"2025-12-06T03:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:34 crc kubenswrapper[4802]: I1206 03:41:34.705057 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:34 crc kubenswrapper[4802]: I1206 03:41:34.705138 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:34 crc kubenswrapper[4802]: I1206 03:41:34.705156 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:34 crc kubenswrapper[4802]: I1206 03:41:34.705184 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:34 crc kubenswrapper[4802]: I1206 03:41:34.705208 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:34Z","lastTransitionTime":"2025-12-06T03:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:34 crc kubenswrapper[4802]: I1206 03:41:34.807676 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:34 crc kubenswrapper[4802]: I1206 03:41:34.807737 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:34 crc kubenswrapper[4802]: I1206 03:41:34.807785 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:34 crc kubenswrapper[4802]: I1206 03:41:34.807811 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:34 crc kubenswrapper[4802]: I1206 03:41:34.807828 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:34Z","lastTransitionTime":"2025-12-06T03:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:34 crc kubenswrapper[4802]: I1206 03:41:34.910987 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:34 crc kubenswrapper[4802]: I1206 03:41:34.911021 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:34 crc kubenswrapper[4802]: I1206 03:41:34.911029 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:34 crc kubenswrapper[4802]: I1206 03:41:34.911042 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:34 crc kubenswrapper[4802]: I1206 03:41:34.911052 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:34Z","lastTransitionTime":"2025-12-06T03:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:35 crc kubenswrapper[4802]: I1206 03:41:35.014079 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:35 crc kubenswrapper[4802]: I1206 03:41:35.014114 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:35 crc kubenswrapper[4802]: I1206 03:41:35.014131 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:35 crc kubenswrapper[4802]: I1206 03:41:35.014147 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:35 crc kubenswrapper[4802]: I1206 03:41:35.014156 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:35Z","lastTransitionTime":"2025-12-06T03:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:35 crc kubenswrapper[4802]: I1206 03:41:35.117192 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:35 crc kubenswrapper[4802]: I1206 03:41:35.117279 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:35 crc kubenswrapper[4802]: I1206 03:41:35.117294 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:35 crc kubenswrapper[4802]: I1206 03:41:35.117350 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:35 crc kubenswrapper[4802]: I1206 03:41:35.117361 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:35Z","lastTransitionTime":"2025-12-06T03:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:35 crc kubenswrapper[4802]: I1206 03:41:35.220672 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:35 crc kubenswrapper[4802]: I1206 03:41:35.220733 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:35 crc kubenswrapper[4802]: I1206 03:41:35.220806 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:35 crc kubenswrapper[4802]: I1206 03:41:35.220836 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:35 crc kubenswrapper[4802]: I1206 03:41:35.220860 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:35Z","lastTransitionTime":"2025-12-06T03:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:35 crc kubenswrapper[4802]: I1206 03:41:35.324259 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:35 crc kubenswrapper[4802]: I1206 03:41:35.324334 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:35 crc kubenswrapper[4802]: I1206 03:41:35.324358 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:35 crc kubenswrapper[4802]: I1206 03:41:35.324389 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:35 crc kubenswrapper[4802]: I1206 03:41:35.324414 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:35Z","lastTransitionTime":"2025-12-06T03:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:35 crc kubenswrapper[4802]: I1206 03:41:35.427706 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:35 crc kubenswrapper[4802]: I1206 03:41:35.427801 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:35 crc kubenswrapper[4802]: I1206 03:41:35.427827 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:35 crc kubenswrapper[4802]: I1206 03:41:35.427857 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:35 crc kubenswrapper[4802]: I1206 03:41:35.427881 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:35Z","lastTransitionTime":"2025-12-06T03:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:35 crc kubenswrapper[4802]: I1206 03:41:35.531288 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:35 crc kubenswrapper[4802]: I1206 03:41:35.531346 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:35 crc kubenswrapper[4802]: I1206 03:41:35.531364 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:35 crc kubenswrapper[4802]: I1206 03:41:35.531389 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:35 crc kubenswrapper[4802]: I1206 03:41:35.531408 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:35Z","lastTransitionTime":"2025-12-06T03:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:35 crc kubenswrapper[4802]: I1206 03:41:35.634423 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:35 crc kubenswrapper[4802]: I1206 03:41:35.634553 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:35 crc kubenswrapper[4802]: I1206 03:41:35.634572 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:35 crc kubenswrapper[4802]: I1206 03:41:35.634597 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:35 crc kubenswrapper[4802]: I1206 03:41:35.634614 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:35Z","lastTransitionTime":"2025-12-06T03:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:35 crc kubenswrapper[4802]: I1206 03:41:35.738015 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:35 crc kubenswrapper[4802]: I1206 03:41:35.738082 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:35 crc kubenswrapper[4802]: I1206 03:41:35.738099 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:35 crc kubenswrapper[4802]: I1206 03:41:35.738125 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:35 crc kubenswrapper[4802]: I1206 03:41:35.738146 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:35Z","lastTransitionTime":"2025-12-06T03:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:35 crc kubenswrapper[4802]: I1206 03:41:35.840363 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:35 crc kubenswrapper[4802]: I1206 03:41:35.840440 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:35 crc kubenswrapper[4802]: I1206 03:41:35.840464 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:35 crc kubenswrapper[4802]: I1206 03:41:35.840497 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:35 crc kubenswrapper[4802]: I1206 03:41:35.840523 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:35Z","lastTransitionTime":"2025-12-06T03:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:35 crc kubenswrapper[4802]: I1206 03:41:35.943516 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:35 crc kubenswrapper[4802]: I1206 03:41:35.943576 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:35 crc kubenswrapper[4802]: I1206 03:41:35.943598 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:35 crc kubenswrapper[4802]: I1206 03:41:35.943629 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:35 crc kubenswrapper[4802]: I1206 03:41:35.943651 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:35Z","lastTransitionTime":"2025-12-06T03:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:36 crc kubenswrapper[4802]: I1206 03:41:36.046515 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:36 crc kubenswrapper[4802]: I1206 03:41:36.046622 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:36 crc kubenswrapper[4802]: I1206 03:41:36.046642 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:36 crc kubenswrapper[4802]: I1206 03:41:36.046703 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:36 crc kubenswrapper[4802]: I1206 03:41:36.046720 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:36Z","lastTransitionTime":"2025-12-06T03:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:36 crc kubenswrapper[4802]: I1206 03:41:36.150374 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:36 crc kubenswrapper[4802]: I1206 03:41:36.150441 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:36 crc kubenswrapper[4802]: I1206 03:41:36.150481 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:36 crc kubenswrapper[4802]: I1206 03:41:36.150514 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:36 crc kubenswrapper[4802]: I1206 03:41:36.150537 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:36Z","lastTransitionTime":"2025-12-06T03:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:36 crc kubenswrapper[4802]: I1206 03:41:36.254018 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:36 crc kubenswrapper[4802]: I1206 03:41:36.254083 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:36 crc kubenswrapper[4802]: I1206 03:41:36.254106 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:36 crc kubenswrapper[4802]: I1206 03:41:36.254133 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:36 crc kubenswrapper[4802]: I1206 03:41:36.254150 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:36Z","lastTransitionTime":"2025-12-06T03:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:36 crc kubenswrapper[4802]: I1206 03:41:36.356791 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:36 crc kubenswrapper[4802]: I1206 03:41:36.356871 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:36 crc kubenswrapper[4802]: I1206 03:41:36.356886 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:36 crc kubenswrapper[4802]: I1206 03:41:36.356902 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:36 crc kubenswrapper[4802]: I1206 03:41:36.356914 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:36Z","lastTransitionTime":"2025-12-06T03:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:36 crc kubenswrapper[4802]: I1206 03:41:36.449035 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:41:36 crc kubenswrapper[4802]: I1206 03:41:36.449105 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:41:36 crc kubenswrapper[4802]: I1206 03:41:36.449056 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:41:36 crc kubenswrapper[4802]: I1206 03:41:36.449160 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:41:36 crc kubenswrapper[4802]: E1206 03:41:36.449290 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5ltx2" podUID="8881599e-71e5-4578-b3f9-21f4fa2f57aa" Dec 06 03:41:36 crc kubenswrapper[4802]: E1206 03:41:36.449843 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:41:36 crc kubenswrapper[4802]: E1206 03:41:36.449947 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:41:36 crc kubenswrapper[4802]: E1206 03:41:36.450054 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:41:36 crc kubenswrapper[4802]: I1206 03:41:36.459956 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:36 crc kubenswrapper[4802]: I1206 03:41:36.459979 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:36 crc kubenswrapper[4802]: I1206 03:41:36.459987 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:36 crc kubenswrapper[4802]: I1206 03:41:36.459999 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:36 crc kubenswrapper[4802]: I1206 03:41:36.460007 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:36Z","lastTransitionTime":"2025-12-06T03:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:36 crc kubenswrapper[4802]: I1206 03:41:36.563043 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:36 crc kubenswrapper[4802]: I1206 03:41:36.563100 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:36 crc kubenswrapper[4802]: I1206 03:41:36.563115 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:36 crc kubenswrapper[4802]: I1206 03:41:36.563133 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:36 crc kubenswrapper[4802]: I1206 03:41:36.563145 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:36Z","lastTransitionTime":"2025-12-06T03:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:36 crc kubenswrapper[4802]: I1206 03:41:36.666789 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:36 crc kubenswrapper[4802]: I1206 03:41:36.666849 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:36 crc kubenswrapper[4802]: I1206 03:41:36.666866 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:36 crc kubenswrapper[4802]: I1206 03:41:36.666892 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:36 crc kubenswrapper[4802]: I1206 03:41:36.666908 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:36Z","lastTransitionTime":"2025-12-06T03:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:36 crc kubenswrapper[4802]: I1206 03:41:36.770080 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:36 crc kubenswrapper[4802]: I1206 03:41:36.770140 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:36 crc kubenswrapper[4802]: I1206 03:41:36.770160 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:36 crc kubenswrapper[4802]: I1206 03:41:36.770222 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:36 crc kubenswrapper[4802]: I1206 03:41:36.770240 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:36Z","lastTransitionTime":"2025-12-06T03:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:36 crc kubenswrapper[4802]: I1206 03:41:36.873095 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:36 crc kubenswrapper[4802]: I1206 03:41:36.873157 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:36 crc kubenswrapper[4802]: I1206 03:41:36.873174 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:36 crc kubenswrapper[4802]: I1206 03:41:36.873198 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:36 crc kubenswrapper[4802]: I1206 03:41:36.873215 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:36Z","lastTransitionTime":"2025-12-06T03:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:36 crc kubenswrapper[4802]: I1206 03:41:36.976227 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:36 crc kubenswrapper[4802]: I1206 03:41:36.976280 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:36 crc kubenswrapper[4802]: I1206 03:41:36.976299 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:36 crc kubenswrapper[4802]: I1206 03:41:36.976323 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:36 crc kubenswrapper[4802]: I1206 03:41:36.976343 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:36Z","lastTransitionTime":"2025-12-06T03:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.078694 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.078743 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.078774 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.078791 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.078802 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:37Z","lastTransitionTime":"2025-12-06T03:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.181012 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.181047 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.181057 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.181069 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.181079 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:37Z","lastTransitionTime":"2025-12-06T03:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.284231 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.284299 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.284323 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.284354 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.284377 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:37Z","lastTransitionTime":"2025-12-06T03:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.386993 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.387035 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.387044 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.387058 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.387067 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:37Z","lastTransitionTime":"2025-12-06T03:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.468914 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kj7hf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd46dfa5-7d4a-49ed-8199-523f10cebdf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5eee8a0cf96ceb3785dac9df31bf8d31715d1c88fbea17c5e582b0a4554bb63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmkcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kj7hf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.482257 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2466610962332c6ca120fe055e4e8a3fc623088119dc1a412e06b745176f89e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.489632 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.489704 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.489717 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.489736 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.489764 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:37Z","lastTransitionTime":"2025-12-06T03:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.495819 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.509325 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6df38316-e0d3-4018-8d27-3620eba3a68d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57381e9589a2dc04a4d2733c37985f42594182128a374b7be0f471ea69138b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bf5e41701c7a2d45adc4740ceb4ac1785db17bc7cb15daed439f7556812d1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpxxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.525085 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fad9b63d84f41fccf696cd313d7ca7f72b3d3251d6c1b7341573a3fadf1e378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.541098 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb3f278fff47c900c56d8a563d5ed75a2ae84e2034b0f45aad4083790b9e7946\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b10ca16ac26904ec5e6245344b71d836888cae9b069cae6a7a4831d609752db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.557643 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9de75acd-1ba9-4322-ae49-791724b4b6a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://990943a2a48d79458297f4e5cf4178bbd6e9eafbcadc154a803846c589474e85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18a08a15442f04e37a5aa71afbcc293e80b876477d60cbe71d562c75d141a9ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18a08a15442f04e37a5aa71afbcc293e80b876477d60cbe71d562c75d141a9ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.572008 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2093b6-c4b5-4db8-be16-fbc859885e8d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0d6e582161f0d9126126e03b27211be22f6b5f797154928c92fa13a8118f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03004a29aefe7fe2452fb9f6663c0807afccb055d7ddbf6b907f305e6a19c81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1801ac64db4ba91c2e8a249e3651d3e50ab62e9f2c87312ce9bf7f38518a82cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6e454507cadbb982a566a3692995aaa3c733ff1cd7fb4870476602190e1eeba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.586292 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.592400 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.592437 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.592448 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.592465 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.592477 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:37Z","lastTransitionTime":"2025-12-06T03:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.604280 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.621235 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtw2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3a3bac2-3923-4507-a4de-1556b3ad473e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fed5ee4189d54456a3216f4fdb1ecf003fa47e29dbfe4b2b2f301edfb897d6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtw2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.651702 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7730de4f033561ad77046e028cd1a3f0bfe2775387cbdfa35f47ea79909ec4cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a158aebd701ef46225d3e37109ae28ab58d107087bf06650c4b6d0055a3bbba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2357f37497fde64937eca7b38693893af745ab278653ebc79406e122a6d4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8941bd25107333b264d3f081a811fdfeea23ad92193fc61367665fd2cd1209c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f9cd4dc75e63f09f60b48b6da284643910bb32ba51f2a808cb4d9803f30042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6796f65ee9c5bd0c18f205697135d17677c5daf3ba378cc992cba795e0188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf554aaba0e773f6a6c1f29bb1819b7ae25b2b2db255c4bd0a16c512f314e9cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf554aaba0e773f6a6c1f29bb1819b7ae25b2b2db255c4bd0a16c512f314e9cf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:41:23Z\\\",\\\"message\\\":\\\"42b7e8 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 03:41:22.975067 6823 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1206 03:41:22.975567 6823 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 03:41:22.975037 6823 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-kj7hf in node crc\\\\nI1206 03:41:22.975614 6823 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-kj7hf after 0 failed attempt(s)\\\\nI1206 03:41:22.975624 6823 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-kj7hf\\\\nF1206 03:41:22.974932 6823 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:41:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pvm6q_openshift-ovn-kubernetes(eafce4e0-e7fb-4877-b0ab-3283829ba1ef)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3cb210bfc97a439a7ff98c02c59f8ad16bc400dc214a616a44b43a8b01e2782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pvm6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.668398 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-b2knv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a3ae46f-03b5-4999-9d58-c38ee222f68a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76665d4c127445009fd66078827b54607c9ef812c8011d1c0b4cde1c0516853a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk4hj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-b2knv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.688634 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22rrq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03112a5d-5882-4fac-9ae5-13525ff82fe0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://831dab65557627cbce0d055459052dae5b594f33b66858367ccd034ff6811924\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0205b089b12c42688203bb7ec0f27dc5f9a577dabdb446b811f184046f85dc3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:41:14Z\\\",\\\"message\\\":\\\"2025-12-06T03:40:28+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_384cd081-7034-46cf-bb48-841728494c60\\\\n2025-12-06T03:40:28+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_384cd081-7034-46cf-bb48-841728494c60 to /host/opt/cni/bin/\\\\n2025-12-06T03:40:29Z [verbose] multus-daemon started\\\\n2025-12-06T03:40:29Z [verbose] Readiness Indicator file check\\\\n2025-12-06T03:41:14Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:41:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trlt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22rrq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.695069 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.695157 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.695259 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.695291 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.695308 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:37Z","lastTransitionTime":"2025-12-06T03:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.705718 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ngspk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34808ab2-d21b-40b6-8fe3-e18bf9e5fc46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52509d21f5b4bfbd458162cba176ff82c627c7c4671204addf9d76f7392fd760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qkf8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9685a4d5d4f40ff4f59b6da6193785f340890453d1e422d3421a621dc7bd02f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qkf8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ngspk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.718940 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5ltx2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8881599e-71e5-4578-b3f9-21f4fa2f57aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbm6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbm6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5ltx2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.733421 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1721e985-40bd-4b2d-a0e8-d4e365094ba6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a895e0cc04880c3287ba725ac5484835235f1e066a94ce98fcf4b6133c5cbd2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a21a97385f856ec996105b597d0ae81c56ed28b625ef3f2905e1af6ec8f6a45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://646e01a58908ba2e703aae5eef53e1b981b0f0c3d1d3202487f2a8b94ee3f55d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31752b84d86bda30792b29e4c6ed963753cc576c5de784395e8e6dcbebaced47\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"ver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:40:26.813730 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814130 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:40:26.814139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:40:26.814143 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:40:26.814145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 03:40:26.821403 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821442 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821482 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821496 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 03:40:26.821509 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821514 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 03:40:26.821612 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 03:40:26.821619 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1206 03:40:26.822049 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16a92d501c9e626fe91e4a6793e426d10eb292cfbab766db8b8cc72e5dae94ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.746270 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd230730-a9a1-47ed-b625-25e32231f271\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce321001135d0a2b878316f8003bf1bc54cba9a36f6a654ba376e1e7c0b4da71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af83814e6b2c5ed803157308ab895551e0826fd02f1ecc8b2751145c09d359da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9e6b48b5701dd330ef91712cdcd160afe9831a3881038bfb1e0397a9c8a3aa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2d359853c738e3e0e264f1a6a919b17afc71a362821b3ea170cda6695e3593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2d359853c738e3e0e264f1a6a919b17afc71a362821b3ea170cda6695e3593f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.798923 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.798993 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.799012 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.799042 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.799060 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:37Z","lastTransitionTime":"2025-12-06T03:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.902025 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.902137 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.902155 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.902179 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:37 crc kubenswrapper[4802]: I1206 03:41:37.902199 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:37Z","lastTransitionTime":"2025-12-06T03:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:38 crc kubenswrapper[4802]: I1206 03:41:38.005541 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:38 crc kubenswrapper[4802]: I1206 03:41:38.005703 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:38 crc kubenswrapper[4802]: I1206 03:41:38.005727 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:38 crc kubenswrapper[4802]: I1206 03:41:38.005853 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:38 crc kubenswrapper[4802]: I1206 03:41:38.005944 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:38Z","lastTransitionTime":"2025-12-06T03:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:38 crc kubenswrapper[4802]: I1206 03:41:38.108500 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:38 crc kubenswrapper[4802]: I1206 03:41:38.108548 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:38 crc kubenswrapper[4802]: I1206 03:41:38.108560 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:38 crc kubenswrapper[4802]: I1206 03:41:38.108576 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:38 crc kubenswrapper[4802]: I1206 03:41:38.108590 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:38Z","lastTransitionTime":"2025-12-06T03:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:38 crc kubenswrapper[4802]: I1206 03:41:38.210802 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:38 crc kubenswrapper[4802]: I1206 03:41:38.210864 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:38 crc kubenswrapper[4802]: I1206 03:41:38.210882 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:38 crc kubenswrapper[4802]: I1206 03:41:38.210908 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:38 crc kubenswrapper[4802]: I1206 03:41:38.210926 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:38Z","lastTransitionTime":"2025-12-06T03:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:38 crc kubenswrapper[4802]: I1206 03:41:38.313829 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:38 crc kubenswrapper[4802]: I1206 03:41:38.313892 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:38 crc kubenswrapper[4802]: I1206 03:41:38.313912 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:38 crc kubenswrapper[4802]: I1206 03:41:38.313936 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:38 crc kubenswrapper[4802]: I1206 03:41:38.313954 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:38Z","lastTransitionTime":"2025-12-06T03:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:38 crc kubenswrapper[4802]: I1206 03:41:38.417647 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:38 crc kubenswrapper[4802]: I1206 03:41:38.417727 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:38 crc kubenswrapper[4802]: I1206 03:41:38.417782 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:38 crc kubenswrapper[4802]: I1206 03:41:38.417810 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:38 crc kubenswrapper[4802]: I1206 03:41:38.417828 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:38Z","lastTransitionTime":"2025-12-06T03:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:38 crc kubenswrapper[4802]: I1206 03:41:38.450195 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:41:38 crc kubenswrapper[4802]: I1206 03:41:38.450232 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:41:38 crc kubenswrapper[4802]: E1206 03:41:38.450382 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:41:38 crc kubenswrapper[4802]: I1206 03:41:38.450434 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:41:38 crc kubenswrapper[4802]: E1206 03:41:38.450555 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5ltx2" podUID="8881599e-71e5-4578-b3f9-21f4fa2f57aa" Dec 06 03:41:38 crc kubenswrapper[4802]: E1206 03:41:38.450705 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:41:38 crc kubenswrapper[4802]: I1206 03:41:38.450894 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:41:38 crc kubenswrapper[4802]: E1206 03:41:38.451030 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:41:38 crc kubenswrapper[4802]: I1206 03:41:38.520853 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:38 crc kubenswrapper[4802]: I1206 03:41:38.520930 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:38 crc kubenswrapper[4802]: I1206 03:41:38.520952 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:38 crc kubenswrapper[4802]: I1206 03:41:38.520982 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:38 crc kubenswrapper[4802]: I1206 03:41:38.521003 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:38Z","lastTransitionTime":"2025-12-06T03:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:38 crc kubenswrapper[4802]: I1206 03:41:38.624331 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:38 crc kubenswrapper[4802]: I1206 03:41:38.624385 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:38 crc kubenswrapper[4802]: I1206 03:41:38.624412 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:38 crc kubenswrapper[4802]: I1206 03:41:38.624436 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:38 crc kubenswrapper[4802]: I1206 03:41:38.624454 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:38Z","lastTransitionTime":"2025-12-06T03:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:38 crc kubenswrapper[4802]: I1206 03:41:38.732111 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:38 crc kubenswrapper[4802]: I1206 03:41:38.732160 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:38 crc kubenswrapper[4802]: I1206 03:41:38.732179 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:38 crc kubenswrapper[4802]: I1206 03:41:38.732202 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:38 crc kubenswrapper[4802]: I1206 03:41:38.732219 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:38Z","lastTransitionTime":"2025-12-06T03:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:38 crc kubenswrapper[4802]: I1206 03:41:38.835513 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:38 crc kubenswrapper[4802]: I1206 03:41:38.835969 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:38 crc kubenswrapper[4802]: I1206 03:41:38.836112 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:38 crc kubenswrapper[4802]: I1206 03:41:38.836271 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:38 crc kubenswrapper[4802]: I1206 03:41:38.836418 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:38Z","lastTransitionTime":"2025-12-06T03:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:38 crc kubenswrapper[4802]: I1206 03:41:38.939462 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:38 crc kubenswrapper[4802]: I1206 03:41:38.939542 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:38 crc kubenswrapper[4802]: I1206 03:41:38.939571 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:38 crc kubenswrapper[4802]: I1206 03:41:38.939606 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:38 crc kubenswrapper[4802]: I1206 03:41:38.939629 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:38Z","lastTransitionTime":"2025-12-06T03:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:39 crc kubenswrapper[4802]: I1206 03:41:39.042144 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:39 crc kubenswrapper[4802]: I1206 03:41:39.042225 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:39 crc kubenswrapper[4802]: I1206 03:41:39.042246 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:39 crc kubenswrapper[4802]: I1206 03:41:39.042279 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:39 crc kubenswrapper[4802]: I1206 03:41:39.042302 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:39Z","lastTransitionTime":"2025-12-06T03:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:39 crc kubenswrapper[4802]: I1206 03:41:39.145551 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:39 crc kubenswrapper[4802]: I1206 03:41:39.145620 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:39 crc kubenswrapper[4802]: I1206 03:41:39.145644 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:39 crc kubenswrapper[4802]: I1206 03:41:39.145673 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:39 crc kubenswrapper[4802]: I1206 03:41:39.145695 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:39Z","lastTransitionTime":"2025-12-06T03:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:39 crc kubenswrapper[4802]: I1206 03:41:39.247855 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:39 crc kubenswrapper[4802]: I1206 03:41:39.247899 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:39 crc kubenswrapper[4802]: I1206 03:41:39.247910 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:39 crc kubenswrapper[4802]: I1206 03:41:39.247927 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:39 crc kubenswrapper[4802]: I1206 03:41:39.247940 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:39Z","lastTransitionTime":"2025-12-06T03:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:39 crc kubenswrapper[4802]: I1206 03:41:39.351096 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:39 crc kubenswrapper[4802]: I1206 03:41:39.351140 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:39 crc kubenswrapper[4802]: I1206 03:41:39.351153 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:39 crc kubenswrapper[4802]: I1206 03:41:39.351172 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:39 crc kubenswrapper[4802]: I1206 03:41:39.351184 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:39Z","lastTransitionTime":"2025-12-06T03:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:39 crc kubenswrapper[4802]: I1206 03:41:39.454299 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:39 crc kubenswrapper[4802]: I1206 03:41:39.454371 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:39 crc kubenswrapper[4802]: I1206 03:41:39.454393 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:39 crc kubenswrapper[4802]: I1206 03:41:39.454421 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:39 crc kubenswrapper[4802]: I1206 03:41:39.454447 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:39Z","lastTransitionTime":"2025-12-06T03:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:39 crc kubenswrapper[4802]: I1206 03:41:39.557666 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:39 crc kubenswrapper[4802]: I1206 03:41:39.557747 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:39 crc kubenswrapper[4802]: I1206 03:41:39.557809 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:39 crc kubenswrapper[4802]: I1206 03:41:39.557842 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:39 crc kubenswrapper[4802]: I1206 03:41:39.557867 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:39Z","lastTransitionTime":"2025-12-06T03:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:39 crc kubenswrapper[4802]: I1206 03:41:39.660488 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:39 crc kubenswrapper[4802]: I1206 03:41:39.660544 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:39 crc kubenswrapper[4802]: I1206 03:41:39.660560 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:39 crc kubenswrapper[4802]: I1206 03:41:39.660584 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:39 crc kubenswrapper[4802]: I1206 03:41:39.660599 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:39Z","lastTransitionTime":"2025-12-06T03:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:39 crc kubenswrapper[4802]: I1206 03:41:39.764124 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:39 crc kubenswrapper[4802]: I1206 03:41:39.764527 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:39 crc kubenswrapper[4802]: I1206 03:41:39.764685 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:39 crc kubenswrapper[4802]: I1206 03:41:39.764933 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:39 crc kubenswrapper[4802]: I1206 03:41:39.765123 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:39Z","lastTransitionTime":"2025-12-06T03:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:39 crc kubenswrapper[4802]: I1206 03:41:39.867489 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:39 crc kubenswrapper[4802]: I1206 03:41:39.867556 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:39 crc kubenswrapper[4802]: I1206 03:41:39.867570 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:39 crc kubenswrapper[4802]: I1206 03:41:39.867596 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:39 crc kubenswrapper[4802]: I1206 03:41:39.867611 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:39Z","lastTransitionTime":"2025-12-06T03:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:39 crc kubenswrapper[4802]: I1206 03:41:39.970718 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:39 crc kubenswrapper[4802]: I1206 03:41:39.970818 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:39 crc kubenswrapper[4802]: I1206 03:41:39.970837 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:39 crc kubenswrapper[4802]: I1206 03:41:39.970867 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:39 crc kubenswrapper[4802]: I1206 03:41:39.970893 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:39Z","lastTransitionTime":"2025-12-06T03:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.073639 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.073694 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.073706 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.073725 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.073738 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:40Z","lastTransitionTime":"2025-12-06T03:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.177415 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.178179 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.178320 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.178476 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.178579 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:40Z","lastTransitionTime":"2025-12-06T03:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.282336 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.282403 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.282422 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.282450 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.282469 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:40Z","lastTransitionTime":"2025-12-06T03:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.385288 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.385343 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.385355 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.385374 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.385389 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:40Z","lastTransitionTime":"2025-12-06T03:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.450035 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.450127 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.450051 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.450132 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:41:40 crc kubenswrapper[4802]: E1206 03:41:40.450445 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5ltx2" podUID="8881599e-71e5-4578-b3f9-21f4fa2f57aa" Dec 06 03:41:40 crc kubenswrapper[4802]: E1206 03:41:40.450666 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:41:40 crc kubenswrapper[4802]: E1206 03:41:40.450842 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:41:40 crc kubenswrapper[4802]: E1206 03:41:40.451042 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.452147 4802 scope.go:117] "RemoveContainer" containerID="bf554aaba0e773f6a6c1f29bb1819b7ae25b2b2db255c4bd0a16c512f314e9cf" Dec 06 03:41:40 crc kubenswrapper[4802]: E1206 03:41:40.452405 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pvm6q_openshift-ovn-kubernetes(eafce4e0-e7fb-4877-b0ab-3283829ba1ef)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" podUID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.488353 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.488427 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.488450 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.488668 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.488727 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:40Z","lastTransitionTime":"2025-12-06T03:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.567474 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.567552 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.567563 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.567584 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.567597 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:40Z","lastTransitionTime":"2025-12-06T03:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:40 crc kubenswrapper[4802]: E1206 03:41:40.587998 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0e01ca3a-758f-44bc-b4a2-30c1bc04fda0\\\",\\\"systemUUID\\\":\\\"c7c815e4-d2a4-4244-b035-f988fc95e215\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:40Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.593687 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.593786 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.593818 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.593849 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.593868 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:40Z","lastTransitionTime":"2025-12-06T03:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:40 crc kubenswrapper[4802]: E1206 03:41:40.614024 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0e01ca3a-758f-44bc-b4a2-30c1bc04fda0\\\",\\\"systemUUID\\\":\\\"c7c815e4-d2a4-4244-b035-f988fc95e215\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:40Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.619858 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.619916 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.619934 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.619958 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.619975 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:40Z","lastTransitionTime":"2025-12-06T03:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:40 crc kubenswrapper[4802]: E1206 03:41:40.642608 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0e01ca3a-758f-44bc-b4a2-30c1bc04fda0\\\",\\\"systemUUID\\\":\\\"c7c815e4-d2a4-4244-b035-f988fc95e215\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:40Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.648351 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.648547 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.648619 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.648710 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.648839 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:40Z","lastTransitionTime":"2025-12-06T03:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:40 crc kubenswrapper[4802]: E1206 03:41:40.667281 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0e01ca3a-758f-44bc-b4a2-30c1bc04fda0\\\",\\\"systemUUID\\\":\\\"c7c815e4-d2a4-4244-b035-f988fc95e215\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:40Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.673462 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.673521 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.673534 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.673557 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.673571 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:40Z","lastTransitionTime":"2025-12-06T03:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:40 crc kubenswrapper[4802]: E1206 03:41:40.689335 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0e01ca3a-758f-44bc-b4a2-30c1bc04fda0\\\",\\\"systemUUID\\\":\\\"c7c815e4-d2a4-4244-b035-f988fc95e215\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:40Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:40 crc kubenswrapper[4802]: E1206 03:41:40.689522 4802 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.691798 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.691837 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.691848 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.691868 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.691881 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:40Z","lastTransitionTime":"2025-12-06T03:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.794953 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.795038 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.795063 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.795094 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.795184 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:40Z","lastTransitionTime":"2025-12-06T03:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.897973 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.898019 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.898031 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.898053 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:40 crc kubenswrapper[4802]: I1206 03:41:40.898066 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:40Z","lastTransitionTime":"2025-12-06T03:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:41 crc kubenswrapper[4802]: I1206 03:41:41.000713 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:41 crc kubenswrapper[4802]: I1206 03:41:41.000812 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:41 crc kubenswrapper[4802]: I1206 03:41:41.000837 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:41 crc kubenswrapper[4802]: I1206 03:41:41.000864 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:41 crc kubenswrapper[4802]: I1206 03:41:41.000888 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:41Z","lastTransitionTime":"2025-12-06T03:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:41 crc kubenswrapper[4802]: I1206 03:41:41.105049 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:41 crc kubenswrapper[4802]: I1206 03:41:41.105099 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:41 crc kubenswrapper[4802]: I1206 03:41:41.105117 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:41 crc kubenswrapper[4802]: I1206 03:41:41.105140 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:41 crc kubenswrapper[4802]: I1206 03:41:41.105157 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:41Z","lastTransitionTime":"2025-12-06T03:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:41 crc kubenswrapper[4802]: I1206 03:41:41.208913 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:41 crc kubenswrapper[4802]: I1206 03:41:41.209258 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:41 crc kubenswrapper[4802]: I1206 03:41:41.209400 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:41 crc kubenswrapper[4802]: I1206 03:41:41.209589 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:41 crc kubenswrapper[4802]: I1206 03:41:41.209816 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:41Z","lastTransitionTime":"2025-12-06T03:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:41 crc kubenswrapper[4802]: I1206 03:41:41.313531 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:41 crc kubenswrapper[4802]: I1206 03:41:41.313996 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:41 crc kubenswrapper[4802]: I1206 03:41:41.314180 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:41 crc kubenswrapper[4802]: I1206 03:41:41.314352 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:41 crc kubenswrapper[4802]: I1206 03:41:41.314483 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:41Z","lastTransitionTime":"2025-12-06T03:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:41 crc kubenswrapper[4802]: I1206 03:41:41.418141 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:41 crc kubenswrapper[4802]: I1206 03:41:41.418194 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:41 crc kubenswrapper[4802]: I1206 03:41:41.418210 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:41 crc kubenswrapper[4802]: I1206 03:41:41.418233 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:41 crc kubenswrapper[4802]: I1206 03:41:41.418251 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:41Z","lastTransitionTime":"2025-12-06T03:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:41 crc kubenswrapper[4802]: I1206 03:41:41.521043 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:41 crc kubenswrapper[4802]: I1206 03:41:41.521111 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:41 crc kubenswrapper[4802]: I1206 03:41:41.521130 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:41 crc kubenswrapper[4802]: I1206 03:41:41.521155 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:41 crc kubenswrapper[4802]: I1206 03:41:41.521172 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:41Z","lastTransitionTime":"2025-12-06T03:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:41 crc kubenswrapper[4802]: I1206 03:41:41.624867 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:41 crc kubenswrapper[4802]: I1206 03:41:41.624955 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:41 crc kubenswrapper[4802]: I1206 03:41:41.624974 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:41 crc kubenswrapper[4802]: I1206 03:41:41.625020 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:41 crc kubenswrapper[4802]: I1206 03:41:41.625039 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:41Z","lastTransitionTime":"2025-12-06T03:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:41 crc kubenswrapper[4802]: I1206 03:41:41.728346 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:41 crc kubenswrapper[4802]: I1206 03:41:41.729007 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:41 crc kubenswrapper[4802]: I1206 03:41:41.729054 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:41 crc kubenswrapper[4802]: I1206 03:41:41.729084 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:41 crc kubenswrapper[4802]: I1206 03:41:41.729102 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:41Z","lastTransitionTime":"2025-12-06T03:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:41 crc kubenswrapper[4802]: I1206 03:41:41.832365 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:41 crc kubenswrapper[4802]: I1206 03:41:41.832397 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:41 crc kubenswrapper[4802]: I1206 03:41:41.832407 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:41 crc kubenswrapper[4802]: I1206 03:41:41.832423 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:41 crc kubenswrapper[4802]: I1206 03:41:41.832434 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:41Z","lastTransitionTime":"2025-12-06T03:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:41 crc kubenswrapper[4802]: I1206 03:41:41.934776 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:41 crc kubenswrapper[4802]: I1206 03:41:41.934818 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:41 crc kubenswrapper[4802]: I1206 03:41:41.934829 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:41 crc kubenswrapper[4802]: I1206 03:41:41.934845 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:41 crc kubenswrapper[4802]: I1206 03:41:41.934857 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:41Z","lastTransitionTime":"2025-12-06T03:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:42 crc kubenswrapper[4802]: I1206 03:41:42.037582 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:42 crc kubenswrapper[4802]: I1206 03:41:42.037653 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:42 crc kubenswrapper[4802]: I1206 03:41:42.037665 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:42 crc kubenswrapper[4802]: I1206 03:41:42.037685 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:42 crc kubenswrapper[4802]: I1206 03:41:42.037725 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:42Z","lastTransitionTime":"2025-12-06T03:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:42 crc kubenswrapper[4802]: I1206 03:41:42.140693 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:42 crc kubenswrapper[4802]: I1206 03:41:42.140745 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:42 crc kubenswrapper[4802]: I1206 03:41:42.140791 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:42 crc kubenswrapper[4802]: I1206 03:41:42.140815 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:42 crc kubenswrapper[4802]: I1206 03:41:42.140833 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:42Z","lastTransitionTime":"2025-12-06T03:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:42 crc kubenswrapper[4802]: I1206 03:41:42.243075 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:42 crc kubenswrapper[4802]: I1206 03:41:42.243105 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:42 crc kubenswrapper[4802]: I1206 03:41:42.243113 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:42 crc kubenswrapper[4802]: I1206 03:41:42.243125 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:42 crc kubenswrapper[4802]: I1206 03:41:42.243134 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:42Z","lastTransitionTime":"2025-12-06T03:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:42 crc kubenswrapper[4802]: I1206 03:41:42.345901 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:42 crc kubenswrapper[4802]: I1206 03:41:42.345953 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:42 crc kubenswrapper[4802]: I1206 03:41:42.345973 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:42 crc kubenswrapper[4802]: I1206 03:41:42.345997 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:42 crc kubenswrapper[4802]: I1206 03:41:42.346016 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:42Z","lastTransitionTime":"2025-12-06T03:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:42 crc kubenswrapper[4802]: I1206 03:41:42.447872 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:42 crc kubenswrapper[4802]: I1206 03:41:42.448179 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:42 crc kubenswrapper[4802]: I1206 03:41:42.448414 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:42 crc kubenswrapper[4802]: I1206 03:41:42.448566 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:42 crc kubenswrapper[4802]: I1206 03:41:42.448692 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:42Z","lastTransitionTime":"2025-12-06T03:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:42 crc kubenswrapper[4802]: I1206 03:41:42.449065 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:41:42 crc kubenswrapper[4802]: I1206 03:41:42.449149 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:41:42 crc kubenswrapper[4802]: I1206 03:41:42.449079 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:41:42 crc kubenswrapper[4802]: I1206 03:41:42.449148 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:41:42 crc kubenswrapper[4802]: E1206 03:41:42.449246 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:41:42 crc kubenswrapper[4802]: E1206 03:41:42.449352 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:41:42 crc kubenswrapper[4802]: E1206 03:41:42.449535 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5ltx2" podUID="8881599e-71e5-4578-b3f9-21f4fa2f57aa" Dec 06 03:41:42 crc kubenswrapper[4802]: E1206 03:41:42.449704 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:41:42 crc kubenswrapper[4802]: I1206 03:41:42.550361 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:42 crc kubenswrapper[4802]: I1206 03:41:42.550411 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:42 crc kubenswrapper[4802]: I1206 03:41:42.550426 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:42 crc kubenswrapper[4802]: I1206 03:41:42.550449 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:42 crc kubenswrapper[4802]: I1206 03:41:42.550466 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:42Z","lastTransitionTime":"2025-12-06T03:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:42 crc kubenswrapper[4802]: I1206 03:41:42.653029 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:42 crc kubenswrapper[4802]: I1206 03:41:42.653068 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:42 crc kubenswrapper[4802]: I1206 03:41:42.653079 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:42 crc kubenswrapper[4802]: I1206 03:41:42.653097 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:42 crc kubenswrapper[4802]: I1206 03:41:42.653111 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:42Z","lastTransitionTime":"2025-12-06T03:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:42 crc kubenswrapper[4802]: I1206 03:41:42.755880 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:42 crc kubenswrapper[4802]: I1206 03:41:42.755947 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:42 crc kubenswrapper[4802]: I1206 03:41:42.755965 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:42 crc kubenswrapper[4802]: I1206 03:41:42.755992 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:42 crc kubenswrapper[4802]: I1206 03:41:42.756005 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:42Z","lastTransitionTime":"2025-12-06T03:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:42 crc kubenswrapper[4802]: I1206 03:41:42.859159 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:42 crc kubenswrapper[4802]: I1206 03:41:42.859223 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:42 crc kubenswrapper[4802]: I1206 03:41:42.859242 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:42 crc kubenswrapper[4802]: I1206 03:41:42.859269 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:42 crc kubenswrapper[4802]: I1206 03:41:42.859288 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:42Z","lastTransitionTime":"2025-12-06T03:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:42 crc kubenswrapper[4802]: I1206 03:41:42.962722 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:42 crc kubenswrapper[4802]: I1206 03:41:42.962861 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:42 crc kubenswrapper[4802]: I1206 03:41:42.962888 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:42 crc kubenswrapper[4802]: I1206 03:41:42.962921 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:42 crc kubenswrapper[4802]: I1206 03:41:42.962950 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:42Z","lastTransitionTime":"2025-12-06T03:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:43 crc kubenswrapper[4802]: I1206 03:41:43.066269 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:43 crc kubenswrapper[4802]: I1206 03:41:43.066334 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:43 crc kubenswrapper[4802]: I1206 03:41:43.066353 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:43 crc kubenswrapper[4802]: I1206 03:41:43.066377 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:43 crc kubenswrapper[4802]: I1206 03:41:43.066396 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:43Z","lastTransitionTime":"2025-12-06T03:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:43 crc kubenswrapper[4802]: I1206 03:41:43.169241 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:43 crc kubenswrapper[4802]: I1206 03:41:43.169309 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:43 crc kubenswrapper[4802]: I1206 03:41:43.169333 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:43 crc kubenswrapper[4802]: I1206 03:41:43.169356 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:43 crc kubenswrapper[4802]: I1206 03:41:43.169376 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:43Z","lastTransitionTime":"2025-12-06T03:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:43 crc kubenswrapper[4802]: I1206 03:41:43.272511 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:43 crc kubenswrapper[4802]: I1206 03:41:43.272575 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:43 crc kubenswrapper[4802]: I1206 03:41:43.272599 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:43 crc kubenswrapper[4802]: I1206 03:41:43.272629 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:43 crc kubenswrapper[4802]: I1206 03:41:43.272652 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:43Z","lastTransitionTime":"2025-12-06T03:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:43 crc kubenswrapper[4802]: I1206 03:41:43.376176 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:43 crc kubenswrapper[4802]: I1206 03:41:43.376244 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:43 crc kubenswrapper[4802]: I1206 03:41:43.376267 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:43 crc kubenswrapper[4802]: I1206 03:41:43.376304 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:43 crc kubenswrapper[4802]: I1206 03:41:43.376327 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:43Z","lastTransitionTime":"2025-12-06T03:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:43 crc kubenswrapper[4802]: I1206 03:41:43.479712 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:43 crc kubenswrapper[4802]: I1206 03:41:43.479825 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:43 crc kubenswrapper[4802]: I1206 03:41:43.479846 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:43 crc kubenswrapper[4802]: I1206 03:41:43.479873 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:43 crc kubenswrapper[4802]: I1206 03:41:43.479892 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:43Z","lastTransitionTime":"2025-12-06T03:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:43 crc kubenswrapper[4802]: I1206 03:41:43.582343 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:43 crc kubenswrapper[4802]: I1206 03:41:43.582397 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:43 crc kubenswrapper[4802]: I1206 03:41:43.582415 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:43 crc kubenswrapper[4802]: I1206 03:41:43.582439 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:43 crc kubenswrapper[4802]: I1206 03:41:43.582457 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:43Z","lastTransitionTime":"2025-12-06T03:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:43 crc kubenswrapper[4802]: I1206 03:41:43.685881 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:43 crc kubenswrapper[4802]: I1206 03:41:43.685963 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:43 crc kubenswrapper[4802]: I1206 03:41:43.685984 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:43 crc kubenswrapper[4802]: I1206 03:41:43.686010 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:43 crc kubenswrapper[4802]: I1206 03:41:43.686030 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:43Z","lastTransitionTime":"2025-12-06T03:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:43 crc kubenswrapper[4802]: I1206 03:41:43.789040 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:43 crc kubenswrapper[4802]: I1206 03:41:43.789117 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:43 crc kubenswrapper[4802]: I1206 03:41:43.789137 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:43 crc kubenswrapper[4802]: I1206 03:41:43.789163 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:43 crc kubenswrapper[4802]: I1206 03:41:43.789180 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:43Z","lastTransitionTime":"2025-12-06T03:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:43 crc kubenswrapper[4802]: I1206 03:41:43.892788 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:43 crc kubenswrapper[4802]: I1206 03:41:43.892858 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:43 crc kubenswrapper[4802]: I1206 03:41:43.892881 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:43 crc kubenswrapper[4802]: I1206 03:41:43.892910 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:43 crc kubenswrapper[4802]: I1206 03:41:43.892935 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:43Z","lastTransitionTime":"2025-12-06T03:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:43 crc kubenswrapper[4802]: I1206 03:41:43.995542 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:43 crc kubenswrapper[4802]: I1206 03:41:43.995603 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:43 crc kubenswrapper[4802]: I1206 03:41:43.995621 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:43 crc kubenswrapper[4802]: I1206 03:41:43.995645 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:43 crc kubenswrapper[4802]: I1206 03:41:43.995663 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:43Z","lastTransitionTime":"2025-12-06T03:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:44 crc kubenswrapper[4802]: I1206 03:41:44.098986 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:44 crc kubenswrapper[4802]: I1206 03:41:44.099060 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:44 crc kubenswrapper[4802]: I1206 03:41:44.099092 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:44 crc kubenswrapper[4802]: I1206 03:41:44.099123 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:44 crc kubenswrapper[4802]: I1206 03:41:44.099147 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:44Z","lastTransitionTime":"2025-12-06T03:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:44 crc kubenswrapper[4802]: I1206 03:41:44.202412 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:44 crc kubenswrapper[4802]: I1206 03:41:44.202522 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:44 crc kubenswrapper[4802]: I1206 03:41:44.202554 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:44 crc kubenswrapper[4802]: I1206 03:41:44.202583 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:44 crc kubenswrapper[4802]: I1206 03:41:44.202601 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:44Z","lastTransitionTime":"2025-12-06T03:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:44 crc kubenswrapper[4802]: I1206 03:41:44.305657 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:44 crc kubenswrapper[4802]: I1206 03:41:44.305722 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:44 crc kubenswrapper[4802]: I1206 03:41:44.305739 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:44 crc kubenswrapper[4802]: I1206 03:41:44.305789 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:44 crc kubenswrapper[4802]: I1206 03:41:44.305842 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:44Z","lastTransitionTime":"2025-12-06T03:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:44 crc kubenswrapper[4802]: I1206 03:41:44.408901 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:44 crc kubenswrapper[4802]: I1206 03:41:44.408957 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:44 crc kubenswrapper[4802]: I1206 03:41:44.408974 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:44 crc kubenswrapper[4802]: I1206 03:41:44.409001 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:44 crc kubenswrapper[4802]: I1206 03:41:44.409017 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:44Z","lastTransitionTime":"2025-12-06T03:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:44 crc kubenswrapper[4802]: I1206 03:41:44.449281 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:41:44 crc kubenswrapper[4802]: I1206 03:41:44.449523 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:41:44 crc kubenswrapper[4802]: I1206 03:41:44.449520 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:41:44 crc kubenswrapper[4802]: I1206 03:41:44.449524 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:41:44 crc kubenswrapper[4802]: E1206 03:41:44.449689 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:41:44 crc kubenswrapper[4802]: E1206 03:41:44.449886 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:41:44 crc kubenswrapper[4802]: E1206 03:41:44.450073 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5ltx2" podUID="8881599e-71e5-4578-b3f9-21f4fa2f57aa" Dec 06 03:41:44 crc kubenswrapper[4802]: E1206 03:41:44.450184 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:41:44 crc kubenswrapper[4802]: I1206 03:41:44.511606 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:44 crc kubenswrapper[4802]: I1206 03:41:44.511690 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:44 crc kubenswrapper[4802]: I1206 03:41:44.511724 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:44 crc kubenswrapper[4802]: I1206 03:41:44.511797 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:44 crc kubenswrapper[4802]: I1206 03:41:44.511824 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:44Z","lastTransitionTime":"2025-12-06T03:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:44 crc kubenswrapper[4802]: I1206 03:41:44.615287 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:44 crc kubenswrapper[4802]: I1206 03:41:44.615377 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:44 crc kubenswrapper[4802]: I1206 03:41:44.615402 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:44 crc kubenswrapper[4802]: I1206 03:41:44.615434 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:44 crc kubenswrapper[4802]: I1206 03:41:44.615459 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:44Z","lastTransitionTime":"2025-12-06T03:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:44 crc kubenswrapper[4802]: I1206 03:41:44.718924 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:44 crc kubenswrapper[4802]: I1206 03:41:44.718997 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:44 crc kubenswrapper[4802]: I1206 03:41:44.719025 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:44 crc kubenswrapper[4802]: I1206 03:41:44.719058 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:44 crc kubenswrapper[4802]: I1206 03:41:44.719077 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:44Z","lastTransitionTime":"2025-12-06T03:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:44 crc kubenswrapper[4802]: I1206 03:41:44.822414 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:44 crc kubenswrapper[4802]: I1206 03:41:44.822493 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:44 crc kubenswrapper[4802]: I1206 03:41:44.822517 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:44 crc kubenswrapper[4802]: I1206 03:41:44.822553 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:44 crc kubenswrapper[4802]: I1206 03:41:44.822577 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:44Z","lastTransitionTime":"2025-12-06T03:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:44 crc kubenswrapper[4802]: I1206 03:41:44.925169 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:44 crc kubenswrapper[4802]: I1206 03:41:44.925230 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:44 crc kubenswrapper[4802]: I1206 03:41:44.925248 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:44 crc kubenswrapper[4802]: I1206 03:41:44.925273 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:44 crc kubenswrapper[4802]: I1206 03:41:44.925291 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:44Z","lastTransitionTime":"2025-12-06T03:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:44 crc kubenswrapper[4802]: I1206 03:41:44.951696 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8881599e-71e5-4578-b3f9-21f4fa2f57aa-metrics-certs\") pod \"network-metrics-daemon-5ltx2\" (UID: \"8881599e-71e5-4578-b3f9-21f4fa2f57aa\") " pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:41:44 crc kubenswrapper[4802]: E1206 03:41:44.951866 4802 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 03:41:44 crc kubenswrapper[4802]: E1206 03:41:44.951925 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8881599e-71e5-4578-b3f9-21f4fa2f57aa-metrics-certs podName:8881599e-71e5-4578-b3f9-21f4fa2f57aa nodeName:}" failed. No retries permitted until 2025-12-06 03:42:48.95190682 +0000 UTC m=+161.823815972 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8881599e-71e5-4578-b3f9-21f4fa2f57aa-metrics-certs") pod "network-metrics-daemon-5ltx2" (UID: "8881599e-71e5-4578-b3f9-21f4fa2f57aa") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 03:41:45 crc kubenswrapper[4802]: I1206 03:41:45.027943 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:45 crc kubenswrapper[4802]: I1206 03:41:45.027993 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:45 crc kubenswrapper[4802]: I1206 03:41:45.028004 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:45 crc kubenswrapper[4802]: I1206 03:41:45.028021 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:45 crc kubenswrapper[4802]: I1206 03:41:45.028033 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:45Z","lastTransitionTime":"2025-12-06T03:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:45 crc kubenswrapper[4802]: I1206 03:41:45.130834 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:45 crc kubenswrapper[4802]: I1206 03:41:45.130950 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:45 crc kubenswrapper[4802]: I1206 03:41:45.130969 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:45 crc kubenswrapper[4802]: I1206 03:41:45.131038 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:45 crc kubenswrapper[4802]: I1206 03:41:45.131062 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:45Z","lastTransitionTime":"2025-12-06T03:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:45 crc kubenswrapper[4802]: I1206 03:41:45.234458 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:45 crc kubenswrapper[4802]: I1206 03:41:45.234518 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:45 crc kubenswrapper[4802]: I1206 03:41:45.234539 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:45 crc kubenswrapper[4802]: I1206 03:41:45.234564 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:45 crc kubenswrapper[4802]: I1206 03:41:45.234617 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:45Z","lastTransitionTime":"2025-12-06T03:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:45 crc kubenswrapper[4802]: I1206 03:41:45.337780 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:45 crc kubenswrapper[4802]: I1206 03:41:45.337856 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:45 crc kubenswrapper[4802]: I1206 03:41:45.337874 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:45 crc kubenswrapper[4802]: I1206 03:41:45.337900 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:45 crc kubenswrapper[4802]: I1206 03:41:45.337917 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:45Z","lastTransitionTime":"2025-12-06T03:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:45 crc kubenswrapper[4802]: I1206 03:41:45.441086 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:45 crc kubenswrapper[4802]: I1206 03:41:45.441157 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:45 crc kubenswrapper[4802]: I1206 03:41:45.441181 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:45 crc kubenswrapper[4802]: I1206 03:41:45.441213 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:45 crc kubenswrapper[4802]: I1206 03:41:45.441237 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:45Z","lastTransitionTime":"2025-12-06T03:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:45 crc kubenswrapper[4802]: I1206 03:41:45.544512 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:45 crc kubenswrapper[4802]: I1206 03:41:45.544580 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:45 crc kubenswrapper[4802]: I1206 03:41:45.544598 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:45 crc kubenswrapper[4802]: I1206 03:41:45.544622 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:45 crc kubenswrapper[4802]: I1206 03:41:45.544640 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:45Z","lastTransitionTime":"2025-12-06T03:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:45 crc kubenswrapper[4802]: I1206 03:41:45.648165 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:45 crc kubenswrapper[4802]: I1206 03:41:45.648226 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:45 crc kubenswrapper[4802]: I1206 03:41:45.648243 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:45 crc kubenswrapper[4802]: I1206 03:41:45.648266 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:45 crc kubenswrapper[4802]: I1206 03:41:45.648282 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:45Z","lastTransitionTime":"2025-12-06T03:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:45 crc kubenswrapper[4802]: I1206 03:41:45.751577 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:45 crc kubenswrapper[4802]: I1206 03:41:45.751640 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:45 crc kubenswrapper[4802]: I1206 03:41:45.751659 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:45 crc kubenswrapper[4802]: I1206 03:41:45.751689 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:45 crc kubenswrapper[4802]: I1206 03:41:45.751706 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:45Z","lastTransitionTime":"2025-12-06T03:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:45 crc kubenswrapper[4802]: I1206 03:41:45.854087 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:45 crc kubenswrapper[4802]: I1206 03:41:45.854145 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:45 crc kubenswrapper[4802]: I1206 03:41:45.854163 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:45 crc kubenswrapper[4802]: I1206 03:41:45.854185 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:45 crc kubenswrapper[4802]: I1206 03:41:45.854202 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:45Z","lastTransitionTime":"2025-12-06T03:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:45 crc kubenswrapper[4802]: I1206 03:41:45.956876 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:45 crc kubenswrapper[4802]: I1206 03:41:45.956939 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:45 crc kubenswrapper[4802]: I1206 03:41:45.956961 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:45 crc kubenswrapper[4802]: I1206 03:41:45.956990 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:45 crc kubenswrapper[4802]: I1206 03:41:45.957013 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:45Z","lastTransitionTime":"2025-12-06T03:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:46 crc kubenswrapper[4802]: I1206 03:41:46.059420 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:46 crc kubenswrapper[4802]: I1206 03:41:46.059504 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:46 crc kubenswrapper[4802]: I1206 03:41:46.059530 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:46 crc kubenswrapper[4802]: I1206 03:41:46.059563 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:46 crc kubenswrapper[4802]: I1206 03:41:46.059588 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:46Z","lastTransitionTime":"2025-12-06T03:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:46 crc kubenswrapper[4802]: I1206 03:41:46.162002 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:46 crc kubenswrapper[4802]: I1206 03:41:46.162054 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:46 crc kubenswrapper[4802]: I1206 03:41:46.162067 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:46 crc kubenswrapper[4802]: I1206 03:41:46.162086 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:46 crc kubenswrapper[4802]: I1206 03:41:46.162101 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:46Z","lastTransitionTime":"2025-12-06T03:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:46 crc kubenswrapper[4802]: I1206 03:41:46.264935 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:46 crc kubenswrapper[4802]: I1206 03:41:46.264994 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:46 crc kubenswrapper[4802]: I1206 03:41:46.265011 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:46 crc kubenswrapper[4802]: I1206 03:41:46.265036 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:46 crc kubenswrapper[4802]: I1206 03:41:46.265057 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:46Z","lastTransitionTime":"2025-12-06T03:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:46 crc kubenswrapper[4802]: I1206 03:41:46.367430 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:46 crc kubenswrapper[4802]: I1206 03:41:46.367487 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:46 crc kubenswrapper[4802]: I1206 03:41:46.367545 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:46 crc kubenswrapper[4802]: I1206 03:41:46.367573 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:46 crc kubenswrapper[4802]: I1206 03:41:46.367590 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:46Z","lastTransitionTime":"2025-12-06T03:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:46 crc kubenswrapper[4802]: I1206 03:41:46.449430 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:41:46 crc kubenswrapper[4802]: I1206 03:41:46.449508 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:41:46 crc kubenswrapper[4802]: I1206 03:41:46.449447 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:41:46 crc kubenswrapper[4802]: I1206 03:41:46.449458 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:41:46 crc kubenswrapper[4802]: E1206 03:41:46.449615 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5ltx2" podUID="8881599e-71e5-4578-b3f9-21f4fa2f57aa" Dec 06 03:41:46 crc kubenswrapper[4802]: E1206 03:41:46.449832 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:41:46 crc kubenswrapper[4802]: E1206 03:41:46.449978 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:41:46 crc kubenswrapper[4802]: E1206 03:41:46.450070 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:41:46 crc kubenswrapper[4802]: I1206 03:41:46.470460 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:46 crc kubenswrapper[4802]: I1206 03:41:46.470529 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:46 crc kubenswrapper[4802]: I1206 03:41:46.470547 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:46 crc kubenswrapper[4802]: I1206 03:41:46.470572 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:46 crc kubenswrapper[4802]: I1206 03:41:46.470589 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:46Z","lastTransitionTime":"2025-12-06T03:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:46 crc kubenswrapper[4802]: I1206 03:41:46.573048 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:46 crc kubenswrapper[4802]: I1206 03:41:46.573118 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:46 crc kubenswrapper[4802]: I1206 03:41:46.573136 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:46 crc kubenswrapper[4802]: I1206 03:41:46.573162 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:46 crc kubenswrapper[4802]: I1206 03:41:46.573180 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:46Z","lastTransitionTime":"2025-12-06T03:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:46 crc kubenswrapper[4802]: I1206 03:41:46.676527 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:46 crc kubenswrapper[4802]: I1206 03:41:46.676584 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:46 crc kubenswrapper[4802]: I1206 03:41:46.676601 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:46 crc kubenswrapper[4802]: I1206 03:41:46.676626 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:46 crc kubenswrapper[4802]: I1206 03:41:46.676644 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:46Z","lastTransitionTime":"2025-12-06T03:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:46 crc kubenswrapper[4802]: I1206 03:41:46.779965 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:46 crc kubenswrapper[4802]: I1206 03:41:46.780032 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:46 crc kubenswrapper[4802]: I1206 03:41:46.780050 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:46 crc kubenswrapper[4802]: I1206 03:41:46.780074 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:46 crc kubenswrapper[4802]: I1206 03:41:46.780092 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:46Z","lastTransitionTime":"2025-12-06T03:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:46 crc kubenswrapper[4802]: I1206 03:41:46.883476 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:46 crc kubenswrapper[4802]: I1206 03:41:46.883533 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:46 crc kubenswrapper[4802]: I1206 03:41:46.883556 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:46 crc kubenswrapper[4802]: I1206 03:41:46.883587 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:46 crc kubenswrapper[4802]: I1206 03:41:46.883609 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:46Z","lastTransitionTime":"2025-12-06T03:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:46 crc kubenswrapper[4802]: I1206 03:41:46.987010 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:46 crc kubenswrapper[4802]: I1206 03:41:46.987073 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:46 crc kubenswrapper[4802]: I1206 03:41:46.987090 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:46 crc kubenswrapper[4802]: I1206 03:41:46.987116 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:46 crc kubenswrapper[4802]: I1206 03:41:46.987133 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:46Z","lastTransitionTime":"2025-12-06T03:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.090163 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.090218 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.090236 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.090258 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.090275 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:47Z","lastTransitionTime":"2025-12-06T03:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.193704 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.193785 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.193805 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.193829 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.193845 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:47Z","lastTransitionTime":"2025-12-06T03:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.296816 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.296874 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.296893 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.296916 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.296933 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:47Z","lastTransitionTime":"2025-12-06T03:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.399353 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.399454 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.399473 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.399540 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.399560 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:47Z","lastTransitionTime":"2025-12-06T03:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.470258 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:47Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.490632 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtw2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3a3bac2-3923-4507-a4de-1556b3ad473e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fed5ee4189d54456a3216f4fdb1ecf003fa47e29dbfe4b2b2f301edfb897d6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtw2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:47Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.502127 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.502181 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.502198 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.502220 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.502239 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:47Z","lastTransitionTime":"2025-12-06T03:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.524501 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7730de4f033561ad77046e028cd1a3f0bfe2775387cbdfa35f47ea79909ec4cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a158aebd701ef46225d3e37109ae28ab58d107087bf06650c4b6d0055a3bbba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2357f37497fde64937eca7b38693893af745ab278653ebc79406e122a6d4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8941bd25107333b264d3f081a811fdfeea23ad92193fc61367665fd2cd1209c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f9cd4dc75e63f09f60b48b6da284643910bb32ba51f2a808cb4d9803f30042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6796f65ee9c5bd0c18f205697135d17677c5daf3ba378cc992cba795e0188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf554aaba0e773f6a6c1f29bb1819b7ae25b2b2db255c4bd0a16c512f314e9cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf554aaba0e773f6a6c1f29bb1819b7ae25b2b2db255c4bd0a16c512f314e9cf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:41:23Z\\\",\\\"message\\\":\\\"42b7e8 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 03:41:22.975067 6823 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1206 03:41:22.975567 6823 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 03:41:22.975037 6823 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-kj7hf in node crc\\\\nI1206 03:41:22.975614 6823 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-kj7hf after 0 failed attempt(s)\\\\nI1206 03:41:22.975624 6823 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-kj7hf\\\\nF1206 03:41:22.974932 6823 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:41:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pvm6q_openshift-ovn-kubernetes(eafce4e0-e7fb-4877-b0ab-3283829ba1ef)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3cb210bfc97a439a7ff98c02c59f8ad16bc400dc214a616a44b43a8b01e2782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pvm6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:47Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.547164 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22rrq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03112a5d-5882-4fac-9ae5-13525ff82fe0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://831dab65557627cbce0d055459052dae5b594f33b66858367ccd034ff6811924\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0205b089b12c42688203bb7ec0f27dc5f9a577dabdb446b811f184046f85dc3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:41:14Z\\\",\\\"message\\\":\\\"2025-12-06T03:40:28+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_384cd081-7034-46cf-bb48-841728494c60\\\\n2025-12-06T03:40:28+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_384cd081-7034-46cf-bb48-841728494c60 to /host/opt/cni/bin/\\\\n2025-12-06T03:40:29Z [verbose] multus-daemon started\\\\n2025-12-06T03:40:29Z [verbose] Readiness Indicator file check\\\\n2025-12-06T03:41:14Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:41:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trlt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22rrq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:47Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.564485 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ngspk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34808ab2-d21b-40b6-8fe3-e18bf9e5fc46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52509d21f5b4bfbd458162cba176ff82c627c7c4671204addf9d76f7392fd760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qkf8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9685a4d5d4f40ff4f59b6da6193785f340890453d1e422d3421a621dc7bd02f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qkf8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ngspk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:47Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.582007 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5ltx2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8881599e-71e5-4578-b3f9-21f4fa2f57aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbm6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbm6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5ltx2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:47Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.604997 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.604793 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1721e985-40bd-4b2d-a0e8-d4e365094ba6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a895e0cc04880c3287ba725ac5484835235f1e066a94ce98fcf4b6133c5cbd2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a21a97385f856ec996105b597d0ae81c56ed28b625ef3f2905e1af6ec8f6a45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://646e01a58908ba2e703aae5eef53e1b981b0f0c3d1d3202487f2a8b94ee3f55d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31752b84d86bda30792b29e4c6ed963753cc576c5de784395e8e6dcbebaced47\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"ver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:40:26.813730 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814130 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:40:26.814139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:40:26.814143 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:40:26.814145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 03:40:26.821403 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821442 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821482 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821496 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 03:40:26.821509 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821514 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 03:40:26.821612 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 03:40:26.821619 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1206 03:40:26.822049 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16a92d501c9e626fe91e4a6793e426d10eb292cfbab766db8b8cc72e5dae94ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:47Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.605051 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.605272 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.605312 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.605332 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:47Z","lastTransitionTime":"2025-12-06T03:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.623055 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd230730-a9a1-47ed-b625-25e32231f271\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce321001135d0a2b878316f8003bf1bc54cba9a36f6a654ba376e1e7c0b4da71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af83814e6b2c5ed803157308ab895551e0826fd02f1ecc8b2751145c09d359da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9e6b48b5701dd330ef91712cdcd160afe9831a3881038bfb1e0397a9c8a3aa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2d359853c738e3e0e264f1a6a919b17afc71a362821b3ea170cda6695e3593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2d359853c738e3e0e264f1a6a919b17afc71a362821b3ea170cda6695e3593f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:47Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.639488 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-b2knv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a3ae46f-03b5-4999-9d58-c38ee222f68a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76665d4c127445009fd66078827b54607c9ef812c8011d1c0b4cde1c0516853a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk4hj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-b2knv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:47Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.658023 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2466610962332c6ca120fe055e4e8a3fc623088119dc1a412e06b745176f89e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:47Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.677628 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:47Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.695268 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6df38316-e0d3-4018-8d27-3620eba3a68d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57381e9589a2dc04a4d2733c37985f42594182128a374b7be0f471ea69138b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bf5e41701c7a2d45adc4740ceb4ac1785db17bc7cb15daed439f7556812d1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpxxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:47Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.708172 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.708228 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.708245 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.708269 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.708287 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:47Z","lastTransitionTime":"2025-12-06T03:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.711401 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kj7hf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd46dfa5-7d4a-49ed-8199-523f10cebdf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5eee8a0cf96ceb3785dac9df31bf8d31715d1c88fbea17c5e582b0a4554bb63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmkcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kj7hf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:47Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.731833 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb3f278fff47c900c56d8a563d5ed75a2ae84e2034b0f45aad4083790b9e7946\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b10ca16ac26904ec5e6245344b71d836888cae9b069cae6a7a4831d609752db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:47Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.748042 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9de75acd-1ba9-4322-ae49-791724b4b6a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://990943a2a48d79458297f4e5cf4178bbd6e9eafbcadc154a803846c589474e85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18a08a15442f04e37a5aa71afbcc293e80b876477d60cbe71d562c75d141a9ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18a08a15442f04e37a5aa71afbcc293e80b876477d60cbe71d562c75d141a9ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:47Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.769184 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2093b6-c4b5-4db8-be16-fbc859885e8d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0d6e582161f0d9126126e03b27211be22f6b5f797154928c92fa13a8118f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03004a29aefe7fe2452fb9f6663c0807afccb055d7ddbf6b907f305e6a19c81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1801ac64db4ba91c2e8a249e3651d3e50ab62e9f2c87312ce9bf7f38518a82cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6e454507cadbb982a566a3692995aaa3c733ff1cd7fb4870476602190e1eeba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:47Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.789580 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:47Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.811034 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.811139 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.811166 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.811198 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.811218 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:47Z","lastTransitionTime":"2025-12-06T03:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.814404 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fad9b63d84f41fccf696cd313d7ca7f72b3d3251d6c1b7341573a3fadf1e378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:47Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.914825 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.914890 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.914907 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.914931 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:47 crc kubenswrapper[4802]: I1206 03:41:47.914949 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:47Z","lastTransitionTime":"2025-12-06T03:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:48 crc kubenswrapper[4802]: I1206 03:41:48.018007 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:48 crc kubenswrapper[4802]: I1206 03:41:48.018059 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:48 crc kubenswrapper[4802]: I1206 03:41:48.018074 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:48 crc kubenswrapper[4802]: I1206 03:41:48.018098 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:48 crc kubenswrapper[4802]: I1206 03:41:48.018115 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:48Z","lastTransitionTime":"2025-12-06T03:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:48 crc kubenswrapper[4802]: I1206 03:41:48.121747 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:48 crc kubenswrapper[4802]: I1206 03:41:48.122238 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:48 crc kubenswrapper[4802]: I1206 03:41:48.122477 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:48 crc kubenswrapper[4802]: I1206 03:41:48.122726 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:48 crc kubenswrapper[4802]: I1206 03:41:48.123037 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:48Z","lastTransitionTime":"2025-12-06T03:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:48 crc kubenswrapper[4802]: I1206 03:41:48.226295 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:48 crc kubenswrapper[4802]: I1206 03:41:48.226354 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:48 crc kubenswrapper[4802]: I1206 03:41:48.226371 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:48 crc kubenswrapper[4802]: I1206 03:41:48.226392 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:48 crc kubenswrapper[4802]: I1206 03:41:48.226411 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:48Z","lastTransitionTime":"2025-12-06T03:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:48 crc kubenswrapper[4802]: I1206 03:41:48.329498 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:48 crc kubenswrapper[4802]: I1206 03:41:48.329874 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:48 crc kubenswrapper[4802]: I1206 03:41:48.330006 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:48 crc kubenswrapper[4802]: I1206 03:41:48.330144 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:48 crc kubenswrapper[4802]: I1206 03:41:48.330263 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:48Z","lastTransitionTime":"2025-12-06T03:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:48 crc kubenswrapper[4802]: I1206 03:41:48.433062 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:48 crc kubenswrapper[4802]: I1206 03:41:48.433655 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:48 crc kubenswrapper[4802]: I1206 03:41:48.434027 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:48 crc kubenswrapper[4802]: I1206 03:41:48.434329 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:48 crc kubenswrapper[4802]: I1206 03:41:48.434637 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:48Z","lastTransitionTime":"2025-12-06T03:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:48 crc kubenswrapper[4802]: I1206 03:41:48.449740 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:41:48 crc kubenswrapper[4802]: I1206 03:41:48.449833 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:41:48 crc kubenswrapper[4802]: I1206 03:41:48.449896 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:41:48 crc kubenswrapper[4802]: E1206 03:41:48.450286 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:41:48 crc kubenswrapper[4802]: E1206 03:41:48.450347 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5ltx2" podUID="8881599e-71e5-4578-b3f9-21f4fa2f57aa" Dec 06 03:41:48 crc kubenswrapper[4802]: I1206 03:41:48.449932 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:41:48 crc kubenswrapper[4802]: E1206 03:41:48.450445 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:41:48 crc kubenswrapper[4802]: E1206 03:41:48.450555 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:41:48 crc kubenswrapper[4802]: I1206 03:41:48.542454 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:48 crc kubenswrapper[4802]: I1206 03:41:48.542878 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:48 crc kubenswrapper[4802]: I1206 03:41:48.543061 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:48 crc kubenswrapper[4802]: I1206 03:41:48.543287 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:48 crc kubenswrapper[4802]: I1206 03:41:48.543468 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:48Z","lastTransitionTime":"2025-12-06T03:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:48 crc kubenswrapper[4802]: I1206 03:41:48.646324 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:48 crc kubenswrapper[4802]: I1206 03:41:48.646383 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:48 crc kubenswrapper[4802]: I1206 03:41:48.646398 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:48 crc kubenswrapper[4802]: I1206 03:41:48.646421 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:48 crc kubenswrapper[4802]: I1206 03:41:48.646437 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:48Z","lastTransitionTime":"2025-12-06T03:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:48 crc kubenswrapper[4802]: I1206 03:41:48.750132 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:48 crc kubenswrapper[4802]: I1206 03:41:48.750228 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:48 crc kubenswrapper[4802]: I1206 03:41:48.750279 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:48 crc kubenswrapper[4802]: I1206 03:41:48.750304 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:48 crc kubenswrapper[4802]: I1206 03:41:48.750321 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:48Z","lastTransitionTime":"2025-12-06T03:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:48 crc kubenswrapper[4802]: I1206 03:41:48.853275 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:48 crc kubenswrapper[4802]: I1206 03:41:48.853344 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:48 crc kubenswrapper[4802]: I1206 03:41:48.853363 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:48 crc kubenswrapper[4802]: I1206 03:41:48.853426 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:48 crc kubenswrapper[4802]: I1206 03:41:48.853448 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:48Z","lastTransitionTime":"2025-12-06T03:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:48 crc kubenswrapper[4802]: I1206 03:41:48.957092 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:48 crc kubenswrapper[4802]: I1206 03:41:48.957142 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:48 crc kubenswrapper[4802]: I1206 03:41:48.957159 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:48 crc kubenswrapper[4802]: I1206 03:41:48.957185 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:48 crc kubenswrapper[4802]: I1206 03:41:48.957207 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:48Z","lastTransitionTime":"2025-12-06T03:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:49 crc kubenswrapper[4802]: I1206 03:41:49.060243 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:49 crc kubenswrapper[4802]: I1206 03:41:49.060310 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:49 crc kubenswrapper[4802]: I1206 03:41:49.060333 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:49 crc kubenswrapper[4802]: I1206 03:41:49.060362 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:49 crc kubenswrapper[4802]: I1206 03:41:49.060383 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:49Z","lastTransitionTime":"2025-12-06T03:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:49 crc kubenswrapper[4802]: I1206 03:41:49.164120 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:49 crc kubenswrapper[4802]: I1206 03:41:49.164214 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:49 crc kubenswrapper[4802]: I1206 03:41:49.164234 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:49 crc kubenswrapper[4802]: I1206 03:41:49.164263 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:49 crc kubenswrapper[4802]: I1206 03:41:49.164287 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:49Z","lastTransitionTime":"2025-12-06T03:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:49 crc kubenswrapper[4802]: I1206 03:41:49.267396 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:49 crc kubenswrapper[4802]: I1206 03:41:49.267470 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:49 crc kubenswrapper[4802]: I1206 03:41:49.267494 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:49 crc kubenswrapper[4802]: I1206 03:41:49.267522 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:49 crc kubenswrapper[4802]: I1206 03:41:49.267541 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:49Z","lastTransitionTime":"2025-12-06T03:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:49 crc kubenswrapper[4802]: I1206 03:41:49.370247 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:49 crc kubenswrapper[4802]: I1206 03:41:49.370310 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:49 crc kubenswrapper[4802]: I1206 03:41:49.370332 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:49 crc kubenswrapper[4802]: I1206 03:41:49.370359 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:49 crc kubenswrapper[4802]: I1206 03:41:49.370382 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:49Z","lastTransitionTime":"2025-12-06T03:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:49 crc kubenswrapper[4802]: I1206 03:41:49.469497 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 06 03:41:49 crc kubenswrapper[4802]: I1206 03:41:49.475677 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:49 crc kubenswrapper[4802]: I1206 03:41:49.475741 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:49 crc kubenswrapper[4802]: I1206 03:41:49.475788 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:49 crc kubenswrapper[4802]: I1206 03:41:49.475806 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:49 crc kubenswrapper[4802]: I1206 03:41:49.475819 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:49Z","lastTransitionTime":"2025-12-06T03:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:49 crc kubenswrapper[4802]: I1206 03:41:49.579300 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:49 crc kubenswrapper[4802]: I1206 03:41:49.579359 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:49 crc kubenswrapper[4802]: I1206 03:41:49.579378 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:49 crc kubenswrapper[4802]: I1206 03:41:49.579404 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:49 crc kubenswrapper[4802]: I1206 03:41:49.579424 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:49Z","lastTransitionTime":"2025-12-06T03:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:49 crc kubenswrapper[4802]: I1206 03:41:49.683050 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:49 crc kubenswrapper[4802]: I1206 03:41:49.683097 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:49 crc kubenswrapper[4802]: I1206 03:41:49.683109 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:49 crc kubenswrapper[4802]: I1206 03:41:49.683129 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:49 crc kubenswrapper[4802]: I1206 03:41:49.683142 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:49Z","lastTransitionTime":"2025-12-06T03:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:49 crc kubenswrapper[4802]: I1206 03:41:49.786395 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:49 crc kubenswrapper[4802]: I1206 03:41:49.786453 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:49 crc kubenswrapper[4802]: I1206 03:41:49.786470 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:49 crc kubenswrapper[4802]: I1206 03:41:49.786495 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:49 crc kubenswrapper[4802]: I1206 03:41:49.786514 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:49Z","lastTransitionTime":"2025-12-06T03:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:49 crc kubenswrapper[4802]: I1206 03:41:49.889300 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:49 crc kubenswrapper[4802]: I1206 03:41:49.889368 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:49 crc kubenswrapper[4802]: I1206 03:41:49.889388 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:49 crc kubenswrapper[4802]: I1206 03:41:49.889417 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:49 crc kubenswrapper[4802]: I1206 03:41:49.889435 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:49Z","lastTransitionTime":"2025-12-06T03:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:49 crc kubenswrapper[4802]: I1206 03:41:49.992889 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:49 crc kubenswrapper[4802]: I1206 03:41:49.992944 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:49 crc kubenswrapper[4802]: I1206 03:41:49.992961 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:49 crc kubenswrapper[4802]: I1206 03:41:49.992988 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:49 crc kubenswrapper[4802]: I1206 03:41:49.993006 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:49Z","lastTransitionTime":"2025-12-06T03:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:50 crc kubenswrapper[4802]: I1206 03:41:50.096146 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:50 crc kubenswrapper[4802]: I1206 03:41:50.096230 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:50 crc kubenswrapper[4802]: I1206 03:41:50.096258 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:50 crc kubenswrapper[4802]: I1206 03:41:50.096291 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:50 crc kubenswrapper[4802]: I1206 03:41:50.096317 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:50Z","lastTransitionTime":"2025-12-06T03:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:50 crc kubenswrapper[4802]: I1206 03:41:50.198800 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:50 crc kubenswrapper[4802]: I1206 03:41:50.198840 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:50 crc kubenswrapper[4802]: I1206 03:41:50.198854 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:50 crc kubenswrapper[4802]: I1206 03:41:50.198871 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:50 crc kubenswrapper[4802]: I1206 03:41:50.198883 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:50Z","lastTransitionTime":"2025-12-06T03:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:50 crc kubenswrapper[4802]: I1206 03:41:50.300700 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:50 crc kubenswrapper[4802]: I1206 03:41:50.300730 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:50 crc kubenswrapper[4802]: I1206 03:41:50.300742 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:50 crc kubenswrapper[4802]: I1206 03:41:50.300778 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:50 crc kubenswrapper[4802]: I1206 03:41:50.300790 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:50Z","lastTransitionTime":"2025-12-06T03:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:50 crc kubenswrapper[4802]: I1206 03:41:50.402921 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:50 crc kubenswrapper[4802]: I1206 03:41:50.402968 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:50 crc kubenswrapper[4802]: I1206 03:41:50.402976 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:50 crc kubenswrapper[4802]: I1206 03:41:50.402988 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:50 crc kubenswrapper[4802]: I1206 03:41:50.402998 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:50Z","lastTransitionTime":"2025-12-06T03:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:50 crc kubenswrapper[4802]: I1206 03:41:50.450160 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:41:50 crc kubenswrapper[4802]: I1206 03:41:50.450192 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:41:50 crc kubenswrapper[4802]: I1206 03:41:50.450205 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:41:50 crc kubenswrapper[4802]: I1206 03:41:50.450324 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:41:50 crc kubenswrapper[4802]: E1206 03:41:50.450554 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:41:50 crc kubenswrapper[4802]: E1206 03:41:50.450692 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:41:50 crc kubenswrapper[4802]: E1206 03:41:50.450902 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5ltx2" podUID="8881599e-71e5-4578-b3f9-21f4fa2f57aa" Dec 06 03:41:50 crc kubenswrapper[4802]: E1206 03:41:50.450996 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:41:50 crc kubenswrapper[4802]: I1206 03:41:50.505611 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:50 crc kubenswrapper[4802]: I1206 03:41:50.505703 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:50 crc kubenswrapper[4802]: I1206 03:41:50.505721 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:50 crc kubenswrapper[4802]: I1206 03:41:50.505744 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:50 crc kubenswrapper[4802]: I1206 03:41:50.505795 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:50Z","lastTransitionTime":"2025-12-06T03:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:50 crc kubenswrapper[4802]: I1206 03:41:50.608487 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:50 crc kubenswrapper[4802]: I1206 03:41:50.608549 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:50 crc kubenswrapper[4802]: I1206 03:41:50.608567 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:50 crc kubenswrapper[4802]: I1206 03:41:50.608592 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:50 crc kubenswrapper[4802]: I1206 03:41:50.608616 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:50Z","lastTransitionTime":"2025-12-06T03:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:50 crc kubenswrapper[4802]: I1206 03:41:50.711337 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:50 crc kubenswrapper[4802]: I1206 03:41:50.711399 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:50 crc kubenswrapper[4802]: I1206 03:41:50.711420 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:50 crc kubenswrapper[4802]: I1206 03:41:50.711454 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:50 crc kubenswrapper[4802]: I1206 03:41:50.711479 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:50Z","lastTransitionTime":"2025-12-06T03:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:50 crc kubenswrapper[4802]: I1206 03:41:50.815163 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:50 crc kubenswrapper[4802]: I1206 03:41:50.815231 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:50 crc kubenswrapper[4802]: I1206 03:41:50.815255 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:50 crc kubenswrapper[4802]: I1206 03:41:50.815287 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:50 crc kubenswrapper[4802]: I1206 03:41:50.815309 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:50Z","lastTransitionTime":"2025-12-06T03:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:50 crc kubenswrapper[4802]: I1206 03:41:50.918554 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:50 crc kubenswrapper[4802]: I1206 03:41:50.918599 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:50 crc kubenswrapper[4802]: I1206 03:41:50.918609 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:50 crc kubenswrapper[4802]: I1206 03:41:50.918624 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:50 crc kubenswrapper[4802]: I1206 03:41:50.918635 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:50Z","lastTransitionTime":"2025-12-06T03:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.004455 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.004493 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.004508 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.004528 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.004539 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:51Z","lastTransitionTime":"2025-12-06T03:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:51 crc kubenswrapper[4802]: E1206 03:41:51.025001 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0e01ca3a-758f-44bc-b4a2-30c1bc04fda0\\\",\\\"systemUUID\\\":\\\"c7c815e4-d2a4-4244-b035-f988fc95e215\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:51Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.035133 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.035188 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.035201 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.035223 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.035238 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:51Z","lastTransitionTime":"2025-12-06T03:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:51 crc kubenswrapper[4802]: E1206 03:41:51.056056 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0e01ca3a-758f-44bc-b4a2-30c1bc04fda0\\\",\\\"systemUUID\\\":\\\"c7c815e4-d2a4-4244-b035-f988fc95e215\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:51Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.062120 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.062195 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.062219 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.062250 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.062273 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:51Z","lastTransitionTime":"2025-12-06T03:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:51 crc kubenswrapper[4802]: E1206 03:41:51.084900 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0e01ca3a-758f-44bc-b4a2-30c1bc04fda0\\\",\\\"systemUUID\\\":\\\"c7c815e4-d2a4-4244-b035-f988fc95e215\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:51Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.090342 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.090437 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.090458 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.090485 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.090503 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:51Z","lastTransitionTime":"2025-12-06T03:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:51 crc kubenswrapper[4802]: E1206 03:41:51.114970 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0e01ca3a-758f-44bc-b4a2-30c1bc04fda0\\\",\\\"systemUUID\\\":\\\"c7c815e4-d2a4-4244-b035-f988fc95e215\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:51Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.119945 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.120054 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.120122 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.120158 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.120177 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:51Z","lastTransitionTime":"2025-12-06T03:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:51 crc kubenswrapper[4802]: E1206 03:41:51.141652 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:41:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0e01ca3a-758f-44bc-b4a2-30c1bc04fda0\\\",\\\"systemUUID\\\":\\\"c7c815e4-d2a4-4244-b035-f988fc95e215\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:51Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:51 crc kubenswrapper[4802]: E1206 03:41:51.141853 4802 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.144095 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.144141 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.144184 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.144204 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.144222 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:51Z","lastTransitionTime":"2025-12-06T03:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.247743 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.247854 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.247876 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.247902 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.247921 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:51Z","lastTransitionTime":"2025-12-06T03:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.351528 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.351591 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.351612 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.351638 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.351655 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:51Z","lastTransitionTime":"2025-12-06T03:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.454316 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.454426 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.454447 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.454479 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.454503 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:51Z","lastTransitionTime":"2025-12-06T03:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.557450 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.557504 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.557520 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.557544 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.557561 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:51Z","lastTransitionTime":"2025-12-06T03:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.660542 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.660613 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.660636 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.660665 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.660686 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:51Z","lastTransitionTime":"2025-12-06T03:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.763827 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.763897 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.763922 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.763951 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.763972 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:51Z","lastTransitionTime":"2025-12-06T03:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.867175 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.867253 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.867278 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.867308 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.867330 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:51Z","lastTransitionTime":"2025-12-06T03:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.970868 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.970993 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.971019 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.971060 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:51 crc kubenswrapper[4802]: I1206 03:41:51.971086 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:51Z","lastTransitionTime":"2025-12-06T03:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:52 crc kubenswrapper[4802]: I1206 03:41:52.073499 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:52 crc kubenswrapper[4802]: I1206 03:41:52.073570 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:52 crc kubenswrapper[4802]: I1206 03:41:52.073595 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:52 crc kubenswrapper[4802]: I1206 03:41:52.073626 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:52 crc kubenswrapper[4802]: I1206 03:41:52.073647 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:52Z","lastTransitionTime":"2025-12-06T03:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:52 crc kubenswrapper[4802]: I1206 03:41:52.176798 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:52 crc kubenswrapper[4802]: I1206 03:41:52.176861 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:52 crc kubenswrapper[4802]: I1206 03:41:52.176873 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:52 crc kubenswrapper[4802]: I1206 03:41:52.176890 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:52 crc kubenswrapper[4802]: I1206 03:41:52.176902 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:52Z","lastTransitionTime":"2025-12-06T03:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:52 crc kubenswrapper[4802]: I1206 03:41:52.280477 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:52 crc kubenswrapper[4802]: I1206 03:41:52.280562 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:52 crc kubenswrapper[4802]: I1206 03:41:52.280583 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:52 crc kubenswrapper[4802]: I1206 03:41:52.280618 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:52 crc kubenswrapper[4802]: I1206 03:41:52.280641 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:52Z","lastTransitionTime":"2025-12-06T03:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:52 crc kubenswrapper[4802]: I1206 03:41:52.385063 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:52 crc kubenswrapper[4802]: I1206 03:41:52.385129 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:52 crc kubenswrapper[4802]: I1206 03:41:52.385148 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:52 crc kubenswrapper[4802]: I1206 03:41:52.385175 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:52 crc kubenswrapper[4802]: I1206 03:41:52.385194 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:52Z","lastTransitionTime":"2025-12-06T03:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:52 crc kubenswrapper[4802]: I1206 03:41:52.449855 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:41:52 crc kubenswrapper[4802]: I1206 03:41:52.449952 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:41:52 crc kubenswrapper[4802]: I1206 03:41:52.449900 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:41:52 crc kubenswrapper[4802]: I1206 03:41:52.450071 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:41:52 crc kubenswrapper[4802]: E1206 03:41:52.450259 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5ltx2" podUID="8881599e-71e5-4578-b3f9-21f4fa2f57aa" Dec 06 03:41:52 crc kubenswrapper[4802]: E1206 03:41:52.450845 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:41:52 crc kubenswrapper[4802]: E1206 03:41:52.451053 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:41:52 crc kubenswrapper[4802]: E1206 03:41:52.451175 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:41:52 crc kubenswrapper[4802]: I1206 03:41:52.451572 4802 scope.go:117] "RemoveContainer" containerID="bf554aaba0e773f6a6c1f29bb1819b7ae25b2b2db255c4bd0a16c512f314e9cf" Dec 06 03:41:52 crc kubenswrapper[4802]: E1206 03:41:52.452001 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pvm6q_openshift-ovn-kubernetes(eafce4e0-e7fb-4877-b0ab-3283829ba1ef)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" podUID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" Dec 06 03:41:52 crc kubenswrapper[4802]: I1206 03:41:52.488527 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:52 crc kubenswrapper[4802]: I1206 03:41:52.488604 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:52 crc kubenswrapper[4802]: I1206 03:41:52.488624 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:52 crc kubenswrapper[4802]: I1206 03:41:52.488654 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:52 crc kubenswrapper[4802]: I1206 03:41:52.488675 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:52Z","lastTransitionTime":"2025-12-06T03:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:52 crc kubenswrapper[4802]: I1206 03:41:52.592471 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:52 crc kubenswrapper[4802]: I1206 03:41:52.592562 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:52 crc kubenswrapper[4802]: I1206 03:41:52.592592 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:52 crc kubenswrapper[4802]: I1206 03:41:52.592623 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:52 crc kubenswrapper[4802]: I1206 03:41:52.592645 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:52Z","lastTransitionTime":"2025-12-06T03:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:52 crc kubenswrapper[4802]: I1206 03:41:52.695526 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:52 crc kubenswrapper[4802]: I1206 03:41:52.695583 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:52 crc kubenswrapper[4802]: I1206 03:41:52.695601 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:52 crc kubenswrapper[4802]: I1206 03:41:52.695624 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:52 crc kubenswrapper[4802]: I1206 03:41:52.695643 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:52Z","lastTransitionTime":"2025-12-06T03:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:52 crc kubenswrapper[4802]: I1206 03:41:52.798393 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:52 crc kubenswrapper[4802]: I1206 03:41:52.798434 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:52 crc kubenswrapper[4802]: I1206 03:41:52.798443 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:52 crc kubenswrapper[4802]: I1206 03:41:52.798460 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:52 crc kubenswrapper[4802]: I1206 03:41:52.798470 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:52Z","lastTransitionTime":"2025-12-06T03:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:52 crc kubenswrapper[4802]: I1206 03:41:52.900915 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:52 crc kubenswrapper[4802]: I1206 03:41:52.900969 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:52 crc kubenswrapper[4802]: I1206 03:41:52.900982 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:52 crc kubenswrapper[4802]: I1206 03:41:52.900999 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:52 crc kubenswrapper[4802]: I1206 03:41:52.901011 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:52Z","lastTransitionTime":"2025-12-06T03:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:53 crc kubenswrapper[4802]: I1206 03:41:53.003819 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:53 crc kubenswrapper[4802]: I1206 03:41:53.003886 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:53 crc kubenswrapper[4802]: I1206 03:41:53.003903 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:53 crc kubenswrapper[4802]: I1206 03:41:53.003926 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:53 crc kubenswrapper[4802]: I1206 03:41:53.003944 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:53Z","lastTransitionTime":"2025-12-06T03:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:53 crc kubenswrapper[4802]: I1206 03:41:53.106287 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:53 crc kubenswrapper[4802]: I1206 03:41:53.106355 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:53 crc kubenswrapper[4802]: I1206 03:41:53.106372 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:53 crc kubenswrapper[4802]: I1206 03:41:53.106398 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:53 crc kubenswrapper[4802]: I1206 03:41:53.106419 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:53Z","lastTransitionTime":"2025-12-06T03:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:53 crc kubenswrapper[4802]: I1206 03:41:53.209858 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:53 crc kubenswrapper[4802]: I1206 03:41:53.209918 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:53 crc kubenswrapper[4802]: I1206 03:41:53.209934 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:53 crc kubenswrapper[4802]: I1206 03:41:53.209961 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:53 crc kubenswrapper[4802]: I1206 03:41:53.209978 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:53Z","lastTransitionTime":"2025-12-06T03:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:53 crc kubenswrapper[4802]: I1206 03:41:53.312259 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:53 crc kubenswrapper[4802]: I1206 03:41:53.312299 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:53 crc kubenswrapper[4802]: I1206 03:41:53.312307 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:53 crc kubenswrapper[4802]: I1206 03:41:53.312327 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:53 crc kubenswrapper[4802]: I1206 03:41:53.312339 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:53Z","lastTransitionTime":"2025-12-06T03:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:53 crc kubenswrapper[4802]: I1206 03:41:53.415113 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:53 crc kubenswrapper[4802]: I1206 03:41:53.415163 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:53 crc kubenswrapper[4802]: I1206 03:41:53.415175 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:53 crc kubenswrapper[4802]: I1206 03:41:53.415193 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:53 crc kubenswrapper[4802]: I1206 03:41:53.415204 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:53Z","lastTransitionTime":"2025-12-06T03:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:53 crc kubenswrapper[4802]: I1206 03:41:53.517454 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:53 crc kubenswrapper[4802]: I1206 03:41:53.517521 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:53 crc kubenswrapper[4802]: I1206 03:41:53.517538 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:53 crc kubenswrapper[4802]: I1206 03:41:53.517561 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:53 crc kubenswrapper[4802]: I1206 03:41:53.517577 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:53Z","lastTransitionTime":"2025-12-06T03:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:53 crc kubenswrapper[4802]: I1206 03:41:53.620268 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:53 crc kubenswrapper[4802]: I1206 03:41:53.620348 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:53 crc kubenswrapper[4802]: I1206 03:41:53.620372 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:53 crc kubenswrapper[4802]: I1206 03:41:53.620403 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:53 crc kubenswrapper[4802]: I1206 03:41:53.620425 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:53Z","lastTransitionTime":"2025-12-06T03:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:53 crc kubenswrapper[4802]: I1206 03:41:53.727509 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:53 crc kubenswrapper[4802]: I1206 03:41:53.727604 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:53 crc kubenswrapper[4802]: I1206 03:41:53.727663 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:53 crc kubenswrapper[4802]: I1206 03:41:53.727687 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:53 crc kubenswrapper[4802]: I1206 03:41:53.727736 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:53Z","lastTransitionTime":"2025-12-06T03:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:53 crc kubenswrapper[4802]: I1206 03:41:53.830113 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:53 crc kubenswrapper[4802]: I1206 03:41:53.830167 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:53 crc kubenswrapper[4802]: I1206 03:41:53.830179 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:53 crc kubenswrapper[4802]: I1206 03:41:53.830196 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:53 crc kubenswrapper[4802]: I1206 03:41:53.830210 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:53Z","lastTransitionTime":"2025-12-06T03:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:53 crc kubenswrapper[4802]: I1206 03:41:53.932388 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:53 crc kubenswrapper[4802]: I1206 03:41:53.932449 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:53 crc kubenswrapper[4802]: I1206 03:41:53.932466 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:53 crc kubenswrapper[4802]: I1206 03:41:53.932488 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:53 crc kubenswrapper[4802]: I1206 03:41:53.932501 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:53Z","lastTransitionTime":"2025-12-06T03:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:54 crc kubenswrapper[4802]: I1206 03:41:54.035725 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:54 crc kubenswrapper[4802]: I1206 03:41:54.035806 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:54 crc kubenswrapper[4802]: I1206 03:41:54.035819 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:54 crc kubenswrapper[4802]: I1206 03:41:54.035837 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:54 crc kubenswrapper[4802]: I1206 03:41:54.035848 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:54Z","lastTransitionTime":"2025-12-06T03:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:54 crc kubenswrapper[4802]: I1206 03:41:54.139464 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:54 crc kubenswrapper[4802]: I1206 03:41:54.139540 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:54 crc kubenswrapper[4802]: I1206 03:41:54.139564 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:54 crc kubenswrapper[4802]: I1206 03:41:54.139599 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:54 crc kubenswrapper[4802]: I1206 03:41:54.139624 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:54Z","lastTransitionTime":"2025-12-06T03:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:54 crc kubenswrapper[4802]: I1206 03:41:54.242255 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:54 crc kubenswrapper[4802]: I1206 03:41:54.242303 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:54 crc kubenswrapper[4802]: I1206 03:41:54.242316 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:54 crc kubenswrapper[4802]: I1206 03:41:54.242333 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:54 crc kubenswrapper[4802]: I1206 03:41:54.242346 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:54Z","lastTransitionTime":"2025-12-06T03:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:54 crc kubenswrapper[4802]: I1206 03:41:54.345622 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:54 crc kubenswrapper[4802]: I1206 03:41:54.345693 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:54 crc kubenswrapper[4802]: I1206 03:41:54.345711 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:54 crc kubenswrapper[4802]: I1206 03:41:54.345735 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:54 crc kubenswrapper[4802]: I1206 03:41:54.345789 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:54Z","lastTransitionTime":"2025-12-06T03:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:54 crc kubenswrapper[4802]: I1206 03:41:54.448340 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:54 crc kubenswrapper[4802]: I1206 03:41:54.448418 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:54 crc kubenswrapper[4802]: I1206 03:41:54.448437 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:54 crc kubenswrapper[4802]: I1206 03:41:54.448468 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:54 crc kubenswrapper[4802]: I1206 03:41:54.448487 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:54Z","lastTransitionTime":"2025-12-06T03:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:54 crc kubenswrapper[4802]: I1206 03:41:54.449224 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:41:54 crc kubenswrapper[4802]: I1206 03:41:54.449249 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:41:54 crc kubenswrapper[4802]: I1206 03:41:54.449274 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:41:54 crc kubenswrapper[4802]: E1206 03:41:54.449319 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:41:54 crc kubenswrapper[4802]: I1206 03:41:54.449332 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:41:54 crc kubenswrapper[4802]: E1206 03:41:54.449493 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5ltx2" podUID="8881599e-71e5-4578-b3f9-21f4fa2f57aa" Dec 06 03:41:54 crc kubenswrapper[4802]: E1206 03:41:54.449594 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:41:54 crc kubenswrapper[4802]: E1206 03:41:54.449662 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:41:54 crc kubenswrapper[4802]: I1206 03:41:54.551723 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:54 crc kubenswrapper[4802]: I1206 03:41:54.551796 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:54 crc kubenswrapper[4802]: I1206 03:41:54.551808 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:54 crc kubenswrapper[4802]: I1206 03:41:54.551824 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:54 crc kubenswrapper[4802]: I1206 03:41:54.551835 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:54Z","lastTransitionTime":"2025-12-06T03:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:54 crc kubenswrapper[4802]: I1206 03:41:54.655067 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:54 crc kubenswrapper[4802]: I1206 03:41:54.655137 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:54 crc kubenswrapper[4802]: I1206 03:41:54.655160 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:54 crc kubenswrapper[4802]: I1206 03:41:54.655190 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:54 crc kubenswrapper[4802]: I1206 03:41:54.655244 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:54Z","lastTransitionTime":"2025-12-06T03:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:54 crc kubenswrapper[4802]: I1206 03:41:54.758325 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:54 crc kubenswrapper[4802]: I1206 03:41:54.758407 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:54 crc kubenswrapper[4802]: I1206 03:41:54.758427 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:54 crc kubenswrapper[4802]: I1206 03:41:54.758453 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:54 crc kubenswrapper[4802]: I1206 03:41:54.758470 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:54Z","lastTransitionTime":"2025-12-06T03:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:54 crc kubenswrapper[4802]: I1206 03:41:54.861172 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:54 crc kubenswrapper[4802]: I1206 03:41:54.861240 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:54 crc kubenswrapper[4802]: I1206 03:41:54.861258 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:54 crc kubenswrapper[4802]: I1206 03:41:54.861283 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:54 crc kubenswrapper[4802]: I1206 03:41:54.861308 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:54Z","lastTransitionTime":"2025-12-06T03:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:54 crc kubenswrapper[4802]: I1206 03:41:54.964876 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:54 crc kubenswrapper[4802]: I1206 03:41:54.964955 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:54 crc kubenswrapper[4802]: I1206 03:41:54.964978 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:54 crc kubenswrapper[4802]: I1206 03:41:54.965011 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:54 crc kubenswrapper[4802]: I1206 03:41:54.965031 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:54Z","lastTransitionTime":"2025-12-06T03:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:55 crc kubenswrapper[4802]: I1206 03:41:55.066941 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:55 crc kubenswrapper[4802]: I1206 03:41:55.066996 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:55 crc kubenswrapper[4802]: I1206 03:41:55.067012 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:55 crc kubenswrapper[4802]: I1206 03:41:55.067036 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:55 crc kubenswrapper[4802]: I1206 03:41:55.067054 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:55Z","lastTransitionTime":"2025-12-06T03:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:55 crc kubenswrapper[4802]: I1206 03:41:55.170834 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:55 crc kubenswrapper[4802]: I1206 03:41:55.170893 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:55 crc kubenswrapper[4802]: I1206 03:41:55.170910 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:55 crc kubenswrapper[4802]: I1206 03:41:55.170933 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:55 crc kubenswrapper[4802]: I1206 03:41:55.170957 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:55Z","lastTransitionTime":"2025-12-06T03:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:55 crc kubenswrapper[4802]: I1206 03:41:55.274903 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:55 crc kubenswrapper[4802]: I1206 03:41:55.274967 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:55 crc kubenswrapper[4802]: I1206 03:41:55.274985 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:55 crc kubenswrapper[4802]: I1206 03:41:55.275007 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:55 crc kubenswrapper[4802]: I1206 03:41:55.275025 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:55Z","lastTransitionTime":"2025-12-06T03:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:55 crc kubenswrapper[4802]: I1206 03:41:55.378418 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:55 crc kubenswrapper[4802]: I1206 03:41:55.378479 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:55 crc kubenswrapper[4802]: I1206 03:41:55.378497 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:55 crc kubenswrapper[4802]: I1206 03:41:55.378521 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:55 crc kubenswrapper[4802]: I1206 03:41:55.378538 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:55Z","lastTransitionTime":"2025-12-06T03:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:55 crc kubenswrapper[4802]: I1206 03:41:55.481836 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:55 crc kubenswrapper[4802]: I1206 03:41:55.481886 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:55 crc kubenswrapper[4802]: I1206 03:41:55.481904 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:55 crc kubenswrapper[4802]: I1206 03:41:55.481928 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:55 crc kubenswrapper[4802]: I1206 03:41:55.481946 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:55Z","lastTransitionTime":"2025-12-06T03:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:55 crc kubenswrapper[4802]: I1206 03:41:55.585927 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:55 crc kubenswrapper[4802]: I1206 03:41:55.586025 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:55 crc kubenswrapper[4802]: I1206 03:41:55.586051 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:55 crc kubenswrapper[4802]: I1206 03:41:55.586084 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:55 crc kubenswrapper[4802]: I1206 03:41:55.586107 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:55Z","lastTransitionTime":"2025-12-06T03:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:55 crc kubenswrapper[4802]: I1206 03:41:55.689047 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:55 crc kubenswrapper[4802]: I1206 03:41:55.689091 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:55 crc kubenswrapper[4802]: I1206 03:41:55.689101 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:55 crc kubenswrapper[4802]: I1206 03:41:55.689117 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:55 crc kubenswrapper[4802]: I1206 03:41:55.689127 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:55Z","lastTransitionTime":"2025-12-06T03:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:55 crc kubenswrapper[4802]: I1206 03:41:55.791418 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:55 crc kubenswrapper[4802]: I1206 03:41:55.791490 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:55 crc kubenswrapper[4802]: I1206 03:41:55.791507 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:55 crc kubenswrapper[4802]: I1206 03:41:55.791533 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:55 crc kubenswrapper[4802]: I1206 03:41:55.791552 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:55Z","lastTransitionTime":"2025-12-06T03:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:55 crc kubenswrapper[4802]: I1206 03:41:55.894398 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:55 crc kubenswrapper[4802]: I1206 03:41:55.894462 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:55 crc kubenswrapper[4802]: I1206 03:41:55.894480 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:55 crc kubenswrapper[4802]: I1206 03:41:55.894505 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:55 crc kubenswrapper[4802]: I1206 03:41:55.894521 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:55Z","lastTransitionTime":"2025-12-06T03:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:55 crc kubenswrapper[4802]: I1206 03:41:55.997598 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:55 crc kubenswrapper[4802]: I1206 03:41:55.997667 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:55 crc kubenswrapper[4802]: I1206 03:41:55.997689 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:55 crc kubenswrapper[4802]: I1206 03:41:55.997713 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:55 crc kubenswrapper[4802]: I1206 03:41:55.997726 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:55Z","lastTransitionTime":"2025-12-06T03:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:56 crc kubenswrapper[4802]: I1206 03:41:56.100672 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:56 crc kubenswrapper[4802]: I1206 03:41:56.100733 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:56 crc kubenswrapper[4802]: I1206 03:41:56.100783 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:56 crc kubenswrapper[4802]: I1206 03:41:56.100807 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:56 crc kubenswrapper[4802]: I1206 03:41:56.100825 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:56Z","lastTransitionTime":"2025-12-06T03:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:56 crc kubenswrapper[4802]: I1206 03:41:56.204134 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:56 crc kubenswrapper[4802]: I1206 03:41:56.204200 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:56 crc kubenswrapper[4802]: I1206 03:41:56.204217 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:56 crc kubenswrapper[4802]: I1206 03:41:56.204243 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:56 crc kubenswrapper[4802]: I1206 03:41:56.204260 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:56Z","lastTransitionTime":"2025-12-06T03:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:56 crc kubenswrapper[4802]: I1206 03:41:56.306994 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:56 crc kubenswrapper[4802]: I1206 03:41:56.307068 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:56 crc kubenswrapper[4802]: I1206 03:41:56.307086 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:56 crc kubenswrapper[4802]: I1206 03:41:56.307112 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:56 crc kubenswrapper[4802]: I1206 03:41:56.307129 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:56Z","lastTransitionTime":"2025-12-06T03:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:56 crc kubenswrapper[4802]: I1206 03:41:56.410593 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:56 crc kubenswrapper[4802]: I1206 03:41:56.410638 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:56 crc kubenswrapper[4802]: I1206 03:41:56.410649 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:56 crc kubenswrapper[4802]: I1206 03:41:56.410665 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:56 crc kubenswrapper[4802]: I1206 03:41:56.410677 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:56Z","lastTransitionTime":"2025-12-06T03:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:56 crc kubenswrapper[4802]: I1206 03:41:56.449303 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:41:56 crc kubenswrapper[4802]: I1206 03:41:56.449376 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:41:56 crc kubenswrapper[4802]: I1206 03:41:56.449373 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:41:56 crc kubenswrapper[4802]: E1206 03:41:56.449447 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5ltx2" podUID="8881599e-71e5-4578-b3f9-21f4fa2f57aa" Dec 06 03:41:56 crc kubenswrapper[4802]: E1206 03:41:56.449564 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:41:56 crc kubenswrapper[4802]: I1206 03:41:56.449595 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:41:56 crc kubenswrapper[4802]: E1206 03:41:56.449685 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:41:56 crc kubenswrapper[4802]: E1206 03:41:56.449879 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:41:56 crc kubenswrapper[4802]: I1206 03:41:56.513721 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:56 crc kubenswrapper[4802]: I1206 03:41:56.513811 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:56 crc kubenswrapper[4802]: I1206 03:41:56.513829 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:56 crc kubenswrapper[4802]: I1206 03:41:56.513856 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:56 crc kubenswrapper[4802]: I1206 03:41:56.513874 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:56Z","lastTransitionTime":"2025-12-06T03:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:56 crc kubenswrapper[4802]: I1206 03:41:56.617120 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:56 crc kubenswrapper[4802]: I1206 03:41:56.617201 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:56 crc kubenswrapper[4802]: I1206 03:41:56.617225 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:56 crc kubenswrapper[4802]: I1206 03:41:56.617257 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:56 crc kubenswrapper[4802]: I1206 03:41:56.617283 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:56Z","lastTransitionTime":"2025-12-06T03:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:56 crc kubenswrapper[4802]: I1206 03:41:56.720513 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:56 crc kubenswrapper[4802]: I1206 03:41:56.720573 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:56 crc kubenswrapper[4802]: I1206 03:41:56.720589 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:56 crc kubenswrapper[4802]: I1206 03:41:56.720617 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:56 crc kubenswrapper[4802]: I1206 03:41:56.720635 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:56Z","lastTransitionTime":"2025-12-06T03:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:56 crc kubenswrapper[4802]: I1206 03:41:56.824252 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:56 crc kubenswrapper[4802]: I1206 03:41:56.824320 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:56 crc kubenswrapper[4802]: I1206 03:41:56.824353 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:56 crc kubenswrapper[4802]: I1206 03:41:56.824398 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:56 crc kubenswrapper[4802]: I1206 03:41:56.824424 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:56Z","lastTransitionTime":"2025-12-06T03:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:56 crc kubenswrapper[4802]: I1206 03:41:56.926932 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:56 crc kubenswrapper[4802]: I1206 03:41:56.927007 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:56 crc kubenswrapper[4802]: I1206 03:41:56.927032 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:56 crc kubenswrapper[4802]: I1206 03:41:56.927066 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:56 crc kubenswrapper[4802]: I1206 03:41:56.927090 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:56Z","lastTransitionTime":"2025-12-06T03:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.029916 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.029974 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.029990 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.030020 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.030037 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:57Z","lastTransitionTime":"2025-12-06T03:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.132792 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.132858 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.132882 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.132912 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.132935 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:57Z","lastTransitionTime":"2025-12-06T03:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.236625 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.236688 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.236711 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.236738 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.236792 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:57Z","lastTransitionTime":"2025-12-06T03:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.339559 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.339629 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.339649 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.339675 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.339698 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:57Z","lastTransitionTime":"2025-12-06T03:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.442057 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.442124 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.442146 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.442181 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.442204 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:57Z","lastTransitionTime":"2025-12-06T03:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.468150 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:57Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.491582 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xtw2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3a3bac2-3923-4507-a4de-1556b3ad473e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fed5ee4189d54456a3216f4fdb1ecf003fa47e29dbfe4b2b2f301edfb897d6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec6ae071e7bfec0dec2643eedf708867d8400816d2ed832c49a2661f572c7fef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7a821224ce6752f788f9f75591bdd98a5be5f82f94d2f4828ea5bd75e021cce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8fb21794f610afab0c1bf740018c3655229218b12749f615e477dbed2b9efd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cb6acaa8cf3f2c0be44f134f6fc15c60bf229998d90f2cacb6d51ebddcb6526\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24df18575298e8b953d2d1c34da099978f24cbdecb8b632683901feabaf9fab5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe85a4bab1e15f3b38594271760bd34032d836ae55baaa77dd2d265ba57e3a85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr56s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xtw2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:57Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.523823 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7730de4f033561ad77046e028cd1a3f0bfe2775387cbdfa35f47ea79909ec4cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a158aebd701ef46225d3e37109ae28ab58d107087bf06650c4b6d0055a3bbba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2357f37497fde64937eca7b38693893af745ab278653ebc79406e122a6d4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8941bd25107333b264d3f081a811fdfeea23ad92193fc61367665fd2cd1209c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f9cd4dc75e63f09f60b48b6da284643910bb32ba51f2a808cb4d9803f30042\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6796f65ee9c5bd0c18f205697135d17677c5daf3ba378cc992cba795e0188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf554aaba0e773f6a6c1f29bb1819b7ae25b2b2db255c4bd0a16c512f314e9cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf554aaba0e773f6a6c1f29bb1819b7ae25b2b2db255c4bd0a16c512f314e9cf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:41:23Z\\\",\\\"message\\\":\\\"42b7e8 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 03:41:22.975067 6823 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1206 03:41:22.975567 6823 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 03:41:22.975037 6823 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-kj7hf in node crc\\\\nI1206 03:41:22.975614 6823 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-kj7hf after 0 failed attempt(s)\\\\nI1206 03:41:22.975624 6823 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-kj7hf\\\\nF1206 03:41:22.974932 6823 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:41:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pvm6q_openshift-ovn-kubernetes(eafce4e0-e7fb-4877-b0ab-3283829ba1ef)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3cb210bfc97a439a7ff98c02c59f8ad16bc400dc214a616a44b43a8b01e2782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-62rc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pvm6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:57Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.544926 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.544991 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.545015 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.545046 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.545068 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:57Z","lastTransitionTime":"2025-12-06T03:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.558234 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22rrq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03112a5d-5882-4fac-9ae5-13525ff82fe0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://831dab65557627cbce0d055459052dae5b594f33b66858367ccd034ff6811924\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0205b089b12c42688203bb7ec0f27dc5f9a577dabdb446b811f184046f85dc3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:41:14Z\\\",\\\"message\\\":\\\"2025-12-06T03:40:28+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_384cd081-7034-46cf-bb48-841728494c60\\\\n2025-12-06T03:40:28+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_384cd081-7034-46cf-bb48-841728494c60 to /host/opt/cni/bin/\\\\n2025-12-06T03:40:29Z [verbose] multus-daemon started\\\\n2025-12-06T03:40:29Z [verbose] Readiness Indicator file check\\\\n2025-12-06T03:41:14Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:41:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trlt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22rrq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:57Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.576922 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ngspk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34808ab2-d21b-40b6-8fe3-e18bf9e5fc46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52509d21f5b4bfbd458162cba176ff82c627c7c4671204addf9d76f7392fd760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qkf8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9685a4d5d4f40ff4f59b6da6193785f340890453d1e422d3421a621dc7bd02f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qkf8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ngspk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:57Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.593449 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5ltx2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8881599e-71e5-4578-b3f9-21f4fa2f57aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbm6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbm6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5ltx2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:57Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.619637 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"646d0204-8842-4019-acfb-e46780c70e43\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4aa3d0fabd0ec4495a3b66418d59d266828d55d591782797864a9a0b261dd9a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7ee33d59413614773a292125b39f656b7a5962fa254a217966e625bd239ba7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://648ea0415efa4e81ad24f8af337ed2c8c0ba0a041aaf3ad7bb4e471eef14a877\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fc93fc85f068f1dc9e3c410331b403e582a0cce3828d53e4887ada0bdd0a77e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ab8a6728c484ad31865d6c09bc04fb41e5cc612d64b7d90e6b4edb4fa317236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adad8cbf6f6508ca0ad781d9e59b9d90566c02acaeb3a4b3208725c75bfda226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adad8cbf6f6508ca0ad781d9e59b9d90566c02acaeb3a4b3208725c75bfda226\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84ce35f5ad94653c857069ade3fcdf326b1941c3b4fae4b75a695a5b1917f9aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84ce35f5ad94653c857069ade3fcdf326b1941c3b4fae4b75a695a5b1917f9aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://17f78cae7920f1f196ec9baf053ff98df6935df8078f8114a2b443aa2111d50f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17f78cae7920f1f196ec9baf053ff98df6935df8078f8114a2b443aa2111d50f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:57Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.633439 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1721e985-40bd-4b2d-a0e8-d4e365094ba6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a895e0cc04880c3287ba725ac5484835235f1e066a94ce98fcf4b6133c5cbd2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a21a97385f856ec996105b597d0ae81c56ed28b625ef3f2905e1af6ec8f6a45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://646e01a58908ba2e703aae5eef53e1b981b0f0c3d1d3202487f2a8b94ee3f55d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31752b84d86bda30792b29e4c6ed963753cc576c5de784395e8e6dcbebaced47\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"ver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:40:26.813730 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814130 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:40:26.814136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:40:26.814139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:40:26.814143 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:40:26.814145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 03:40:26.821403 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821442 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 03:40:26.821482 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821496 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 03:40:26.821509 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 03:40:26.821514 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 03:40:26.821612 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 03:40:26.821619 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1206 03:40:26.822049 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16a92d501c9e626fe91e4a6793e426d10eb292cfbab766db8b8cc72e5dae94ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:57Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.643993 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd230730-a9a1-47ed-b625-25e32231f271\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce321001135d0a2b878316f8003bf1bc54cba9a36f6a654ba376e1e7c0b4da71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af83814e6b2c5ed803157308ab895551e0826fd02f1ecc8b2751145c09d359da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9e6b48b5701dd330ef91712cdcd160afe9831a3881038bfb1e0397a9c8a3aa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2d359853c738e3e0e264f1a6a919b17afc71a362821b3ea170cda6695e3593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2d359853c738e3e0e264f1a6a919b17afc71a362821b3ea170cda6695e3593f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:57Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.647947 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.647992 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.648000 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.648016 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.648026 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:57Z","lastTransitionTime":"2025-12-06T03:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.657460 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-b2knv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a3ae46f-03b5-4999-9d58-c38ee222f68a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76665d4c127445009fd66078827b54607c9ef812c8011d1c0b4cde1c0516853a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk4hj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-b2knv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:57Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.670226 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2466610962332c6ca120fe055e4e8a3fc623088119dc1a412e06b745176f89e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:57Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.687199 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:57Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.699384 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6df38316-e0d3-4018-8d27-3620eba3a68d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57381e9589a2dc04a4d2733c37985f42594182128a374b7be0f471ea69138b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bf5e41701c7a2d45adc4740ceb4ac1785db17bc7cb15daed439f7556812d1e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfj4n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpxxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:57Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.713519 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kj7hf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd46dfa5-7d4a-49ed-8199-523f10cebdf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5eee8a0cf96ceb3785dac9df31bf8d31715d1c88fbea17c5e582b0a4554bb63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmkcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kj7hf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:57Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.725592 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb3f278fff47c900c56d8a563d5ed75a2ae84e2034b0f45aad4083790b9e7946\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b10ca16ac26904ec5e6245344b71d836888cae9b069cae6a7a4831d609752db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:57Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.737876 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9de75acd-1ba9-4322-ae49-791724b4b6a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://990943a2a48d79458297f4e5cf4178bbd6e9eafbcadc154a803846c589474e85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18a08a15442f04e37a5aa71afbcc293e80b876477d60cbe71d562c75d141a9ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18a08a15442f04e37a5aa71afbcc293e80b876477d60cbe71d562c75d141a9ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:40:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:57Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.749937 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.750007 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.750028 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.750052 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.750069 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:57Z","lastTransitionTime":"2025-12-06T03:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.753697 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2093b6-c4b5-4db8-be16-fbc859885e8d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0d6e582161f0d9126126e03b27211be22f6b5f797154928c92fa13a8118f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03004a29aefe7fe2452fb9f6663c0807afccb055d7ddbf6b907f305e6a19c81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1801ac64db4ba91c2e8a249e3651d3e50ab62e9f2c87312ce9bf7f38518a82cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6e454507cadbb982a566a3692995aaa3c733ff1cd7fb4870476602190e1eeba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:40:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:57Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.767219 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:57Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.779425 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:40:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fad9b63d84f41fccf696cd313d7ca7f72b3d3251d6c1b7341573a3fadf1e378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:41:57Z is after 2025-08-24T17:21:41Z" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.852498 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.852555 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.852572 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.852595 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.852612 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:57Z","lastTransitionTime":"2025-12-06T03:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.955691 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.955747 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.955799 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.955823 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:57 crc kubenswrapper[4802]: I1206 03:41:57.955840 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:57Z","lastTransitionTime":"2025-12-06T03:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:58 crc kubenswrapper[4802]: I1206 03:41:58.058363 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:58 crc kubenswrapper[4802]: I1206 03:41:58.058493 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:58 crc kubenswrapper[4802]: I1206 03:41:58.058525 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:58 crc kubenswrapper[4802]: I1206 03:41:58.058550 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:58 crc kubenswrapper[4802]: I1206 03:41:58.058569 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:58Z","lastTransitionTime":"2025-12-06T03:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:58 crc kubenswrapper[4802]: I1206 03:41:58.162342 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:58 crc kubenswrapper[4802]: I1206 03:41:58.162372 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:58 crc kubenswrapper[4802]: I1206 03:41:58.162380 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:58 crc kubenswrapper[4802]: I1206 03:41:58.162395 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:58 crc kubenswrapper[4802]: I1206 03:41:58.162403 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:58Z","lastTransitionTime":"2025-12-06T03:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:58 crc kubenswrapper[4802]: I1206 03:41:58.264997 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:58 crc kubenswrapper[4802]: I1206 03:41:58.265064 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:58 crc kubenswrapper[4802]: I1206 03:41:58.265082 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:58 crc kubenswrapper[4802]: I1206 03:41:58.265105 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:58 crc kubenswrapper[4802]: I1206 03:41:58.265124 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:58Z","lastTransitionTime":"2025-12-06T03:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:58 crc kubenswrapper[4802]: I1206 03:41:58.368729 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:58 crc kubenswrapper[4802]: I1206 03:41:58.368863 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:58 crc kubenswrapper[4802]: I1206 03:41:58.368888 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:58 crc kubenswrapper[4802]: I1206 03:41:58.368922 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:58 crc kubenswrapper[4802]: I1206 03:41:58.368946 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:58Z","lastTransitionTime":"2025-12-06T03:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:58 crc kubenswrapper[4802]: I1206 03:41:58.449577 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:41:58 crc kubenswrapper[4802]: I1206 03:41:58.449833 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:41:58 crc kubenswrapper[4802]: E1206 03:41:58.449844 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:41:58 crc kubenswrapper[4802]: I1206 03:41:58.449899 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:41:58 crc kubenswrapper[4802]: I1206 03:41:58.449940 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:41:58 crc kubenswrapper[4802]: E1206 03:41:58.450519 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5ltx2" podUID="8881599e-71e5-4578-b3f9-21f4fa2f57aa" Dec 06 03:41:58 crc kubenswrapper[4802]: E1206 03:41:58.450793 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:41:58 crc kubenswrapper[4802]: E1206 03:41:58.450957 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:41:58 crc kubenswrapper[4802]: I1206 03:41:58.472161 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:58 crc kubenswrapper[4802]: I1206 03:41:58.472237 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:58 crc kubenswrapper[4802]: I1206 03:41:58.472262 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:58 crc kubenswrapper[4802]: I1206 03:41:58.472293 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:58 crc kubenswrapper[4802]: I1206 03:41:58.472318 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:58Z","lastTransitionTime":"2025-12-06T03:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:58 crc kubenswrapper[4802]: I1206 03:41:58.574986 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:58 crc kubenswrapper[4802]: I1206 03:41:58.575054 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:58 crc kubenswrapper[4802]: I1206 03:41:58.575072 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:58 crc kubenswrapper[4802]: I1206 03:41:58.575096 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:58 crc kubenswrapper[4802]: I1206 03:41:58.575114 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:58Z","lastTransitionTime":"2025-12-06T03:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:58 crc kubenswrapper[4802]: I1206 03:41:58.678327 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:58 crc kubenswrapper[4802]: I1206 03:41:58.678393 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:58 crc kubenswrapper[4802]: I1206 03:41:58.678411 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:58 crc kubenswrapper[4802]: I1206 03:41:58.678436 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:58 crc kubenswrapper[4802]: I1206 03:41:58.678456 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:58Z","lastTransitionTime":"2025-12-06T03:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:58 crc kubenswrapper[4802]: I1206 03:41:58.781687 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:58 crc kubenswrapper[4802]: I1206 03:41:58.781817 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:58 crc kubenswrapper[4802]: I1206 03:41:58.781842 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:58 crc kubenswrapper[4802]: I1206 03:41:58.781870 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:58 crc kubenswrapper[4802]: I1206 03:41:58.781887 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:58Z","lastTransitionTime":"2025-12-06T03:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:58 crc kubenswrapper[4802]: I1206 03:41:58.885076 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:58 crc kubenswrapper[4802]: I1206 03:41:58.885148 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:58 crc kubenswrapper[4802]: I1206 03:41:58.885172 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:58 crc kubenswrapper[4802]: I1206 03:41:58.885204 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:58 crc kubenswrapper[4802]: I1206 03:41:58.885228 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:58Z","lastTransitionTime":"2025-12-06T03:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:58 crc kubenswrapper[4802]: I1206 03:41:58.988355 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:58 crc kubenswrapper[4802]: I1206 03:41:58.988457 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:58 crc kubenswrapper[4802]: I1206 03:41:58.988484 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:58 crc kubenswrapper[4802]: I1206 03:41:58.988523 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:58 crc kubenswrapper[4802]: I1206 03:41:58.988550 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:58Z","lastTransitionTime":"2025-12-06T03:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:59 crc kubenswrapper[4802]: I1206 03:41:59.092189 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:59 crc kubenswrapper[4802]: I1206 03:41:59.092243 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:59 crc kubenswrapper[4802]: I1206 03:41:59.092260 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:59 crc kubenswrapper[4802]: I1206 03:41:59.092282 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:59 crc kubenswrapper[4802]: I1206 03:41:59.092299 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:59Z","lastTransitionTime":"2025-12-06T03:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:59 crc kubenswrapper[4802]: I1206 03:41:59.195659 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:59 crc kubenswrapper[4802]: I1206 03:41:59.195722 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:59 crc kubenswrapper[4802]: I1206 03:41:59.195739 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:59 crc kubenswrapper[4802]: I1206 03:41:59.195797 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:59 crc kubenswrapper[4802]: I1206 03:41:59.195847 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:59Z","lastTransitionTime":"2025-12-06T03:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:59 crc kubenswrapper[4802]: I1206 03:41:59.298570 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:59 crc kubenswrapper[4802]: I1206 03:41:59.298662 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:59 crc kubenswrapper[4802]: I1206 03:41:59.298686 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:59 crc kubenswrapper[4802]: I1206 03:41:59.298736 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:59 crc kubenswrapper[4802]: I1206 03:41:59.298813 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:59Z","lastTransitionTime":"2025-12-06T03:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:59 crc kubenswrapper[4802]: I1206 03:41:59.401231 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:59 crc kubenswrapper[4802]: I1206 03:41:59.401316 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:59 crc kubenswrapper[4802]: I1206 03:41:59.401340 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:59 crc kubenswrapper[4802]: I1206 03:41:59.401370 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:59 crc kubenswrapper[4802]: I1206 03:41:59.401395 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:59Z","lastTransitionTime":"2025-12-06T03:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:59 crc kubenswrapper[4802]: I1206 03:41:59.504565 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:59 crc kubenswrapper[4802]: I1206 03:41:59.504630 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:59 crc kubenswrapper[4802]: I1206 03:41:59.504647 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:59 crc kubenswrapper[4802]: I1206 03:41:59.504721 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:59 crc kubenswrapper[4802]: I1206 03:41:59.504741 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:59Z","lastTransitionTime":"2025-12-06T03:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:59 crc kubenswrapper[4802]: I1206 03:41:59.607464 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:59 crc kubenswrapper[4802]: I1206 03:41:59.607524 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:59 crc kubenswrapper[4802]: I1206 03:41:59.607548 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:59 crc kubenswrapper[4802]: I1206 03:41:59.607580 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:59 crc kubenswrapper[4802]: I1206 03:41:59.607603 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:59Z","lastTransitionTime":"2025-12-06T03:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:59 crc kubenswrapper[4802]: I1206 03:41:59.710867 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:59 crc kubenswrapper[4802]: I1206 03:41:59.710925 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:59 crc kubenswrapper[4802]: I1206 03:41:59.710942 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:59 crc kubenswrapper[4802]: I1206 03:41:59.710965 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:59 crc kubenswrapper[4802]: I1206 03:41:59.710985 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:59Z","lastTransitionTime":"2025-12-06T03:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:59 crc kubenswrapper[4802]: I1206 03:41:59.814151 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:59 crc kubenswrapper[4802]: I1206 03:41:59.814203 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:59 crc kubenswrapper[4802]: I1206 03:41:59.814219 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:59 crc kubenswrapper[4802]: I1206 03:41:59.814243 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:59 crc kubenswrapper[4802]: I1206 03:41:59.814261 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:59Z","lastTransitionTime":"2025-12-06T03:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:41:59 crc kubenswrapper[4802]: I1206 03:41:59.918155 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:41:59 crc kubenswrapper[4802]: I1206 03:41:59.918224 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:41:59 crc kubenswrapper[4802]: I1206 03:41:59.918248 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:41:59 crc kubenswrapper[4802]: I1206 03:41:59.918292 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:41:59 crc kubenswrapper[4802]: I1206 03:41:59.918314 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:41:59Z","lastTransitionTime":"2025-12-06T03:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:42:00 crc kubenswrapper[4802]: I1206 03:42:00.021670 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:42:00 crc kubenswrapper[4802]: I1206 03:42:00.021744 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:42:00 crc kubenswrapper[4802]: I1206 03:42:00.021837 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:42:00 crc kubenswrapper[4802]: I1206 03:42:00.021875 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:42:00 crc kubenswrapper[4802]: I1206 03:42:00.021898 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:42:00Z","lastTransitionTime":"2025-12-06T03:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:42:00 crc kubenswrapper[4802]: I1206 03:42:00.124700 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:42:00 crc kubenswrapper[4802]: I1206 03:42:00.124827 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:42:00 crc kubenswrapper[4802]: I1206 03:42:00.124850 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:42:00 crc kubenswrapper[4802]: I1206 03:42:00.124875 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:42:00 crc kubenswrapper[4802]: I1206 03:42:00.124894 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:42:00Z","lastTransitionTime":"2025-12-06T03:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:42:00 crc kubenswrapper[4802]: I1206 03:42:00.228190 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:42:00 crc kubenswrapper[4802]: I1206 03:42:00.228252 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:42:00 crc kubenswrapper[4802]: I1206 03:42:00.228270 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:42:00 crc kubenswrapper[4802]: I1206 03:42:00.228301 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:42:00 crc kubenswrapper[4802]: I1206 03:42:00.228320 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:42:00Z","lastTransitionTime":"2025-12-06T03:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:42:00 crc kubenswrapper[4802]: I1206 03:42:00.331544 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:42:00 crc kubenswrapper[4802]: I1206 03:42:00.331946 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:42:00 crc kubenswrapper[4802]: I1206 03:42:00.331966 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:42:00 crc kubenswrapper[4802]: I1206 03:42:00.331992 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:42:00 crc kubenswrapper[4802]: I1206 03:42:00.332008 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:42:00Z","lastTransitionTime":"2025-12-06T03:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:42:00 crc kubenswrapper[4802]: I1206 03:42:00.435429 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:42:00 crc kubenswrapper[4802]: I1206 03:42:00.435506 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:42:00 crc kubenswrapper[4802]: I1206 03:42:00.435530 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:42:00 crc kubenswrapper[4802]: I1206 03:42:00.435562 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:42:00 crc kubenswrapper[4802]: I1206 03:42:00.435583 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:42:00Z","lastTransitionTime":"2025-12-06T03:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:42:00 crc kubenswrapper[4802]: I1206 03:42:00.449169 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:42:00 crc kubenswrapper[4802]: I1206 03:42:00.449209 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:42:00 crc kubenswrapper[4802]: I1206 03:42:00.449250 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:42:00 crc kubenswrapper[4802]: I1206 03:42:00.449325 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:42:00 crc kubenswrapper[4802]: E1206 03:42:00.449368 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:42:00 crc kubenswrapper[4802]: E1206 03:42:00.449594 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5ltx2" podUID="8881599e-71e5-4578-b3f9-21f4fa2f57aa" Dec 06 03:42:00 crc kubenswrapper[4802]: E1206 03:42:00.449694 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:42:00 crc kubenswrapper[4802]: E1206 03:42:00.449889 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:42:00 crc kubenswrapper[4802]: I1206 03:42:00.538193 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:42:00 crc kubenswrapper[4802]: I1206 03:42:00.538246 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:42:00 crc kubenswrapper[4802]: I1206 03:42:00.538262 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:42:00 crc kubenswrapper[4802]: I1206 03:42:00.538286 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:42:00 crc kubenswrapper[4802]: I1206 03:42:00.538305 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:42:00Z","lastTransitionTime":"2025-12-06T03:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:42:00 crc kubenswrapper[4802]: I1206 03:42:00.640727 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:42:00 crc kubenswrapper[4802]: I1206 03:42:00.640829 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:42:00 crc kubenswrapper[4802]: I1206 03:42:00.640846 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:42:00 crc kubenswrapper[4802]: I1206 03:42:00.640871 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:42:00 crc kubenswrapper[4802]: I1206 03:42:00.640888 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:42:00Z","lastTransitionTime":"2025-12-06T03:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:42:00 crc kubenswrapper[4802]: I1206 03:42:00.743676 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:42:00 crc kubenswrapper[4802]: I1206 03:42:00.743717 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:42:00 crc kubenswrapper[4802]: I1206 03:42:00.743726 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:42:00 crc kubenswrapper[4802]: I1206 03:42:00.743739 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:42:00 crc kubenswrapper[4802]: I1206 03:42:00.743763 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:42:00Z","lastTransitionTime":"2025-12-06T03:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:42:00 crc kubenswrapper[4802]: I1206 03:42:00.846665 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:42:00 crc kubenswrapper[4802]: I1206 03:42:00.846715 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:42:00 crc kubenswrapper[4802]: I1206 03:42:00.846733 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:42:00 crc kubenswrapper[4802]: I1206 03:42:00.846819 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:42:00 crc kubenswrapper[4802]: I1206 03:42:00.846846 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:42:00Z","lastTransitionTime":"2025-12-06T03:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:42:00 crc kubenswrapper[4802]: I1206 03:42:00.949499 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:42:00 crc kubenswrapper[4802]: I1206 03:42:00.949555 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:42:00 crc kubenswrapper[4802]: I1206 03:42:00.949571 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:42:00 crc kubenswrapper[4802]: I1206 03:42:00.949595 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:42:00 crc kubenswrapper[4802]: I1206 03:42:00.949615 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:42:00Z","lastTransitionTime":"2025-12-06T03:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.052516 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.052870 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.052957 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.052988 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.053046 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:42:01Z","lastTransitionTime":"2025-12-06T03:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.105414 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-22rrq_03112a5d-5882-4fac-9ae5-13525ff82fe0/kube-multus/1.log" Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.106107 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-22rrq_03112a5d-5882-4fac-9ae5-13525ff82fe0/kube-multus/0.log" Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.106192 4802 generic.go:334] "Generic (PLEG): container finished" podID="03112a5d-5882-4fac-9ae5-13525ff82fe0" containerID="831dab65557627cbce0d055459052dae5b594f33b66858367ccd034ff6811924" exitCode=1 Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.106243 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-22rrq" event={"ID":"03112a5d-5882-4fac-9ae5-13525ff82fe0","Type":"ContainerDied","Data":"831dab65557627cbce0d055459052dae5b594f33b66858367ccd034ff6811924"} Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.106337 4802 scope.go:117] "RemoveContainer" containerID="c0205b089b12c42688203bb7ec0f27dc5f9a577dabdb446b811f184046f85dc3" Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.107135 4802 scope.go:117] "RemoveContainer" containerID="831dab65557627cbce0d055459052dae5b594f33b66858367ccd034ff6811924" Dec 06 03:42:01 crc kubenswrapper[4802]: E1206 03:42:01.107405 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-22rrq_openshift-multus(03112a5d-5882-4fac-9ae5-13525ff82fe0)\"" pod="openshift-multus/multus-22rrq" podUID="03112a5d-5882-4fac-9ae5-13525ff82fe0" Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.156968 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.157027 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.157044 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.157068 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.157086 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:42:01Z","lastTransitionTime":"2025-12-06T03:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.163024 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podStartSLOduration=95.162999932 podStartE2EDuration="1m35.162999932s" podCreationTimestamp="2025-12-06 03:40:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:42:01.1419903 +0000 UTC m=+114.013899512" watchObservedRunningTime="2025-12-06 03:42:01.162999932 +0000 UTC m=+114.034909114" Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.183681 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-kj7hf" podStartSLOduration=95.183647463 podStartE2EDuration="1m35.183647463s" podCreationTimestamp="2025-12-06 03:40:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:42:01.162905979 +0000 UTC m=+114.034815161" watchObservedRunningTime="2025-12-06 03:42:01.183647463 +0000 UTC m=+114.055556655" Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.260769 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.260820 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.260832 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.260877 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.260890 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:42:01Z","lastTransitionTime":"2025-12-06T03:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.304513 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=41.304485983 podStartE2EDuration="41.304485983s" podCreationTimestamp="2025-12-06 03:41:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:42:01.284787836 +0000 UTC m=+114.156697048" watchObservedRunningTime="2025-12-06 03:42:01.304485983 +0000 UTC m=+114.176395175" Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.305131 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=95.30512182 podStartE2EDuration="1m35.30512182s" podCreationTimestamp="2025-12-06 03:40:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:42:01.303514717 +0000 UTC m=+114.175423909" watchObservedRunningTime="2025-12-06 03:42:01.30512182 +0000 UTC m=+114.177031012" Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.363168 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.363213 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.363224 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.363243 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.363256 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:42:01Z","lastTransitionTime":"2025-12-06T03:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.365026 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.365075 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.365134 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.365158 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.365175 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:42:01Z","lastTransitionTime":"2025-12-06T03:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.391603 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-xtw2k" podStartSLOduration=95.391574031 podStartE2EDuration="1m35.391574031s" podCreationTimestamp="2025-12-06 03:40:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:42:01.391395915 +0000 UTC m=+114.263305087" watchObservedRunningTime="2025-12-06 03:42:01.391574031 +0000 UTC m=+114.263483243" Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.414826 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=59.414797251 podStartE2EDuration="59.414797251s" podCreationTimestamp="2025-12-06 03:41:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:42:01.414073702 +0000 UTC m=+114.285982874" watchObservedRunningTime="2025-12-06 03:42:01.414797251 +0000 UTC m=+114.286706443" Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.428794 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-b2knv" podStartSLOduration=96.428742504 podStartE2EDuration="1m36.428742504s" podCreationTimestamp="2025-12-06 03:40:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:42:01.428479637 +0000 UTC m=+114.300388809" watchObservedRunningTime="2025-12-06 03:42:01.428742504 +0000 UTC m=+114.300651686" Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.430445 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-vhfrf"] Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.431134 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vhfrf" Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.433111 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.434136 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.434268 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.435004 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.464615 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/8956b0ad-1f3b-42e5-8bfc-3225d63953fe-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-vhfrf\" (UID: \"8956b0ad-1f3b-42e5-8bfc-3225d63953fe\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vhfrf" Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.464658 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8956b0ad-1f3b-42e5-8bfc-3225d63953fe-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-vhfrf\" (UID: \"8956b0ad-1f3b-42e5-8bfc-3225d63953fe\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vhfrf" Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.464683 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8956b0ad-1f3b-42e5-8bfc-3225d63953fe-service-ca\") pod \"cluster-version-operator-5c965bbfc6-vhfrf\" (UID: \"8956b0ad-1f3b-42e5-8bfc-3225d63953fe\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vhfrf" Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.464718 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8956b0ad-1f3b-42e5-8bfc-3225d63953fe-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-vhfrf\" (UID: \"8956b0ad-1f3b-42e5-8bfc-3225d63953fe\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vhfrf" Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.464780 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/8956b0ad-1f3b-42e5-8bfc-3225d63953fe-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-vhfrf\" (UID: \"8956b0ad-1f3b-42e5-8bfc-3225d63953fe\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vhfrf" Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.468050 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ngspk" podStartSLOduration=94.468032754 podStartE2EDuration="1m34.468032754s" podCreationTimestamp="2025-12-06 03:40:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:42:01.467584552 +0000 UTC m=+114.339493704" watchObservedRunningTime="2025-12-06 03:42:01.468032754 +0000 UTC m=+114.339941916" Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.505104 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=12.505085214 podStartE2EDuration="12.505085214s" podCreationTimestamp="2025-12-06 03:41:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:42:01.504279133 +0000 UTC m=+114.376188315" watchObservedRunningTime="2025-12-06 03:42:01.505085214 +0000 UTC m=+114.376994366" Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.522849 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=94.522832448 podStartE2EDuration="1m34.522832448s" podCreationTimestamp="2025-12-06 03:40:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:42:01.52176838 +0000 UTC m=+114.393677542" watchObservedRunningTime="2025-12-06 03:42:01.522832448 +0000 UTC m=+114.394741610" Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.566365 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/8956b0ad-1f3b-42e5-8bfc-3225d63953fe-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-vhfrf\" (UID: \"8956b0ad-1f3b-42e5-8bfc-3225d63953fe\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vhfrf" Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.566444 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/8956b0ad-1f3b-42e5-8bfc-3225d63953fe-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-vhfrf\" (UID: \"8956b0ad-1f3b-42e5-8bfc-3225d63953fe\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vhfrf" Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.566468 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8956b0ad-1f3b-42e5-8bfc-3225d63953fe-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-vhfrf\" (UID: \"8956b0ad-1f3b-42e5-8bfc-3225d63953fe\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vhfrf" Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.566489 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8956b0ad-1f3b-42e5-8bfc-3225d63953fe-service-ca\") pod \"cluster-version-operator-5c965bbfc6-vhfrf\" (UID: \"8956b0ad-1f3b-42e5-8bfc-3225d63953fe\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vhfrf" Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.566521 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8956b0ad-1f3b-42e5-8bfc-3225d63953fe-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-vhfrf\" (UID: \"8956b0ad-1f3b-42e5-8bfc-3225d63953fe\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vhfrf" Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.566894 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/8956b0ad-1f3b-42e5-8bfc-3225d63953fe-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-vhfrf\" (UID: \"8956b0ad-1f3b-42e5-8bfc-3225d63953fe\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vhfrf" Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.566952 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/8956b0ad-1f3b-42e5-8bfc-3225d63953fe-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-vhfrf\" (UID: \"8956b0ad-1f3b-42e5-8bfc-3225d63953fe\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vhfrf" Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.567734 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8956b0ad-1f3b-42e5-8bfc-3225d63953fe-service-ca\") pod \"cluster-version-operator-5c965bbfc6-vhfrf\" (UID: \"8956b0ad-1f3b-42e5-8bfc-3225d63953fe\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vhfrf" Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.576611 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8956b0ad-1f3b-42e5-8bfc-3225d63953fe-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-vhfrf\" (UID: \"8956b0ad-1f3b-42e5-8bfc-3225d63953fe\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vhfrf" Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.582512 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8956b0ad-1f3b-42e5-8bfc-3225d63953fe-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-vhfrf\" (UID: \"8956b0ad-1f3b-42e5-8bfc-3225d63953fe\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vhfrf" Dec 06 03:42:01 crc kubenswrapper[4802]: I1206 03:42:01.750891 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vhfrf" Dec 06 03:42:02 crc kubenswrapper[4802]: I1206 03:42:02.115182 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-22rrq_03112a5d-5882-4fac-9ae5-13525ff82fe0/kube-multus/1.log" Dec 06 03:42:02 crc kubenswrapper[4802]: I1206 03:42:02.120365 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vhfrf" event={"ID":"8956b0ad-1f3b-42e5-8bfc-3225d63953fe","Type":"ContainerStarted","Data":"c4e018da82c41be82f2de95cd6af2a856d66d3d4a5d225900c1acaaa16247753"} Dec 06 03:42:02 crc kubenswrapper[4802]: I1206 03:42:02.120456 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vhfrf" event={"ID":"8956b0ad-1f3b-42e5-8bfc-3225d63953fe","Type":"ContainerStarted","Data":"cf0d9216e1d610a59f6dc5ec52f093f2af37db10c570336ff7fae1fbbc872368"} Dec 06 03:42:02 crc kubenswrapper[4802]: I1206 03:42:02.144156 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vhfrf" podStartSLOduration=96.144126703 podStartE2EDuration="1m36.144126703s" podCreationTimestamp="2025-12-06 03:40:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:42:02.142606443 +0000 UTC m=+115.014515625" watchObservedRunningTime="2025-12-06 03:42:02.144126703 +0000 UTC m=+115.016035895" Dec 06 03:42:02 crc kubenswrapper[4802]: I1206 03:42:02.450022 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:42:02 crc kubenswrapper[4802]: I1206 03:42:02.450163 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:42:02 crc kubenswrapper[4802]: I1206 03:42:02.450164 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:42:02 crc kubenswrapper[4802]: I1206 03:42:02.450258 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:42:02 crc kubenswrapper[4802]: E1206 03:42:02.450986 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:42:02 crc kubenswrapper[4802]: E1206 03:42:02.450542 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:42:02 crc kubenswrapper[4802]: E1206 03:42:02.450722 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:42:02 crc kubenswrapper[4802]: E1206 03:42:02.450445 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5ltx2" podUID="8881599e-71e5-4578-b3f9-21f4fa2f57aa" Dec 06 03:42:04 crc kubenswrapper[4802]: I1206 03:42:04.449842 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:42:04 crc kubenswrapper[4802]: I1206 03:42:04.449907 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:42:04 crc kubenswrapper[4802]: I1206 03:42:04.449883 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:42:04 crc kubenswrapper[4802]: I1206 03:42:04.449861 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:42:04 crc kubenswrapper[4802]: E1206 03:42:04.450038 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5ltx2" podUID="8881599e-71e5-4578-b3f9-21f4fa2f57aa" Dec 06 03:42:04 crc kubenswrapper[4802]: E1206 03:42:04.450171 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:42:04 crc kubenswrapper[4802]: E1206 03:42:04.450333 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:42:04 crc kubenswrapper[4802]: E1206 03:42:04.450558 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:42:06 crc kubenswrapper[4802]: I1206 03:42:06.449365 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:42:06 crc kubenswrapper[4802]: I1206 03:42:06.449462 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:42:06 crc kubenswrapper[4802]: I1206 03:42:06.449482 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:42:06 crc kubenswrapper[4802]: E1206 03:42:06.449672 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5ltx2" podUID="8881599e-71e5-4578-b3f9-21f4fa2f57aa" Dec 06 03:42:06 crc kubenswrapper[4802]: I1206 03:42:06.449719 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:42:06 crc kubenswrapper[4802]: E1206 03:42:06.449936 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:42:06 crc kubenswrapper[4802]: E1206 03:42:06.450095 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:42:06 crc kubenswrapper[4802]: E1206 03:42:06.450271 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:42:07 crc kubenswrapper[4802]: E1206 03:42:07.449121 4802 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 06 03:42:07 crc kubenswrapper[4802]: I1206 03:42:07.452839 4802 scope.go:117] "RemoveContainer" containerID="bf554aaba0e773f6a6c1f29bb1819b7ae25b2b2db255c4bd0a16c512f314e9cf" Dec 06 03:42:07 crc kubenswrapper[4802]: E1206 03:42:07.543187 4802 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 06 03:42:08 crc kubenswrapper[4802]: I1206 03:42:08.143834 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pvm6q_eafce4e0-e7fb-4877-b0ab-3283829ba1ef/ovnkube-controller/3.log" Dec 06 03:42:08 crc kubenswrapper[4802]: I1206 03:42:08.146520 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" event={"ID":"eafce4e0-e7fb-4877-b0ab-3283829ba1ef","Type":"ContainerStarted","Data":"4180d357d73e8243e980d1bcaef03f37d0eb15e93c2e8a73d9c204909ff5ac54"} Dec 06 03:42:08 crc kubenswrapper[4802]: I1206 03:42:08.147237 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:42:08 crc kubenswrapper[4802]: I1206 03:42:08.191286 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" podStartSLOduration=102.191268646 podStartE2EDuration="1m42.191268646s" podCreationTimestamp="2025-12-06 03:40:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:42:08.188272996 +0000 UTC m=+121.060182158" watchObservedRunningTime="2025-12-06 03:42:08.191268646 +0000 UTC m=+121.063177808" Dec 06 03:42:08 crc kubenswrapper[4802]: I1206 03:42:08.409728 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-5ltx2"] Dec 06 03:42:08 crc kubenswrapper[4802]: I1206 03:42:08.409908 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:42:08 crc kubenswrapper[4802]: E1206 03:42:08.410046 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5ltx2" podUID="8881599e-71e5-4578-b3f9-21f4fa2f57aa" Dec 06 03:42:08 crc kubenswrapper[4802]: I1206 03:42:08.449968 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:42:08 crc kubenswrapper[4802]: I1206 03:42:08.450021 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:42:08 crc kubenswrapper[4802]: I1206 03:42:08.449985 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:42:08 crc kubenswrapper[4802]: E1206 03:42:08.450096 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:42:08 crc kubenswrapper[4802]: E1206 03:42:08.450154 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:42:08 crc kubenswrapper[4802]: E1206 03:42:08.450210 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:42:10 crc kubenswrapper[4802]: I1206 03:42:10.449992 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:42:10 crc kubenswrapper[4802]: I1206 03:42:10.450100 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:42:10 crc kubenswrapper[4802]: I1206 03:42:10.450057 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:42:10 crc kubenswrapper[4802]: I1206 03:42:10.450267 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:42:10 crc kubenswrapper[4802]: E1206 03:42:10.450253 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:42:10 crc kubenswrapper[4802]: E1206 03:42:10.450367 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:42:10 crc kubenswrapper[4802]: E1206 03:42:10.450474 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5ltx2" podUID="8881599e-71e5-4578-b3f9-21f4fa2f57aa" Dec 06 03:42:10 crc kubenswrapper[4802]: E1206 03:42:10.450735 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:42:12 crc kubenswrapper[4802]: I1206 03:42:12.449626 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:42:12 crc kubenswrapper[4802]: E1206 03:42:12.450111 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:42:12 crc kubenswrapper[4802]: I1206 03:42:12.449680 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:42:12 crc kubenswrapper[4802]: E1206 03:42:12.450198 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:42:12 crc kubenswrapper[4802]: I1206 03:42:12.449633 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:42:12 crc kubenswrapper[4802]: I1206 03:42:12.449702 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:42:12 crc kubenswrapper[4802]: E1206 03:42:12.450264 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5ltx2" podUID="8881599e-71e5-4578-b3f9-21f4fa2f57aa" Dec 06 03:42:12 crc kubenswrapper[4802]: E1206 03:42:12.450389 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:42:12 crc kubenswrapper[4802]: E1206 03:42:12.544677 4802 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 06 03:42:14 crc kubenswrapper[4802]: I1206 03:42:14.450158 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:42:14 crc kubenswrapper[4802]: I1206 03:42:14.450217 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:42:14 crc kubenswrapper[4802]: I1206 03:42:14.450246 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:42:14 crc kubenswrapper[4802]: I1206 03:42:14.450188 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:42:14 crc kubenswrapper[4802]: E1206 03:42:14.451148 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5ltx2" podUID="8881599e-71e5-4578-b3f9-21f4fa2f57aa" Dec 06 03:42:14 crc kubenswrapper[4802]: E1206 03:42:14.451303 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:42:14 crc kubenswrapper[4802]: E1206 03:42:14.451432 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:42:14 crc kubenswrapper[4802]: E1206 03:42:14.451634 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:42:16 crc kubenswrapper[4802]: I1206 03:42:16.450019 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:42:16 crc kubenswrapper[4802]: I1206 03:42:16.450123 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:42:16 crc kubenswrapper[4802]: E1206 03:42:16.450319 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5ltx2" podUID="8881599e-71e5-4578-b3f9-21f4fa2f57aa" Dec 06 03:42:16 crc kubenswrapper[4802]: I1206 03:42:16.450356 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:42:16 crc kubenswrapper[4802]: I1206 03:42:16.450453 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:42:16 crc kubenswrapper[4802]: E1206 03:42:16.450839 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:42:16 crc kubenswrapper[4802]: I1206 03:42:16.450914 4802 scope.go:117] "RemoveContainer" containerID="831dab65557627cbce0d055459052dae5b594f33b66858367ccd034ff6811924" Dec 06 03:42:16 crc kubenswrapper[4802]: E1206 03:42:16.450993 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:42:16 crc kubenswrapper[4802]: E1206 03:42:16.451176 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:42:17 crc kubenswrapper[4802]: I1206 03:42:17.178702 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-22rrq_03112a5d-5882-4fac-9ae5-13525ff82fe0/kube-multus/1.log" Dec 06 03:42:17 crc kubenswrapper[4802]: I1206 03:42:17.179064 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-22rrq" event={"ID":"03112a5d-5882-4fac-9ae5-13525ff82fe0","Type":"ContainerStarted","Data":"a85adeebc0758de20abbd8fcca0cea561a182955a07b43139ba2e29f72134676"} Dec 06 03:42:17 crc kubenswrapper[4802]: I1206 03:42:17.200596 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-22rrq" podStartSLOduration=111.2005775 podStartE2EDuration="1m51.2005775s" podCreationTimestamp="2025-12-06 03:40:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:42:17.200096977 +0000 UTC m=+130.072006129" watchObservedRunningTime="2025-12-06 03:42:17.2005775 +0000 UTC m=+130.072486652" Dec 06 03:42:17 crc kubenswrapper[4802]: E1206 03:42:17.545318 4802 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 06 03:42:18 crc kubenswrapper[4802]: I1206 03:42:18.450184 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:42:18 crc kubenswrapper[4802]: I1206 03:42:18.450274 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:42:18 crc kubenswrapper[4802]: I1206 03:42:18.450336 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:42:18 crc kubenswrapper[4802]: I1206 03:42:18.450274 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:42:18 crc kubenswrapper[4802]: E1206 03:42:18.450531 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:42:18 crc kubenswrapper[4802]: E1206 03:42:18.450631 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:42:18 crc kubenswrapper[4802]: E1206 03:42:18.450867 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5ltx2" podUID="8881599e-71e5-4578-b3f9-21f4fa2f57aa" Dec 06 03:42:18 crc kubenswrapper[4802]: E1206 03:42:18.450988 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:42:20 crc kubenswrapper[4802]: I1206 03:42:20.449543 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:42:20 crc kubenswrapper[4802]: I1206 03:42:20.449673 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:42:20 crc kubenswrapper[4802]: I1206 03:42:20.449604 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:42:20 crc kubenswrapper[4802]: I1206 03:42:20.449583 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:42:20 crc kubenswrapper[4802]: E1206 03:42:20.449895 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5ltx2" podUID="8881599e-71e5-4578-b3f9-21f4fa2f57aa" Dec 06 03:42:20 crc kubenswrapper[4802]: E1206 03:42:20.450211 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:42:20 crc kubenswrapper[4802]: E1206 03:42:20.450319 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:42:20 crc kubenswrapper[4802]: E1206 03:42:20.450356 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:42:22 crc kubenswrapper[4802]: I1206 03:42:22.449158 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:42:22 crc kubenswrapper[4802]: I1206 03:42:22.449228 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:42:22 crc kubenswrapper[4802]: I1206 03:42:22.449259 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:42:22 crc kubenswrapper[4802]: I1206 03:42:22.449288 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:42:22 crc kubenswrapper[4802]: E1206 03:42:22.449468 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5ltx2" podUID="8881599e-71e5-4578-b3f9-21f4fa2f57aa" Dec 06 03:42:22 crc kubenswrapper[4802]: E1206 03:42:22.449625 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:42:22 crc kubenswrapper[4802]: E1206 03:42:22.449826 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:42:22 crc kubenswrapper[4802]: E1206 03:42:22.450051 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:42:24 crc kubenswrapper[4802]: I1206 03:42:24.449046 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:42:24 crc kubenswrapper[4802]: I1206 03:42:24.449101 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:42:24 crc kubenswrapper[4802]: I1206 03:42:24.449282 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:42:24 crc kubenswrapper[4802]: I1206 03:42:24.449349 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:42:24 crc kubenswrapper[4802]: I1206 03:42:24.452489 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 06 03:42:24 crc kubenswrapper[4802]: I1206 03:42:24.452607 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 06 03:42:24 crc kubenswrapper[4802]: I1206 03:42:24.452786 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 06 03:42:24 crc kubenswrapper[4802]: I1206 03:42:24.452946 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 06 03:42:24 crc kubenswrapper[4802]: I1206 03:42:24.453125 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 06 03:42:24 crc kubenswrapper[4802]: I1206 03:42:24.453632 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.355946 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.406456 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-7ckfj"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.407200 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-7ckfj" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.408887 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-4lqzs"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.409882 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-4lqzs" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.417133 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ghcrj"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.417876 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ghcrj" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.420031 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.420286 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.420934 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.421280 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.421634 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.421907 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.422085 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-rqs9q"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.422607 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-rqs9q" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.423059 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-9v855"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.423575 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9v855" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.425198 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-8br44"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.425609 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8br44" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.426723 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.426905 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.427027 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.427142 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.427253 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.429807 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-fnpz5"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.430322 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fnpz5" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.431383 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.431571 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.431710 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.431894 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.432423 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.434127 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.434338 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.434776 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.434951 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.435947 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-8xz4m"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.436200 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.436483 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-7cg8t"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.436800 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-8xz4m" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.436805 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-7cg8t" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.439122 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jv24g"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.439693 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jv24g" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.439829 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.439987 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.440304 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.440359 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.443495 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-thckl"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.444062 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-9s5r8"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.444563 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-cgpdh"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.444815 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-thckl" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.444935 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-9s5r8" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.444999 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-cgpdh" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.445465 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fqdzj"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.445929 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fqdzj" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.450921 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-rqs9q\" (UID: \"263721d5-fcca-411a-876e-0acfe0863d2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-rqs9q" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.450977 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-rqs9q\" (UID: \"263721d5-fcca-411a-876e-0acfe0863d2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-rqs9q" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.451014 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3f3f6253-533f-4d89-ac10-f26eca1ac390-trusted-ca\") pod \"console-operator-58897d9998-4lqzs\" (UID: \"3f3f6253-533f-4d89-ac10-f26eca1ac390\") " pod="openshift-console-operator/console-operator-58897d9998-4lqzs" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.451065 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/2362780e-93d1-4f97-bda0-138ed5180bb0-oauth-serving-cert\") pod \"console-f9d7485db-7ckfj\" (UID: \"2362780e-93d1-4f97-bda0-138ed5180bb0\") " pod="openshift-console/console-f9d7485db-7ckfj" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.451094 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-992nr\" (UniqueName: \"kubernetes.io/projected/3f3f6253-533f-4d89-ac10-f26eca1ac390-kube-api-access-992nr\") pod \"console-operator-58897d9998-4lqzs\" (UID: \"3f3f6253-533f-4d89-ac10-f26eca1ac390\") " pod="openshift-console-operator/console-operator-58897d9998-4lqzs" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.451122 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-rqs9q\" (UID: \"263721d5-fcca-411a-876e-0acfe0863d2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-rqs9q" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.451150 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/bd64a08a-6920-4a0f-b299-aa5a009236a1-encryption-config\") pod \"apiserver-7bbb656c7d-8br44\" (UID: \"bd64a08a-6920-4a0f-b299-aa5a009236a1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8br44" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.451179 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/bd64a08a-6920-4a0f-b299-aa5a009236a1-etcd-client\") pod \"apiserver-7bbb656c7d-8br44\" (UID: \"bd64a08a-6920-4a0f-b299-aa5a009236a1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8br44" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.451221 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-rqs9q\" (UID: \"263721d5-fcca-411a-876e-0acfe0863d2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-rqs9q" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.451227 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-t56zb"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.451250 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2362780e-93d1-4f97-bda0-138ed5180bb0-service-ca\") pod \"console-f9d7485db-7ckfj\" (UID: \"2362780e-93d1-4f97-bda0-138ed5180bb0\") " pod="openshift-console/console-f9d7485db-7ckfj" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.451279 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-rqs9q\" (UID: \"263721d5-fcca-411a-876e-0acfe0863d2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-rqs9q" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.451310 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-rqs9q\" (UID: \"263721d5-fcca-411a-876e-0acfe0863d2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-rqs9q" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.451340 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/2362780e-93d1-4f97-bda0-138ed5180bb0-console-serving-cert\") pod \"console-f9d7485db-7ckfj\" (UID: \"2362780e-93d1-4f97-bda0-138ed5180bb0\") " pod="openshift-console/console-f9d7485db-7ckfj" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.451372 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/dc4fc98a-badd-46dc-b700-6b01aff112b4-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-ghcrj\" (UID: \"dc4fc98a-badd-46dc-b700-6b01aff112b4\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ghcrj" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.451400 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/bd64a08a-6920-4a0f-b299-aa5a009236a1-audit-policies\") pod \"apiserver-7bbb656c7d-8br44\" (UID: \"bd64a08a-6920-4a0f-b299-aa5a009236a1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8br44" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.451430 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db213df1-bfa2-4edd-a3fc-246e35e20585-config\") pod \"route-controller-manager-6576b87f9c-9v855\" (UID: \"db213df1-bfa2-4edd-a3fc-246e35e20585\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9v855" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.451458 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bd64a08a-6920-4a0f-b299-aa5a009236a1-serving-cert\") pod \"apiserver-7bbb656c7d-8br44\" (UID: \"bd64a08a-6920-4a0f-b299-aa5a009236a1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8br44" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.451499 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-rqs9q\" (UID: \"263721d5-fcca-411a-876e-0acfe0863d2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-rqs9q" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.451527 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpqvr\" (UniqueName: \"kubernetes.io/projected/db213df1-bfa2-4edd-a3fc-246e35e20585-kube-api-access-fpqvr\") pod \"route-controller-manager-6576b87f9c-9v855\" (UID: \"db213df1-bfa2-4edd-a3fc-246e35e20585\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9v855" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.451561 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/db213df1-bfa2-4edd-a3fc-246e35e20585-client-ca\") pod \"route-controller-manager-6576b87f9c-9v855\" (UID: \"db213df1-bfa2-4edd-a3fc-246e35e20585\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9v855" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.451591 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/263721d5-fcca-411a-876e-0acfe0863d2d-audit-policies\") pod \"oauth-openshift-558db77b4-rqs9q\" (UID: \"263721d5-fcca-411a-876e-0acfe0863d2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-rqs9q" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.451621 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-rqs9q\" (UID: \"263721d5-fcca-411a-876e-0acfe0863d2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-rqs9q" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.451653 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbrbz\" (UniqueName: \"kubernetes.io/projected/2362780e-93d1-4f97-bda0-138ed5180bb0-kube-api-access-qbrbz\") pod \"console-f9d7485db-7ckfj\" (UID: \"2362780e-93d1-4f97-bda0-138ed5180bb0\") " pod="openshift-console/console-f9d7485db-7ckfj" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.451690 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/263721d5-fcca-411a-876e-0acfe0863d2d-audit-dir\") pod \"oauth-openshift-558db77b4-rqs9q\" (UID: \"263721d5-fcca-411a-876e-0acfe0863d2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-rqs9q" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.451725 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-rqs9q\" (UID: \"263721d5-fcca-411a-876e-0acfe0863d2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-rqs9q" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.451782 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/2362780e-93d1-4f97-bda0-138ed5180bb0-console-config\") pod \"console-f9d7485db-7ckfj\" (UID: \"2362780e-93d1-4f97-bda0-138ed5180bb0\") " pod="openshift-console/console-f9d7485db-7ckfj" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.451818 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/bd64a08a-6920-4a0f-b299-aa5a009236a1-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-8br44\" (UID: \"bd64a08a-6920-4a0f-b299-aa5a009236a1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8br44" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.451848 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6w6b\" (UniqueName: \"kubernetes.io/projected/bd64a08a-6920-4a0f-b299-aa5a009236a1-kube-api-access-g6w6b\") pod \"apiserver-7bbb656c7d-8br44\" (UID: \"bd64a08a-6920-4a0f-b299-aa5a009236a1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8br44" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.451882 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f3f6253-533f-4d89-ac10-f26eca1ac390-serving-cert\") pod \"console-operator-58897d9998-4lqzs\" (UID: \"3f3f6253-533f-4d89-ac10-f26eca1ac390\") " pod="openshift-console-operator/console-operator-58897d9998-4lqzs" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.451928 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfqp4\" (UniqueName: \"kubernetes.io/projected/dc4fc98a-badd-46dc-b700-6b01aff112b4-kube-api-access-lfqp4\") pod \"cluster-samples-operator-665b6dd947-ghcrj\" (UID: \"dc4fc98a-badd-46dc-b700-6b01aff112b4\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ghcrj" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.451957 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/db213df1-bfa2-4edd-a3fc-246e35e20585-serving-cert\") pod \"route-controller-manager-6576b87f9c-9v855\" (UID: \"db213df1-bfa2-4edd-a3fc-246e35e20585\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9v855" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.451991 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-rqs9q\" (UID: \"263721d5-fcca-411a-876e-0acfe0863d2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-rqs9q" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.452019 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f3f6253-533f-4d89-ac10-f26eca1ac390-config\") pod \"console-operator-58897d9998-4lqzs\" (UID: \"3f3f6253-533f-4d89-ac10-f26eca1ac390\") " pod="openshift-console-operator/console-operator-58897d9998-4lqzs" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.452053 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-rqs9q\" (UID: \"263721d5-fcca-411a-876e-0acfe0863d2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-rqs9q" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.452082 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgrht\" (UniqueName: \"kubernetes.io/projected/263721d5-fcca-411a-876e-0acfe0863d2d-kube-api-access-bgrht\") pod \"oauth-openshift-558db77b4-rqs9q\" (UID: \"263721d5-fcca-411a-876e-0acfe0863d2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-rqs9q" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.452113 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/2362780e-93d1-4f97-bda0-138ed5180bb0-console-oauth-config\") pod \"console-f9d7485db-7ckfj\" (UID: \"2362780e-93d1-4f97-bda0-138ed5180bb0\") " pod="openshift-console/console-f9d7485db-7ckfj" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.452143 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2362780e-93d1-4f97-bda0-138ed5180bb0-trusted-ca-bundle\") pod \"console-f9d7485db-7ckfj\" (UID: \"2362780e-93d1-4f97-bda0-138ed5180bb0\") " pod="openshift-console/console-f9d7485db-7ckfj" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.452178 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bd64a08a-6920-4a0f-b299-aa5a009236a1-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-8br44\" (UID: \"bd64a08a-6920-4a0f-b299-aa5a009236a1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8br44" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.452210 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/bd64a08a-6920-4a0f-b299-aa5a009236a1-audit-dir\") pod \"apiserver-7bbb656c7d-8br44\" (UID: \"bd64a08a-6920-4a0f-b299-aa5a009236a1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8br44" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.452119 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-t56zb" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.455149 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.455379 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.455446 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.455663 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.455522 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.455561 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.455380 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.455933 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.456222 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.456450 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.474863 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.478144 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.487008 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.487218 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.488147 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.488362 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.488511 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.488676 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.488736 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.488900 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.489056 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.489387 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.489774 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.490108 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.489784 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.490496 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.490588 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.490764 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.490888 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.491116 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.492065 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.492245 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.492378 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.492524 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-gkpck"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.492598 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.492731 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.493498 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-rrhqx"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.493997 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-rrhqx" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.494259 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-gkpck" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.494446 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-s5b86"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.494965 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-s5b86" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.495180 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.495418 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.495557 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.495643 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.495884 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.495977 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.496131 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.496323 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.496646 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.496890 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.497262 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.497399 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.497567 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.497665 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.497840 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.497965 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.498051 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.498164 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.498256 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.498422 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.498847 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.499087 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.499200 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.499325 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.499436 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.499536 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.499618 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.499877 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.503440 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.503874 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.504572 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.505161 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.506603 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5rwsv"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.507082 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5rwsv" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.507366 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x6zsh"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.507655 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x6zsh" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.508801 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-qfdsg"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.509387 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qfdsg" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.510124 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.511803 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-9rvnv"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.512268 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.512410 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9rvnv" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.514095 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ppzvl"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.514612 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ppzvl" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.514712 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6d29n"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.514736 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.515047 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6d29n" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.515529 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.516211 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-jwc8v"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.523829 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.530654 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.531864 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.544214 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9bfz8"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.544691 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9bfz8" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.544814 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jwc8v" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.544835 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.545080 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-ntmms"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.545728 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-ntmms" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.545876 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4gkmd"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.546377 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4gkmd" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.547618 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wbjzw"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.548103 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gmvnz"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.548446 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gmvnz" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.548672 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wbjzw" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.549533 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.550046 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.551223 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.552417 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2nslt"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.553182 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2nslt" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.553485 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.553769 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mpnf2"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.553842 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/b8455642-8313-4539-9729-fe658ef0e4d6-image-import-ca\") pod \"apiserver-76f77b778f-8xz4m\" (UID: \"b8455642-8313-4539-9729-fe658ef0e4d6\") " pod="openshift-apiserver/apiserver-76f77b778f-8xz4m" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.553882 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e4f6934a-db7e-4031-baa8-2f41dc8c2c18-bound-sa-token\") pod \"ingress-operator-5b745b69d9-qfdsg\" (UID: \"e4f6934a-db7e-4031-baa8-2f41dc8c2c18\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qfdsg" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.553910 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e4f6934a-db7e-4031-baa8-2f41dc8c2c18-metrics-tls\") pod \"ingress-operator-5b745b69d9-qfdsg\" (UID: \"e4f6934a-db7e-4031-baa8-2f41dc8c2c18\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qfdsg" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.553934 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1051665a-7918-487f-ad08-ff6d53d113ca-etcd-client\") pod \"etcd-operator-b45778765-t56zb\" (UID: \"1051665a-7918-487f-ad08-ff6d53d113ca\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t56zb" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.553958 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9454f4b6-e1f9-4dc1-953c-442dbd4088ad-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-x6zsh\" (UID: \"9454f4b6-e1f9-4dc1-953c-442dbd4088ad\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x6zsh" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.553985 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bd64a08a-6920-4a0f-b299-aa5a009236a1-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-8br44\" (UID: \"bd64a08a-6920-4a0f-b299-aa5a009236a1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8br44" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.554010 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwgtl\" (UniqueName: \"kubernetes.io/projected/f57cc683-84dc-4538-96fe-9a198a5b008f-kube-api-access-zwgtl\") pod \"openshift-config-operator-7777fb866f-s5b86\" (UID: \"f57cc683-84dc-4538-96fe-9a198a5b008f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-s5b86" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.554038 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/bd64a08a-6920-4a0f-b299-aa5a009236a1-audit-dir\") pod \"apiserver-7bbb656c7d-8br44\" (UID: \"bd64a08a-6920-4a0f-b299-aa5a009236a1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8br44" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.554087 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/bd64a08a-6920-4a0f-b299-aa5a009236a1-audit-dir\") pod \"apiserver-7bbb656c7d-8br44\" (UID: \"bd64a08a-6920-4a0f-b299-aa5a009236a1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8br44" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.554102 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0fbd23d0-a6b6-4a4e-b418-07fc38e5f497-service-ca-bundle\") pod \"router-default-5444994796-rrhqx\" (UID: \"0fbd23d0-a6b6-4a4e-b418-07fc38e5f497\") " pod="openshift-ingress/router-default-5444994796-rrhqx" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.554131 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ff639bee-182b-4dd2-aabc-d396a3bd9ce5-auth-proxy-config\") pod \"machine-config-operator-74547568cd-9rvnv\" (UID: \"ff639bee-182b-4dd2-aabc-d396a3bd9ce5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9rvnv" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.554150 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93ffc5bd-20a8-450c-861a-0aecc58d602b-config\") pod \"machine-approver-56656f9798-fnpz5\" (UID: \"93ffc5bd-20a8-450c-861a-0aecc58d602b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fnpz5" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.554172 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r765t\" (UniqueName: \"kubernetes.io/projected/4443d2fe-129a-41eb-b7b7-a7ee9bd3b6eb-kube-api-access-r765t\") pod \"authentication-operator-69f744f599-thckl\" (UID: \"4443d2fe-129a-41eb-b7b7-a7ee9bd3b6eb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-thckl" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.554188 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5t2sf\" (UniqueName: \"kubernetes.io/projected/e4f6934a-db7e-4031-baa8-2f41dc8c2c18-kube-api-access-5t2sf\") pod \"ingress-operator-5b745b69d9-qfdsg\" (UID: \"e4f6934a-db7e-4031-baa8-2f41dc8c2c18\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qfdsg" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.554203 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/b8455642-8313-4539-9729-fe658ef0e4d6-audit\") pod \"apiserver-76f77b778f-8xz4m\" (UID: \"b8455642-8313-4539-9729-fe658ef0e4d6\") " pod="openshift-apiserver/apiserver-76f77b778f-8xz4m" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.554229 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e4f6934a-db7e-4031-baa8-2f41dc8c2c18-trusted-ca\") pod \"ingress-operator-5b745b69d9-qfdsg\" (UID: \"e4f6934a-db7e-4031-baa8-2f41dc8c2c18\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qfdsg" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.554249 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f643d9e2-a7df-4aab-a2f5-4bf47363636b-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-jv24g\" (UID: \"f643d9e2-a7df-4aab-a2f5-4bf47363636b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jv24g" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.554261 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mpnf2" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.554264 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4443d2fe-129a-41eb-b7b7-a7ee9bd3b6eb-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-thckl\" (UID: \"4443d2fe-129a-41eb-b7b7-a7ee9bd3b6eb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-thckl" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.554414 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-992nr\" (UniqueName: \"kubernetes.io/projected/3f3f6253-533f-4d89-ac10-f26eca1ac390-kube-api-access-992nr\") pod \"console-operator-58897d9998-4lqzs\" (UID: \"3f3f6253-533f-4d89-ac10-f26eca1ac390\") " pod="openshift-console-operator/console-operator-58897d9998-4lqzs" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.554445 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-rqs9q\" (UID: \"263721d5-fcca-411a-876e-0acfe0863d2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-rqs9q" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.554474 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/93ffc5bd-20a8-450c-861a-0aecc58d602b-machine-approver-tls\") pod \"machine-approver-56656f9798-fnpz5\" (UID: \"93ffc5bd-20a8-450c-861a-0aecc58d602b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fnpz5" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.554498 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/bd64a08a-6920-4a0f-b299-aa5a009236a1-encryption-config\") pod \"apiserver-7bbb656c7d-8br44\" (UID: \"bd64a08a-6920-4a0f-b299-aa5a009236a1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8br44" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.554520 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b8455642-8313-4539-9729-fe658ef0e4d6-etcd-client\") pod \"apiserver-76f77b778f-8xz4m\" (UID: \"b8455642-8313-4539-9729-fe658ef0e4d6\") " pod="openshift-apiserver/apiserver-76f77b778f-8xz4m" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.554564 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/b8455642-8313-4539-9729-fe658ef0e4d6-etcd-serving-ca\") pod \"apiserver-76f77b778f-8xz4m\" (UID: \"b8455642-8313-4539-9729-fe658ef0e4d6\") " pod="openshift-apiserver/apiserver-76f77b778f-8xz4m" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.554586 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbwpp\" (UniqueName: \"kubernetes.io/projected/1051665a-7918-487f-ad08-ff6d53d113ca-kube-api-access-gbwpp\") pod \"etcd-operator-b45778765-t56zb\" (UID: \"1051665a-7918-487f-ad08-ff6d53d113ca\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t56zb" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.554607 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/b8455642-8313-4539-9729-fe658ef0e4d6-node-pullsecrets\") pod \"apiserver-76f77b778f-8xz4m\" (UID: \"b8455642-8313-4539-9729-fe658ef0e4d6\") " pod="openshift-apiserver/apiserver-76f77b778f-8xz4m" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.554629 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4443d2fe-129a-41eb-b7b7-a7ee9bd3b6eb-config\") pod \"authentication-operator-69f744f599-thckl\" (UID: \"4443d2fe-129a-41eb-b7b7-a7ee9bd3b6eb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-thckl" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.554651 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ff639bee-182b-4dd2-aabc-d396a3bd9ce5-images\") pod \"machine-config-operator-74547568cd-9rvnv\" (UID: \"ff639bee-182b-4dd2-aabc-d396a3bd9ce5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9rvnv" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.554676 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-rqs9q\" (UID: \"263721d5-fcca-411a-876e-0acfe0863d2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-rqs9q" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.554702 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/2362780e-93d1-4f97-bda0-138ed5180bb0-console-serving-cert\") pod \"console-f9d7485db-7ckfj\" (UID: \"2362780e-93d1-4f97-bda0-138ed5180bb0\") " pod="openshift-console/console-f9d7485db-7ckfj" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.554736 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/dc4fc98a-badd-46dc-b700-6b01aff112b4-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-ghcrj\" (UID: \"dc4fc98a-badd-46dc-b700-6b01aff112b4\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ghcrj" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.554780 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccnj9\" (UniqueName: \"kubernetes.io/projected/a68fc0dd-997c-4d12-9a20-5345dfe98a41-kube-api-access-ccnj9\") pod \"controller-manager-879f6c89f-cgpdh\" (UID: \"a68fc0dd-997c-4d12-9a20-5345dfe98a41\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cgpdh" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.554804 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/bd64a08a-6920-4a0f-b299-aa5a009236a1-audit-policies\") pod \"apiserver-7bbb656c7d-8br44\" (UID: \"bd64a08a-6920-4a0f-b299-aa5a009236a1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8br44" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.554824 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db213df1-bfa2-4edd-a3fc-246e35e20585-config\") pod \"route-controller-manager-6576b87f9c-9v855\" (UID: \"db213df1-bfa2-4edd-a3fc-246e35e20585\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9v855" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.554849 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/f57cc683-84dc-4538-96fe-9a198a5b008f-available-featuregates\") pod \"openshift-config-operator-7777fb866f-s5b86\" (UID: \"f57cc683-84dc-4538-96fe-9a198a5b008f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-s5b86" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.554871 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnz55\" (UniqueName: \"kubernetes.io/projected/7abae297-572a-4c3c-8e7d-4e2157d838ad-kube-api-access-xnz55\") pod \"dns-operator-744455d44c-gkpck\" (UID: \"7abae297-572a-4c3c-8e7d-4e2157d838ad\") " pod="openshift-dns-operator/dns-operator-744455d44c-gkpck" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.554881 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bd64a08a-6920-4a0f-b299-aa5a009236a1-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-8br44\" (UID: \"bd64a08a-6920-4a0f-b299-aa5a009236a1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8br44" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.554894 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bd64a08a-6920-4a0f-b299-aa5a009236a1-serving-cert\") pod \"apiserver-7bbb656c7d-8br44\" (UID: \"bd64a08a-6920-4a0f-b299-aa5a009236a1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8br44" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.554945 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2n5q\" (UniqueName: \"kubernetes.io/projected/b2994e8f-6482-4dca-baf1-148754e78e85-kube-api-access-l2n5q\") pod \"olm-operator-6b444d44fb-9bfz8\" (UID: \"b2994e8f-6482-4dca-baf1-148754e78e85\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9bfz8" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.554972 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/93ffc5bd-20a8-450c-861a-0aecc58d602b-auth-proxy-config\") pod \"machine-approver-56656f9798-fnpz5\" (UID: \"93ffc5bd-20a8-450c-861a-0aecc58d602b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fnpz5" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.555005 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9454f4b6-e1f9-4dc1-953c-442dbd4088ad-config\") pod \"kube-controller-manager-operator-78b949d7b-x6zsh\" (UID: \"9454f4b6-e1f9-4dc1-953c-442dbd4088ad\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x6zsh" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.555027 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6258654e-b524-4fdd-ba9b-7443b57c2585-config\") pod \"kube-apiserver-operator-766d6c64bb-5rwsv\" (UID: \"6258654e-b524-4fdd-ba9b-7443b57c2585\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5rwsv" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.555053 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-rqs9q\" (UID: \"263721d5-fcca-411a-876e-0acfe0863d2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-rqs9q" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.555076 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/263721d5-fcca-411a-876e-0acfe0863d2d-audit-policies\") pod \"oauth-openshift-558db77b4-rqs9q\" (UID: \"263721d5-fcca-411a-876e-0acfe0863d2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-rqs9q" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.555099 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9172dd32-c65e-4637-ada8-d2c5b0a772ab-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-fqdzj\" (UID: \"9172dd32-c65e-4637-ada8-d2c5b0a772ab\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fqdzj" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.555120 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b8455642-8313-4539-9729-fe658ef0e4d6-serving-cert\") pod \"apiserver-76f77b778f-8xz4m\" (UID: \"b8455642-8313-4539-9729-fe658ef0e4d6\") " pod="openshift-apiserver/apiserver-76f77b778f-8xz4m" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.555145 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b2994e8f-6482-4dca-baf1-148754e78e85-srv-cert\") pod \"olm-operator-6b444d44fb-9bfz8\" (UID: \"b2994e8f-6482-4dca-baf1-148754e78e85\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9bfz8" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.555170 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6258654e-b524-4fdd-ba9b-7443b57c2585-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-5rwsv\" (UID: \"6258654e-b524-4fdd-ba9b-7443b57c2585\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5rwsv" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.555198 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb15a464-d0e8-4e49-b018-14353cc9d6cb-config\") pod \"machine-api-operator-5694c8668f-9s5r8\" (UID: \"eb15a464-d0e8-4e49-b018-14353cc9d6cb\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9s5r8" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.555221 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/263721d5-fcca-411a-876e-0acfe0863d2d-audit-dir\") pod \"oauth-openshift-558db77b4-rqs9q\" (UID: \"263721d5-fcca-411a-876e-0acfe0863d2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-rqs9q" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.555251 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f57cc683-84dc-4538-96fe-9a198a5b008f-serving-cert\") pod \"openshift-config-operator-7777fb866f-s5b86\" (UID: \"f57cc683-84dc-4538-96fe-9a198a5b008f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-s5b86" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.555277 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-rqs9q\" (UID: \"263721d5-fcca-411a-876e-0acfe0863d2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-rqs9q" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.555307 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a92fd9bb-6d1b-4ab1-9f1a-bc9086ecb08e-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-ppzvl\" (UID: \"a92fd9bb-6d1b-4ab1-9f1a-bc9086ecb08e\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ppzvl" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.555329 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a68fc0dd-997c-4d12-9a20-5345dfe98a41-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-cgpdh\" (UID: \"a68fc0dd-997c-4d12-9a20-5345dfe98a41\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cgpdh" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.555353 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/0fbd23d0-a6b6-4a4e-b418-07fc38e5f497-default-certificate\") pod \"router-default-5444994796-rrhqx\" (UID: \"0fbd23d0-a6b6-4a4e-b418-07fc38e5f497\") " pod="openshift-ingress/router-default-5444994796-rrhqx" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.555398 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cch2d\" (UniqueName: \"kubernetes.io/projected/93ffc5bd-20a8-450c-861a-0aecc58d602b-kube-api-access-cch2d\") pod \"machine-approver-56656f9798-fnpz5\" (UID: \"93ffc5bd-20a8-450c-861a-0aecc58d602b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fnpz5" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.555430 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-zfjwx"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.555457 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/263721d5-fcca-411a-876e-0acfe0863d2d-audit-dir\") pod \"oauth-openshift-558db77b4-rqs9q\" (UID: \"263721d5-fcca-411a-876e-0acfe0863d2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-rqs9q" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.555489 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a68fc0dd-997c-4d12-9a20-5345dfe98a41-config\") pod \"controller-manager-879f6c89f-cgpdh\" (UID: \"a68fc0dd-997c-4d12-9a20-5345dfe98a41\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cgpdh" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.555514 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f3f6253-533f-4d89-ac10-f26eca1ac390-serving-cert\") pod \"console-operator-58897d9998-4lqzs\" (UID: \"3f3f6253-533f-4d89-ac10-f26eca1ac390\") " pod="openshift-console-operator/console-operator-58897d9998-4lqzs" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.555547 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/db213df1-bfa2-4edd-a3fc-246e35e20585-serving-cert\") pod \"route-controller-manager-6576b87f9c-9v855\" (UID: \"db213df1-bfa2-4edd-a3fc-246e35e20585\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9v855" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.555569 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/9172dd32-c65e-4637-ada8-d2c5b0a772ab-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-fqdzj\" (UID: \"9172dd32-c65e-4637-ada8-d2c5b0a772ab\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fqdzj" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.555597 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ff639bee-182b-4dd2-aabc-d396a3bd9ce5-proxy-tls\") pod \"machine-config-operator-74547568cd-9rvnv\" (UID: \"ff639bee-182b-4dd2-aabc-d396a3bd9ce5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9rvnv" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.555616 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a68fc0dd-997c-4d12-9a20-5345dfe98a41-client-ca\") pod \"controller-manager-879f6c89f-cgpdh\" (UID: \"a68fc0dd-997c-4d12-9a20-5345dfe98a41\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cgpdh" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.555674 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfqp4\" (UniqueName: \"kubernetes.io/projected/dc4fc98a-badd-46dc-b700-6b01aff112b4-kube-api-access-lfqp4\") pod \"cluster-samples-operator-665b6dd947-ghcrj\" (UID: \"dc4fc98a-badd-46dc-b700-6b01aff112b4\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ghcrj" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.555697 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-rqs9q\" (UID: \"263721d5-fcca-411a-876e-0acfe0863d2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-rqs9q" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.555717 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f3f6253-533f-4d89-ac10-f26eca1ac390-config\") pod \"console-operator-58897d9998-4lqzs\" (UID: \"3f3f6253-533f-4d89-ac10-f26eca1ac390\") " pod="openshift-console-operator/console-operator-58897d9998-4lqzs" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.555738 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/1051665a-7918-487f-ad08-ff6d53d113ca-etcd-service-ca\") pod \"etcd-operator-b45778765-t56zb\" (UID: \"1051665a-7918-487f-ad08-ff6d53d113ca\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t56zb" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.555781 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4443d2fe-129a-41eb-b7b7-a7ee9bd3b6eb-service-ca-bundle\") pod \"authentication-operator-69f744f599-thckl\" (UID: \"4443d2fe-129a-41eb-b7b7-a7ee9bd3b6eb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-thckl" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.555807 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgrht\" (UniqueName: \"kubernetes.io/projected/263721d5-fcca-411a-876e-0acfe0863d2d-kube-api-access-bgrht\") pod \"oauth-openshift-558db77b4-rqs9q\" (UID: \"263721d5-fcca-411a-876e-0acfe0863d2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-rqs9q" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.555812 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/263721d5-fcca-411a-876e-0acfe0863d2d-audit-policies\") pod \"oauth-openshift-558db77b4-rqs9q\" (UID: \"263721d5-fcca-411a-876e-0acfe0863d2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-rqs9q" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.555830 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-rqs9q\" (UID: \"263721d5-fcca-411a-876e-0acfe0863d2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-rqs9q" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.555870 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/2362780e-93d1-4f97-bda0-138ed5180bb0-console-oauth-config\") pod \"console-f9d7485db-7ckfj\" (UID: \"2362780e-93d1-4f97-bda0-138ed5180bb0\") " pod="openshift-console/console-f9d7485db-7ckfj" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.556204 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-zb4m6"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.556342 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2362780e-93d1-4f97-bda0-138ed5180bb0-trusted-ca-bundle\") pod \"console-f9d7485db-7ckfj\" (UID: \"2362780e-93d1-4f97-bda0-138ed5180bb0\") " pod="openshift-console/console-f9d7485db-7ckfj" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.556382 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e61f4275-03bf-4cb0-bb9a-e519c41f159a-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-jwc8v\" (UID: \"e61f4275-03bf-4cb0-bb9a-e519c41f159a\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jwc8v" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.556411 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b8455642-8313-4539-9729-fe658ef0e4d6-trusted-ca-bundle\") pod \"apiserver-76f77b778f-8xz4m\" (UID: \"b8455642-8313-4539-9729-fe658ef0e4d6\") " pod="openshift-apiserver/apiserver-76f77b778f-8xz4m" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.556439 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88cj6\" (UniqueName: \"kubernetes.io/projected/1a34c64b-194e-41a6-8b55-0a87187c4cfc-kube-api-access-88cj6\") pod \"control-plane-machine-set-operator-78cbb6b69f-6d29n\" (UID: \"1a34c64b-194e-41a6-8b55-0a87187c4cfc\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6d29n" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.556457 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-zfjwx" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.556463 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a92fd9bb-6d1b-4ab1-9f1a-bc9086ecb08e-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-ppzvl\" (UID: \"a92fd9bb-6d1b-4ab1-9f1a-bc9086ecb08e\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ppzvl" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.556489 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrwfd\" (UniqueName: \"kubernetes.io/projected/ff639bee-182b-4dd2-aabc-d396a3bd9ce5-kube-api-access-hrwfd\") pod \"machine-config-operator-74547568cd-9rvnv\" (UID: \"ff639bee-182b-4dd2-aabc-d396a3bd9ce5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9rvnv" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.556528 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/eb15a464-d0e8-4e49-b018-14353cc9d6cb-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-9s5r8\" (UID: \"eb15a464-d0e8-4e49-b018-14353cc9d6cb\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9s5r8" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.556554 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-rqs9q\" (UID: \"263721d5-fcca-411a-876e-0acfe0863d2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-rqs9q" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.556579 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-rqs9q\" (UID: \"263721d5-fcca-411a-876e-0acfe0863d2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-rqs9q" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.556603 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3f3f6253-533f-4d89-ac10-f26eca1ac390-trusted-ca\") pod \"console-operator-58897d9998-4lqzs\" (UID: \"3f3f6253-533f-4d89-ac10-f26eca1ac390\") " pod="openshift-console-operator/console-operator-58897d9998-4lqzs" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.556628 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4443d2fe-129a-41eb-b7b7-a7ee9bd3b6eb-serving-cert\") pod \"authentication-operator-69f744f599-thckl\" (UID: \"4443d2fe-129a-41eb-b7b7-a7ee9bd3b6eb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-thckl" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.556642 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/bd64a08a-6920-4a0f-b299-aa5a009236a1-audit-policies\") pod \"apiserver-7bbb656c7d-8br44\" (UID: \"bd64a08a-6920-4a0f-b299-aa5a009236a1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8br44" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.556654 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/2362780e-93d1-4f97-bda0-138ed5180bb0-oauth-serving-cert\") pod \"console-f9d7485db-7ckfj\" (UID: \"2362780e-93d1-4f97-bda0-138ed5180bb0\") " pod="openshift-console/console-f9d7485db-7ckfj" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.556693 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-zb4m6" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.556708 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9454f4b6-e1f9-4dc1-953c-442dbd4088ad-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-x6zsh\" (UID: \"9454f4b6-e1f9-4dc1-953c-442dbd4088ad\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x6zsh" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.556772 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/1051665a-7918-487f-ad08-ff6d53d113ca-etcd-ca\") pod \"etcd-operator-b45778765-t56zb\" (UID: \"1051665a-7918-487f-ad08-ff6d53d113ca\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t56zb" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.556814 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hxvf\" (UniqueName: \"kubernetes.io/projected/e61f4275-03bf-4cb0-bb9a-e519c41f159a-kube-api-access-4hxvf\") pod \"machine-config-controller-84d6567774-jwc8v\" (UID: \"e61f4275-03bf-4cb0-bb9a-e519c41f159a\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jwc8v" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.556834 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwzv6\" (UniqueName: \"kubernetes.io/projected/eb15a464-d0e8-4e49-b018-14353cc9d6cb-kube-api-access-kwzv6\") pod \"machine-api-operator-5694c8668f-9s5r8\" (UID: \"eb15a464-d0e8-4e49-b018-14353cc9d6cb\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9s5r8" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.556850 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwm24\" (UniqueName: \"kubernetes.io/projected/a92fd9bb-6d1b-4ab1-9f1a-bc9086ecb08e-kube-api-access-fwm24\") pod \"kube-storage-version-migrator-operator-b67b599dd-ppzvl\" (UID: \"a92fd9bb-6d1b-4ab1-9f1a-bc9086ecb08e\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ppzvl" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.556887 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/bd64a08a-6920-4a0f-b299-aa5a009236a1-etcd-client\") pod \"apiserver-7bbb656c7d-8br44\" (UID: \"bd64a08a-6920-4a0f-b299-aa5a009236a1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8br44" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.556905 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2362780e-93d1-4f97-bda0-138ed5180bb0-service-ca\") pod \"console-f9d7485db-7ckfj\" (UID: \"2362780e-93d1-4f97-bda0-138ed5180bb0\") " pod="openshift-console/console-f9d7485db-7ckfj" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.556924 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e61f4275-03bf-4cb0-bb9a-e519c41f159a-proxy-tls\") pod \"machine-config-controller-84d6567774-jwc8v\" (UID: \"e61f4275-03bf-4cb0-bb9a-e519c41f159a\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jwc8v" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.556944 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-rqs9q\" (UID: \"263721d5-fcca-411a-876e-0acfe0863d2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-rqs9q" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.556961 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-rqs9q\" (UID: \"263721d5-fcca-411a-876e-0acfe0863d2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-rqs9q" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.556980 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0fbd23d0-a6b6-4a4e-b418-07fc38e5f497-metrics-certs\") pod \"router-default-5444994796-rrhqx\" (UID: \"0fbd23d0-a6b6-4a4e-b418-07fc38e5f497\") " pod="openshift-ingress/router-default-5444994796-rrhqx" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.557002 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-rqs9q\" (UID: \"263721d5-fcca-411a-876e-0acfe0863d2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-rqs9q" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.557020 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpqvr\" (UniqueName: \"kubernetes.io/projected/db213df1-bfa2-4edd-a3fc-246e35e20585-kube-api-access-fpqvr\") pod \"route-controller-manager-6576b87f9c-9v855\" (UID: \"db213df1-bfa2-4edd-a3fc-246e35e20585\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9v855" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.557038 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/db213df1-bfa2-4edd-a3fc-246e35e20585-client-ca\") pod \"route-controller-manager-6576b87f9c-9v855\" (UID: \"db213df1-bfa2-4edd-a3fc-246e35e20585\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9v855" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.557056 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plz2r\" (UniqueName: \"kubernetes.io/projected/b8455642-8313-4539-9729-fe658ef0e4d6-kube-api-access-plz2r\") pod \"apiserver-76f77b778f-8xz4m\" (UID: \"b8455642-8313-4539-9729-fe658ef0e4d6\") " pod="openshift-apiserver/apiserver-76f77b778f-8xz4m" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.557075 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbrbz\" (UniqueName: \"kubernetes.io/projected/2362780e-93d1-4f97-bda0-138ed5180bb0-kube-api-access-qbrbz\") pod \"console-f9d7485db-7ckfj\" (UID: \"2362780e-93d1-4f97-bda0-138ed5180bb0\") " pod="openshift-console/console-f9d7485db-7ckfj" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.557093 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/1a34c64b-194e-41a6-8b55-0a87187c4cfc-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-6d29n\" (UID: \"1a34c64b-194e-41a6-8b55-0a87187c4cfc\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6d29n" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.557920 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db213df1-bfa2-4edd-a3fc-246e35e20585-config\") pod \"route-controller-manager-6576b87f9c-9v855\" (UID: \"db213df1-bfa2-4edd-a3fc-246e35e20585\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9v855" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.558727 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/2362780e-93d1-4f97-bda0-138ed5180bb0-oauth-serving-cert\") pod \"console-f9d7485db-7ckfj\" (UID: \"2362780e-93d1-4f97-bda0-138ed5180bb0\") " pod="openshift-console/console-f9d7485db-7ckfj" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.558727 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f3f6253-533f-4d89-ac10-f26eca1ac390-config\") pod \"console-operator-58897d9998-4lqzs\" (UID: \"3f3f6253-533f-4d89-ac10-f26eca1ac390\") " pod="openshift-console-operator/console-operator-58897d9998-4lqzs" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.558736 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ddpmx"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.560011 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-rqs9q\" (UID: \"263721d5-fcca-411a-876e-0acfe0863d2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-rqs9q" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.560327 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/2362780e-93d1-4f97-bda0-138ed5180bb0-console-oauth-config\") pod \"console-f9d7485db-7ckfj\" (UID: \"2362780e-93d1-4f97-bda0-138ed5180bb0\") " pod="openshift-console/console-f9d7485db-7ckfj" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.561914 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2362780e-93d1-4f97-bda0-138ed5180bb0-service-ca\") pod \"console-f9d7485db-7ckfj\" (UID: \"2362780e-93d1-4f97-bda0-138ed5180bb0\") " pod="openshift-console/console-f9d7485db-7ckfj" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.563516 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2362780e-93d1-4f97-bda0-138ed5180bb0-trusted-ca-bundle\") pod \"console-f9d7485db-7ckfj\" (UID: \"2362780e-93d1-4f97-bda0-138ed5180bb0\") " pod="openshift-console/console-f9d7485db-7ckfj" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.564298 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/db213df1-bfa2-4edd-a3fc-246e35e20585-client-ca\") pod \"route-controller-manager-6576b87f9c-9v855\" (UID: \"db213df1-bfa2-4edd-a3fc-246e35e20585\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9v855" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.564346 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxx8q\" (UniqueName: \"kubernetes.io/projected/9172dd32-c65e-4637-ada8-d2c5b0a772ab-kube-api-access-rxx8q\") pod \"cluster-image-registry-operator-dc59b4c8b-fqdzj\" (UID: \"9172dd32-c65e-4637-ada8-d2c5b0a772ab\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fqdzj" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.564366 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f643d9e2-a7df-4aab-a2f5-4bf47363636b-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-jv24g\" (UID: \"f643d9e2-a7df-4aab-a2f5-4bf47363636b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jv24g" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.564386 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dvdr\" (UniqueName: \"kubernetes.io/projected/0057cff6-6085-4b41-bf4f-e1121bece99a-kube-api-access-7dvdr\") pod \"downloads-7954f5f757-7cg8t\" (UID: \"0057cff6-6085-4b41-bf4f-e1121bece99a\") " pod="openshift-console/downloads-7954f5f757-7cg8t" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.564412 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9172dd32-c65e-4637-ada8-d2c5b0a772ab-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-fqdzj\" (UID: \"9172dd32-c65e-4637-ada8-d2c5b0a772ab\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fqdzj" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.564432 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/eb15a464-d0e8-4e49-b018-14353cc9d6cb-images\") pod \"machine-api-operator-5694c8668f-9s5r8\" (UID: \"eb15a464-d0e8-4e49-b018-14353cc9d6cb\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9s5r8" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.564448 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9bfx\" (UniqueName: \"kubernetes.io/projected/f643d9e2-a7df-4aab-a2f5-4bf47363636b-kube-api-access-m9bfx\") pod \"openshift-controller-manager-operator-756b6f6bc6-jv24g\" (UID: \"f643d9e2-a7df-4aab-a2f5-4bf47363636b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jv24g" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.564466 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/b8455642-8313-4539-9729-fe658ef0e4d6-encryption-config\") pod \"apiserver-76f77b778f-8xz4m\" (UID: \"b8455642-8313-4539-9729-fe658ef0e4d6\") " pod="openshift-apiserver/apiserver-76f77b778f-8xz4m" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.564498 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b2994e8f-6482-4dca-baf1-148754e78e85-profile-collector-cert\") pod \"olm-operator-6b444d44fb-9bfz8\" (UID: \"b2994e8f-6482-4dca-baf1-148754e78e85\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9bfz8" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.564526 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/2362780e-93d1-4f97-bda0-138ed5180bb0-console-config\") pod \"console-f9d7485db-7ckfj\" (UID: \"2362780e-93d1-4f97-bda0-138ed5180bb0\") " pod="openshift-console/console-f9d7485db-7ckfj" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.564546 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1051665a-7918-487f-ad08-ff6d53d113ca-serving-cert\") pod \"etcd-operator-b45778765-t56zb\" (UID: \"1051665a-7918-487f-ad08-ff6d53d113ca\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t56zb" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.564578 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1051665a-7918-487f-ad08-ff6d53d113ca-config\") pod \"etcd-operator-b45778765-t56zb\" (UID: \"1051665a-7918-487f-ad08-ff6d53d113ca\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t56zb" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.564594 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6258654e-b524-4fdd-ba9b-7443b57c2585-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-5rwsv\" (UID: \"6258654e-b524-4fdd-ba9b-7443b57c2585\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5rwsv" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.564611 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/0fbd23d0-a6b6-4a4e-b418-07fc38e5f497-stats-auth\") pod \"router-default-5444994796-rrhqx\" (UID: \"0fbd23d0-a6b6-4a4e-b418-07fc38e5f497\") " pod="openshift-ingress/router-default-5444994796-rrhqx" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.564627 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b8455642-8313-4539-9729-fe658ef0e4d6-config\") pod \"apiserver-76f77b778f-8xz4m\" (UID: \"b8455642-8313-4539-9729-fe658ef0e4d6\") " pod="openshift-apiserver/apiserver-76f77b778f-8xz4m" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.564643 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6cpv\" (UniqueName: \"kubernetes.io/projected/0fbd23d0-a6b6-4a4e-b418-07fc38e5f497-kube-api-access-c6cpv\") pod \"router-default-5444994796-rrhqx\" (UID: \"0fbd23d0-a6b6-4a4e-b418-07fc38e5f497\") " pod="openshift-ingress/router-default-5444994796-rrhqx" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.564660 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6w6b\" (UniqueName: \"kubernetes.io/projected/bd64a08a-6920-4a0f-b299-aa5a009236a1-kube-api-access-g6w6b\") pod \"apiserver-7bbb656c7d-8br44\" (UID: \"bd64a08a-6920-4a0f-b299-aa5a009236a1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8br44" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.564697 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b8455642-8313-4539-9729-fe658ef0e4d6-audit-dir\") pod \"apiserver-76f77b778f-8xz4m\" (UID: \"b8455642-8313-4539-9729-fe658ef0e4d6\") " pod="openshift-apiserver/apiserver-76f77b778f-8xz4m" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.564712 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a68fc0dd-997c-4d12-9a20-5345dfe98a41-serving-cert\") pod \"controller-manager-879f6c89f-cgpdh\" (UID: \"a68fc0dd-997c-4d12-9a20-5345dfe98a41\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cgpdh" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.564729 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/bd64a08a-6920-4a0f-b299-aa5a009236a1-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-8br44\" (UID: \"bd64a08a-6920-4a0f-b299-aa5a009236a1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8br44" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.564811 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7abae297-572a-4c3c-8e7d-4e2157d838ad-metrics-tls\") pod \"dns-operator-744455d44c-gkpck\" (UID: \"7abae297-572a-4c3c-8e7d-4e2157d838ad\") " pod="openshift-dns-operator/dns-operator-744455d44c-gkpck" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.564950 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.565378 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/2362780e-93d1-4f97-bda0-138ed5180bb0-console-config\") pod \"console-f9d7485db-7ckfj\" (UID: \"2362780e-93d1-4f97-bda0-138ed5180bb0\") " pod="openshift-console/console-f9d7485db-7ckfj" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.565940 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/bd64a08a-6920-4a0f-b299-aa5a009236a1-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-8br44\" (UID: \"bd64a08a-6920-4a0f-b299-aa5a009236a1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8br44" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.566326 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bd64a08a-6920-4a0f-b299-aa5a009236a1-serving-cert\") pod \"apiserver-7bbb656c7d-8br44\" (UID: \"bd64a08a-6920-4a0f-b299-aa5a009236a1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8br44" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.566722 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-rqs9q\" (UID: \"263721d5-fcca-411a-876e-0acfe0863d2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-rqs9q" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.570363 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-wrh75"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.570469 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/bd64a08a-6920-4a0f-b299-aa5a009236a1-etcd-client\") pod \"apiserver-7bbb656c7d-8br44\" (UID: \"bd64a08a-6920-4a0f-b299-aa5a009236a1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8br44" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.570846 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-4lqzs"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.570875 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-blwnp"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.570954 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/bd64a08a-6920-4a0f-b299-aa5a009236a1-encryption-config\") pod \"apiserver-7bbb656c7d-8br44\" (UID: \"bd64a08a-6920-4a0f-b299-aa5a009236a1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8br44" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.571140 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/2362780e-93d1-4f97-bda0-138ed5180bb0-console-serving-cert\") pod \"console-f9d7485db-7ckfj\" (UID: \"2362780e-93d1-4f97-bda0-138ed5180bb0\") " pod="openshift-console/console-f9d7485db-7ckfj" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.571251 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ddpmx" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.573612 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-rqs9q\" (UID: \"263721d5-fcca-411a-876e-0acfe0863d2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-rqs9q" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.574037 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-rqs9q\" (UID: \"263721d5-fcca-411a-876e-0acfe0863d2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-rqs9q" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.574182 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-wrh75" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.574248 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3f3f6253-533f-4d89-ac10-f26eca1ac390-trusted-ca\") pod \"console-operator-58897d9998-4lqzs\" (UID: \"3f3f6253-533f-4d89-ac10-f26eca1ac390\") " pod="openshift-console-operator/console-operator-58897d9998-4lqzs" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.579887 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416530-88l5w"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.580384 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f3f6253-533f-4d89-ac10-f26eca1ac390-serving-cert\") pod \"console-operator-58897d9998-4lqzs\" (UID: \"3f3f6253-533f-4d89-ac10-f26eca1ac390\") " pod="openshift-console-operator/console-operator-58897d9998-4lqzs" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.580009 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.581338 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/dc4fc98a-badd-46dc-b700-6b01aff112b4-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-ghcrj\" (UID: \"dc4fc98a-badd-46dc-b700-6b01aff112b4\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ghcrj" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.583221 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-9v855"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.583363 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-cgpdh"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.583382 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-thckl"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.583436 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416530-88l5w" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.583497 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-9s5r8"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.583524 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fqdzj"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.583690 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-s5b86"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.583715 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-8xz4m"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.584233 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-7ckfj"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.585102 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-whhs8"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.586451 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-rqs9q\" (UID: \"263721d5-fcca-411a-876e-0acfe0863d2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-rqs9q" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.586636 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-rqs9q\" (UID: \"263721d5-fcca-411a-876e-0acfe0863d2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-rqs9q" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.586654 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-rqs9q\" (UID: \"263721d5-fcca-411a-876e-0acfe0863d2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-rqs9q" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.586932 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-rqs9q\" (UID: \"263721d5-fcca-411a-876e-0acfe0863d2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-rqs9q" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.587010 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.587193 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/db213df1-bfa2-4edd-a3fc-246e35e20585-serving-cert\") pod \"route-controller-manager-6576b87f9c-9v855\" (UID: \"db213df1-bfa2-4edd-a3fc-246e35e20585\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9v855" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.587379 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-rqs9q\" (UID: \"263721d5-fcca-411a-876e-0acfe0863d2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-rqs9q" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.587527 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-rqs9q\" (UID: \"263721d5-fcca-411a-876e-0acfe0863d2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-rqs9q" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.589092 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.591868 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ghcrj"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.591903 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-rqs9q"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.591989 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-whhs8" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.593982 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-jwc8v"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.597179 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-rqs9q\" (UID: \"263721d5-fcca-411a-876e-0acfe0863d2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-rqs9q" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.597241 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4gkmd"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.600798 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ppzvl"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.602466 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-qfdsg"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.602736 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-t56zb"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.608178 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5rwsv"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.610575 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.610820 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-7cg8t"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.613261 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-gkpck"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.617498 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6d29n"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.620247 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wbjzw"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.621976 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-9rvnv"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.623944 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2nslt"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.626349 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-ntmms"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.627536 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x6zsh"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.628656 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.632150 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gmvnz"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.634298 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-wrh75"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.636661 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-8br44"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.638571 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-9sxg5"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.639578 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-9sxg5" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.640654 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-khkwh"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.641269 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-khkwh" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.641601 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9bfz8"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.642771 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jv24g"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.643950 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-blwnp"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.645187 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416530-88l5w"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.646808 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-zb4m6"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.648221 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-whhs8"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.649140 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.649307 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mpnf2"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.650634 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ddpmx"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.651356 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-zfjwx"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.652608 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-9sxg5"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.653918 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-5kgv6"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.654645 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-5kgv6" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.655034 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-5kgv6"] Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.665619 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwm24\" (UniqueName: \"kubernetes.io/projected/a92fd9bb-6d1b-4ab1-9f1a-bc9086ecb08e-kube-api-access-fwm24\") pod \"kube-storage-version-migrator-operator-b67b599dd-ppzvl\" (UID: \"a92fd9bb-6d1b-4ab1-9f1a-bc9086ecb08e\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ppzvl" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.665659 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e61f4275-03bf-4cb0-bb9a-e519c41f159a-proxy-tls\") pod \"machine-config-controller-84d6567774-jwc8v\" (UID: \"e61f4275-03bf-4cb0-bb9a-e519c41f159a\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jwc8v" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.665683 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0fbd23d0-a6b6-4a4e-b418-07fc38e5f497-metrics-certs\") pod \"router-default-5444994796-rrhqx\" (UID: \"0fbd23d0-a6b6-4a4e-b418-07fc38e5f497\") " pod="openshift-ingress/router-default-5444994796-rrhqx" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.665713 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plz2r\" (UniqueName: \"kubernetes.io/projected/b8455642-8313-4539-9729-fe658ef0e4d6-kube-api-access-plz2r\") pod \"apiserver-76f77b778f-8xz4m\" (UID: \"b8455642-8313-4539-9729-fe658ef0e4d6\") " pod="openshift-apiserver/apiserver-76f77b778f-8xz4m" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.665774 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/1a34c64b-194e-41a6-8b55-0a87187c4cfc-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-6d29n\" (UID: \"1a34c64b-194e-41a6-8b55-0a87187c4cfc\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6d29n" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.665803 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dvdr\" (UniqueName: \"kubernetes.io/projected/0057cff6-6085-4b41-bf4f-e1121bece99a-kube-api-access-7dvdr\") pod \"downloads-7954f5f757-7cg8t\" (UID: \"0057cff6-6085-4b41-bf4f-e1121bece99a\") " pod="openshift-console/downloads-7954f5f757-7cg8t" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.665826 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxx8q\" (UniqueName: \"kubernetes.io/projected/9172dd32-c65e-4637-ada8-d2c5b0a772ab-kube-api-access-rxx8q\") pod \"cluster-image-registry-operator-dc59b4c8b-fqdzj\" (UID: \"9172dd32-c65e-4637-ada8-d2c5b0a772ab\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fqdzj" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.665870 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f643d9e2-a7df-4aab-a2f5-4bf47363636b-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-jv24g\" (UID: \"f643d9e2-a7df-4aab-a2f5-4bf47363636b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jv24g" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.665894 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9172dd32-c65e-4637-ada8-d2c5b0a772ab-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-fqdzj\" (UID: \"9172dd32-c65e-4637-ada8-d2c5b0a772ab\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fqdzj" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.665916 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/eb15a464-d0e8-4e49-b018-14353cc9d6cb-images\") pod \"machine-api-operator-5694c8668f-9s5r8\" (UID: \"eb15a464-d0e8-4e49-b018-14353cc9d6cb\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9s5r8" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.665940 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/b8455642-8313-4539-9729-fe658ef0e4d6-encryption-config\") pod \"apiserver-76f77b778f-8xz4m\" (UID: \"b8455642-8313-4539-9729-fe658ef0e4d6\") " pod="openshift-apiserver/apiserver-76f77b778f-8xz4m" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.665965 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9bfx\" (UniqueName: \"kubernetes.io/projected/f643d9e2-a7df-4aab-a2f5-4bf47363636b-kube-api-access-m9bfx\") pod \"openshift-controller-manager-operator-756b6f6bc6-jv24g\" (UID: \"f643d9e2-a7df-4aab-a2f5-4bf47363636b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jv24g" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.665989 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b2994e8f-6482-4dca-baf1-148754e78e85-profile-collector-cert\") pod \"olm-operator-6b444d44fb-9bfz8\" (UID: \"b2994e8f-6482-4dca-baf1-148754e78e85\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9bfz8" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.666010 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6258654e-b524-4fdd-ba9b-7443b57c2585-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-5rwsv\" (UID: \"6258654e-b524-4fdd-ba9b-7443b57c2585\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5rwsv" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.666033 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1051665a-7918-487f-ad08-ff6d53d113ca-serving-cert\") pod \"etcd-operator-b45778765-t56zb\" (UID: \"1051665a-7918-487f-ad08-ff6d53d113ca\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t56zb" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.666056 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1051665a-7918-487f-ad08-ff6d53d113ca-config\") pod \"etcd-operator-b45778765-t56zb\" (UID: \"1051665a-7918-487f-ad08-ff6d53d113ca\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t56zb" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.666079 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/0fbd23d0-a6b6-4a4e-b418-07fc38e5f497-stats-auth\") pod \"router-default-5444994796-rrhqx\" (UID: \"0fbd23d0-a6b6-4a4e-b418-07fc38e5f497\") " pod="openshift-ingress/router-default-5444994796-rrhqx" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.666101 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b8455642-8313-4539-9729-fe658ef0e4d6-config\") pod \"apiserver-76f77b778f-8xz4m\" (UID: \"b8455642-8313-4539-9729-fe658ef0e4d6\") " pod="openshift-apiserver/apiserver-76f77b778f-8xz4m" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.666126 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6cpv\" (UniqueName: \"kubernetes.io/projected/0fbd23d0-a6b6-4a4e-b418-07fc38e5f497-kube-api-access-c6cpv\") pod \"router-default-5444994796-rrhqx\" (UID: \"0fbd23d0-a6b6-4a4e-b418-07fc38e5f497\") " pod="openshift-ingress/router-default-5444994796-rrhqx" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.666152 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b8455642-8313-4539-9729-fe658ef0e4d6-audit-dir\") pod \"apiserver-76f77b778f-8xz4m\" (UID: \"b8455642-8313-4539-9729-fe658ef0e4d6\") " pod="openshift-apiserver/apiserver-76f77b778f-8xz4m" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.666178 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a68fc0dd-997c-4d12-9a20-5345dfe98a41-serving-cert\") pod \"controller-manager-879f6c89f-cgpdh\" (UID: \"a68fc0dd-997c-4d12-9a20-5345dfe98a41\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cgpdh" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.666204 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7abae297-572a-4c3c-8e7d-4e2157d838ad-metrics-tls\") pod \"dns-operator-744455d44c-gkpck\" (UID: \"7abae297-572a-4c3c-8e7d-4e2157d838ad\") " pod="openshift-dns-operator/dns-operator-744455d44c-gkpck" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.666235 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ec27be85-6d9d-4d96-b3b8-b78a7a941acb-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ddpmx\" (UID: \"ec27be85-6d9d-4d96-b3b8-b78a7a941acb\") " pod="openshift-marketplace/marketplace-operator-79b997595-ddpmx" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.666260 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/b8455642-8313-4539-9729-fe658ef0e4d6-image-import-ca\") pod \"apiserver-76f77b778f-8xz4m\" (UID: \"b8455642-8313-4539-9729-fe658ef0e4d6\") " pod="openshift-apiserver/apiserver-76f77b778f-8xz4m" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.666283 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e4f6934a-db7e-4031-baa8-2f41dc8c2c18-bound-sa-token\") pod \"ingress-operator-5b745b69d9-qfdsg\" (UID: \"e4f6934a-db7e-4031-baa8-2f41dc8c2c18\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qfdsg" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.666312 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e4f6934a-db7e-4031-baa8-2f41dc8c2c18-metrics-tls\") pod \"ingress-operator-5b745b69d9-qfdsg\" (UID: \"e4f6934a-db7e-4031-baa8-2f41dc8c2c18\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qfdsg" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.666318 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b8455642-8313-4539-9729-fe658ef0e4d6-audit-dir\") pod \"apiserver-76f77b778f-8xz4m\" (UID: \"b8455642-8313-4539-9729-fe658ef0e4d6\") " pod="openshift-apiserver/apiserver-76f77b778f-8xz4m" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.666336 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1051665a-7918-487f-ad08-ff6d53d113ca-etcd-client\") pod \"etcd-operator-b45778765-t56zb\" (UID: \"1051665a-7918-487f-ad08-ff6d53d113ca\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t56zb" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.666359 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9454f4b6-e1f9-4dc1-953c-442dbd4088ad-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-x6zsh\" (UID: \"9454f4b6-e1f9-4dc1-953c-442dbd4088ad\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x6zsh" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.666386 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwgtl\" (UniqueName: \"kubernetes.io/projected/f57cc683-84dc-4538-96fe-9a198a5b008f-kube-api-access-zwgtl\") pod \"openshift-config-operator-7777fb866f-s5b86\" (UID: \"f57cc683-84dc-4538-96fe-9a198a5b008f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-s5b86" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.666413 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0fbd23d0-a6b6-4a4e-b418-07fc38e5f497-service-ca-bundle\") pod \"router-default-5444994796-rrhqx\" (UID: \"0fbd23d0-a6b6-4a4e-b418-07fc38e5f497\") " pod="openshift-ingress/router-default-5444994796-rrhqx" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.666436 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r765t\" (UniqueName: \"kubernetes.io/projected/4443d2fe-129a-41eb-b7b7-a7ee9bd3b6eb-kube-api-access-r765t\") pod \"authentication-operator-69f744f599-thckl\" (UID: \"4443d2fe-129a-41eb-b7b7-a7ee9bd3b6eb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-thckl" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.666463 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5t2sf\" (UniqueName: \"kubernetes.io/projected/e4f6934a-db7e-4031-baa8-2f41dc8c2c18-kube-api-access-5t2sf\") pod \"ingress-operator-5b745b69d9-qfdsg\" (UID: \"e4f6934a-db7e-4031-baa8-2f41dc8c2c18\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qfdsg" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.666485 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ff639bee-182b-4dd2-aabc-d396a3bd9ce5-auth-proxy-config\") pod \"machine-config-operator-74547568cd-9rvnv\" (UID: \"ff639bee-182b-4dd2-aabc-d396a3bd9ce5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9rvnv" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.666512 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93ffc5bd-20a8-450c-861a-0aecc58d602b-config\") pod \"machine-approver-56656f9798-fnpz5\" (UID: \"93ffc5bd-20a8-450c-861a-0aecc58d602b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fnpz5" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.666542 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/b8455642-8313-4539-9729-fe658ef0e4d6-audit\") pod \"apiserver-76f77b778f-8xz4m\" (UID: \"b8455642-8313-4539-9729-fe658ef0e4d6\") " pod="openshift-apiserver/apiserver-76f77b778f-8xz4m" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.666588 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e4f6934a-db7e-4031-baa8-2f41dc8c2c18-trusted-ca\") pod \"ingress-operator-5b745b69d9-qfdsg\" (UID: \"e4f6934a-db7e-4031-baa8-2f41dc8c2c18\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qfdsg" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.666611 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f643d9e2-a7df-4aab-a2f5-4bf47363636b-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-jv24g\" (UID: \"f643d9e2-a7df-4aab-a2f5-4bf47363636b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jv24g" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.666633 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4443d2fe-129a-41eb-b7b7-a7ee9bd3b6eb-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-thckl\" (UID: \"4443d2fe-129a-41eb-b7b7-a7ee9bd3b6eb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-thckl" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.666658 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qc82p\" (UniqueName: \"kubernetes.io/projected/ec27be85-6d9d-4d96-b3b8-b78a7a941acb-kube-api-access-qc82p\") pod \"marketplace-operator-79b997595-ddpmx\" (UID: \"ec27be85-6d9d-4d96-b3b8-b78a7a941acb\") " pod="openshift-marketplace/marketplace-operator-79b997595-ddpmx" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.666684 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/93ffc5bd-20a8-450c-861a-0aecc58d602b-machine-approver-tls\") pod \"machine-approver-56656f9798-fnpz5\" (UID: \"93ffc5bd-20a8-450c-861a-0aecc58d602b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fnpz5" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.666702 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbwpp\" (UniqueName: \"kubernetes.io/projected/1051665a-7918-487f-ad08-ff6d53d113ca-kube-api-access-gbwpp\") pod \"etcd-operator-b45778765-t56zb\" (UID: \"1051665a-7918-487f-ad08-ff6d53d113ca\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t56zb" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.666719 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b8455642-8313-4539-9729-fe658ef0e4d6-etcd-client\") pod \"apiserver-76f77b778f-8xz4m\" (UID: \"b8455642-8313-4539-9729-fe658ef0e4d6\") " pod="openshift-apiserver/apiserver-76f77b778f-8xz4m" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.666734 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/b8455642-8313-4539-9729-fe658ef0e4d6-etcd-serving-ca\") pod \"apiserver-76f77b778f-8xz4m\" (UID: \"b8455642-8313-4539-9729-fe658ef0e4d6\") " pod="openshift-apiserver/apiserver-76f77b778f-8xz4m" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.666765 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/b8455642-8313-4539-9729-fe658ef0e4d6-node-pullsecrets\") pod \"apiserver-76f77b778f-8xz4m\" (UID: \"b8455642-8313-4539-9729-fe658ef0e4d6\") " pod="openshift-apiserver/apiserver-76f77b778f-8xz4m" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.666789 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4443d2fe-129a-41eb-b7b7-a7ee9bd3b6eb-config\") pod \"authentication-operator-69f744f599-thckl\" (UID: \"4443d2fe-129a-41eb-b7b7-a7ee9bd3b6eb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-thckl" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.666812 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ff639bee-182b-4dd2-aabc-d396a3bd9ce5-images\") pod \"machine-config-operator-74547568cd-9rvnv\" (UID: \"ff639bee-182b-4dd2-aabc-d396a3bd9ce5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9rvnv" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.666836 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccnj9\" (UniqueName: \"kubernetes.io/projected/a68fc0dd-997c-4d12-9a20-5345dfe98a41-kube-api-access-ccnj9\") pod \"controller-manager-879f6c89f-cgpdh\" (UID: \"a68fc0dd-997c-4d12-9a20-5345dfe98a41\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cgpdh" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.666858 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/f57cc683-84dc-4538-96fe-9a198a5b008f-available-featuregates\") pod \"openshift-config-operator-7777fb866f-s5b86\" (UID: \"f57cc683-84dc-4538-96fe-9a198a5b008f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-s5b86" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.666879 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2n5q\" (UniqueName: \"kubernetes.io/projected/b2994e8f-6482-4dca-baf1-148754e78e85-kube-api-access-l2n5q\") pod \"olm-operator-6b444d44fb-9bfz8\" (UID: \"b2994e8f-6482-4dca-baf1-148754e78e85\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9bfz8" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.666898 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1051665a-7918-487f-ad08-ff6d53d113ca-config\") pod \"etcd-operator-b45778765-t56zb\" (UID: \"1051665a-7918-487f-ad08-ff6d53d113ca\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t56zb" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.666908 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnz55\" (UniqueName: \"kubernetes.io/projected/7abae297-572a-4c3c-8e7d-4e2157d838ad-kube-api-access-xnz55\") pod \"dns-operator-744455d44c-gkpck\" (UID: \"7abae297-572a-4c3c-8e7d-4e2157d838ad\") " pod="openshift-dns-operator/dns-operator-744455d44c-gkpck" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.666914 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/eb15a464-d0e8-4e49-b018-14353cc9d6cb-images\") pod \"machine-api-operator-5694c8668f-9s5r8\" (UID: \"eb15a464-d0e8-4e49-b018-14353cc9d6cb\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9s5r8" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.666944 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/93ffc5bd-20a8-450c-861a-0aecc58d602b-auth-proxy-config\") pod \"machine-approver-56656f9798-fnpz5\" (UID: \"93ffc5bd-20a8-450c-861a-0aecc58d602b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fnpz5" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.666970 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6258654e-b524-4fdd-ba9b-7443b57c2585-config\") pod \"kube-apiserver-operator-766d6c64bb-5rwsv\" (UID: \"6258654e-b524-4fdd-ba9b-7443b57c2585\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5rwsv" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.666993 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9454f4b6-e1f9-4dc1-953c-442dbd4088ad-config\") pod \"kube-controller-manager-operator-78b949d7b-x6zsh\" (UID: \"9454f4b6-e1f9-4dc1-953c-442dbd4088ad\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x6zsh" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.667015 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9172dd32-c65e-4637-ada8-d2c5b0a772ab-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-fqdzj\" (UID: \"9172dd32-c65e-4637-ada8-d2c5b0a772ab\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fqdzj" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.667035 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b8455642-8313-4539-9729-fe658ef0e4d6-serving-cert\") pod \"apiserver-76f77b778f-8xz4m\" (UID: \"b8455642-8313-4539-9729-fe658ef0e4d6\") " pod="openshift-apiserver/apiserver-76f77b778f-8xz4m" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.667057 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ec27be85-6d9d-4d96-b3b8-b78a7a941acb-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ddpmx\" (UID: \"ec27be85-6d9d-4d96-b3b8-b78a7a941acb\") " pod="openshift-marketplace/marketplace-operator-79b997595-ddpmx" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.667081 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6258654e-b524-4fdd-ba9b-7443b57c2585-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-5rwsv\" (UID: \"6258654e-b524-4fdd-ba9b-7443b57c2585\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5rwsv" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.667108 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b2994e8f-6482-4dca-baf1-148754e78e85-srv-cert\") pod \"olm-operator-6b444d44fb-9bfz8\" (UID: \"b2994e8f-6482-4dca-baf1-148754e78e85\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9bfz8" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.667131 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb15a464-d0e8-4e49-b018-14353cc9d6cb-config\") pod \"machine-api-operator-5694c8668f-9s5r8\" (UID: \"eb15a464-d0e8-4e49-b018-14353cc9d6cb\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9s5r8" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.667154 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f57cc683-84dc-4538-96fe-9a198a5b008f-serving-cert\") pod \"openshift-config-operator-7777fb866f-s5b86\" (UID: \"f57cc683-84dc-4538-96fe-9a198a5b008f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-s5b86" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.667182 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a92fd9bb-6d1b-4ab1-9f1a-bc9086ecb08e-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-ppzvl\" (UID: \"a92fd9bb-6d1b-4ab1-9f1a-bc9086ecb08e\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ppzvl" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.667209 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a68fc0dd-997c-4d12-9a20-5345dfe98a41-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-cgpdh\" (UID: \"a68fc0dd-997c-4d12-9a20-5345dfe98a41\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cgpdh" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.667234 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d669k\" (UniqueName: \"kubernetes.io/projected/7621a355-6a10-4143-b35e-f1f617cbf849-kube-api-access-d669k\") pod \"service-ca-9c57cc56f-wrh75\" (UID: \"7621a355-6a10-4143-b35e-f1f617cbf849\") " pod="openshift-service-ca/service-ca-9c57cc56f-wrh75" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.667260 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/0fbd23d0-a6b6-4a4e-b418-07fc38e5f497-default-certificate\") pod \"router-default-5444994796-rrhqx\" (UID: \"0fbd23d0-a6b6-4a4e-b418-07fc38e5f497\") " pod="openshift-ingress/router-default-5444994796-rrhqx" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.667285 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cch2d\" (UniqueName: \"kubernetes.io/projected/93ffc5bd-20a8-450c-861a-0aecc58d602b-kube-api-access-cch2d\") pod \"machine-approver-56656f9798-fnpz5\" (UID: \"93ffc5bd-20a8-450c-861a-0aecc58d602b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fnpz5" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.667311 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a68fc0dd-997c-4d12-9a20-5345dfe98a41-config\") pod \"controller-manager-879f6c89f-cgpdh\" (UID: \"a68fc0dd-997c-4d12-9a20-5345dfe98a41\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cgpdh" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.667336 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/7621a355-6a10-4143-b35e-f1f617cbf849-signing-key\") pod \"service-ca-9c57cc56f-wrh75\" (UID: \"7621a355-6a10-4143-b35e-f1f617cbf849\") " pod="openshift-service-ca/service-ca-9c57cc56f-wrh75" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.667384 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/9172dd32-c65e-4637-ada8-d2c5b0a772ab-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-fqdzj\" (UID: \"9172dd32-c65e-4637-ada8-d2c5b0a772ab\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fqdzj" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.667410 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ff639bee-182b-4dd2-aabc-d396a3bd9ce5-proxy-tls\") pod \"machine-config-operator-74547568cd-9rvnv\" (UID: \"ff639bee-182b-4dd2-aabc-d396a3bd9ce5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9rvnv" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.667432 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a68fc0dd-997c-4d12-9a20-5345dfe98a41-client-ca\") pod \"controller-manager-879f6c89f-cgpdh\" (UID: \"a68fc0dd-997c-4d12-9a20-5345dfe98a41\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cgpdh" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.667456 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/1051665a-7918-487f-ad08-ff6d53d113ca-etcd-service-ca\") pod \"etcd-operator-b45778765-t56zb\" (UID: \"1051665a-7918-487f-ad08-ff6d53d113ca\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t56zb" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.667482 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4443d2fe-129a-41eb-b7b7-a7ee9bd3b6eb-service-ca-bundle\") pod \"authentication-operator-69f744f599-thckl\" (UID: \"4443d2fe-129a-41eb-b7b7-a7ee9bd3b6eb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-thckl" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.667513 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b8455642-8313-4539-9729-fe658ef0e4d6-trusted-ca-bundle\") pod \"apiserver-76f77b778f-8xz4m\" (UID: \"b8455642-8313-4539-9729-fe658ef0e4d6\") " pod="openshift-apiserver/apiserver-76f77b778f-8xz4m" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.667573 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88cj6\" (UniqueName: \"kubernetes.io/projected/1a34c64b-194e-41a6-8b55-0a87187c4cfc-kube-api-access-88cj6\") pod \"control-plane-machine-set-operator-78cbb6b69f-6d29n\" (UID: \"1a34c64b-194e-41a6-8b55-0a87187c4cfc\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6d29n" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.667770 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4443d2fe-129a-41eb-b7b7-a7ee9bd3b6eb-config\") pod \"authentication-operator-69f744f599-thckl\" (UID: \"4443d2fe-129a-41eb-b7b7-a7ee9bd3b6eb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-thckl" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.668063 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e61f4275-03bf-4cb0-bb9a-e519c41f159a-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-jwc8v\" (UID: \"e61f4275-03bf-4cb0-bb9a-e519c41f159a\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jwc8v" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.668150 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a92fd9bb-6d1b-4ab1-9f1a-bc9086ecb08e-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-ppzvl\" (UID: \"a92fd9bb-6d1b-4ab1-9f1a-bc9086ecb08e\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ppzvl" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.668174 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrwfd\" (UniqueName: \"kubernetes.io/projected/ff639bee-182b-4dd2-aabc-d396a3bd9ce5-kube-api-access-hrwfd\") pod \"machine-config-operator-74547568cd-9rvnv\" (UID: \"ff639bee-182b-4dd2-aabc-d396a3bd9ce5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9rvnv" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.668199 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/7621a355-6a10-4143-b35e-f1f617cbf849-signing-cabundle\") pod \"service-ca-9c57cc56f-wrh75\" (UID: \"7621a355-6a10-4143-b35e-f1f617cbf849\") " pod="openshift-service-ca/service-ca-9c57cc56f-wrh75" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.668232 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/eb15a464-d0e8-4e49-b018-14353cc9d6cb-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-9s5r8\" (UID: \"eb15a464-d0e8-4e49-b018-14353cc9d6cb\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9s5r8" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.668252 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4443d2fe-129a-41eb-b7b7-a7ee9bd3b6eb-serving-cert\") pod \"authentication-operator-69f744f599-thckl\" (UID: \"4443d2fe-129a-41eb-b7b7-a7ee9bd3b6eb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-thckl" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.668283 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9454f4b6-e1f9-4dc1-953c-442dbd4088ad-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-x6zsh\" (UID: \"9454f4b6-e1f9-4dc1-953c-442dbd4088ad\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x6zsh" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.668300 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/1051665a-7918-487f-ad08-ff6d53d113ca-etcd-ca\") pod \"etcd-operator-b45778765-t56zb\" (UID: \"1051665a-7918-487f-ad08-ff6d53d113ca\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t56zb" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.668316 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hxvf\" (UniqueName: \"kubernetes.io/projected/e61f4275-03bf-4cb0-bb9a-e519c41f159a-kube-api-access-4hxvf\") pod \"machine-config-controller-84d6567774-jwc8v\" (UID: \"e61f4275-03bf-4cb0-bb9a-e519c41f159a\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jwc8v" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.668337 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwzv6\" (UniqueName: \"kubernetes.io/projected/eb15a464-d0e8-4e49-b018-14353cc9d6cb-kube-api-access-kwzv6\") pod \"machine-api-operator-5694c8668f-9s5r8\" (UID: \"eb15a464-d0e8-4e49-b018-14353cc9d6cb\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9s5r8" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.668469 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9172dd32-c65e-4637-ada8-d2c5b0a772ab-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-fqdzj\" (UID: \"9172dd32-c65e-4637-ada8-d2c5b0a772ab\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fqdzj" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.668523 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b8455642-8313-4539-9729-fe658ef0e4d6-config\") pod \"apiserver-76f77b778f-8xz4m\" (UID: \"b8455642-8313-4539-9729-fe658ef0e4d6\") " pod="openshift-apiserver/apiserver-76f77b778f-8xz4m" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.668571 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/f57cc683-84dc-4538-96fe-9a198a5b008f-available-featuregates\") pod \"openshift-config-operator-7777fb866f-s5b86\" (UID: \"f57cc683-84dc-4538-96fe-9a198a5b008f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-s5b86" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.668902 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.668945 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/93ffc5bd-20a8-450c-861a-0aecc58d602b-auth-proxy-config\") pod \"machine-approver-56656f9798-fnpz5\" (UID: \"93ffc5bd-20a8-450c-861a-0aecc58d602b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fnpz5" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.668145 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb15a464-d0e8-4e49-b018-14353cc9d6cb-config\") pod \"machine-api-operator-5694c8668f-9s5r8\" (UID: \"eb15a464-d0e8-4e49-b018-14353cc9d6cb\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9s5r8" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.669157 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a68fc0dd-997c-4d12-9a20-5345dfe98a41-config\") pod \"controller-manager-879f6c89f-cgpdh\" (UID: \"a68fc0dd-997c-4d12-9a20-5345dfe98a41\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cgpdh" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.669316 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0fbd23d0-a6b6-4a4e-b418-07fc38e5f497-service-ca-bundle\") pod \"router-default-5444994796-rrhqx\" (UID: \"0fbd23d0-a6b6-4a4e-b418-07fc38e5f497\") " pod="openshift-ingress/router-default-5444994796-rrhqx" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.669710 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e61f4275-03bf-4cb0-bb9a-e519c41f159a-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-jwc8v\" (UID: \"e61f4275-03bf-4cb0-bb9a-e519c41f159a\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jwc8v" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.669799 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/b8455642-8313-4539-9729-fe658ef0e4d6-node-pullsecrets\") pod \"apiserver-76f77b778f-8xz4m\" (UID: \"b8455642-8313-4539-9729-fe658ef0e4d6\") " pod="openshift-apiserver/apiserver-76f77b778f-8xz4m" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.669829 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/1051665a-7918-487f-ad08-ff6d53d113ca-etcd-ca\") pod \"etcd-operator-b45778765-t56zb\" (UID: \"1051665a-7918-487f-ad08-ff6d53d113ca\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t56zb" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.669911 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a68fc0dd-997c-4d12-9a20-5345dfe98a41-serving-cert\") pod \"controller-manager-879f6c89f-cgpdh\" (UID: \"a68fc0dd-997c-4d12-9a20-5345dfe98a41\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cgpdh" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.670174 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4443d2fe-129a-41eb-b7b7-a7ee9bd3b6eb-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-thckl\" (UID: \"4443d2fe-129a-41eb-b7b7-a7ee9bd3b6eb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-thckl" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.670348 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/1051665a-7918-487f-ad08-ff6d53d113ca-etcd-service-ca\") pod \"etcd-operator-b45778765-t56zb\" (UID: \"1051665a-7918-487f-ad08-ff6d53d113ca\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t56zb" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.670437 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93ffc5bd-20a8-450c-861a-0aecc58d602b-config\") pod \"machine-approver-56656f9798-fnpz5\" (UID: \"93ffc5bd-20a8-450c-861a-0aecc58d602b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fnpz5" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.670997 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1051665a-7918-487f-ad08-ff6d53d113ca-serving-cert\") pod \"etcd-operator-b45778765-t56zb\" (UID: \"1051665a-7918-487f-ad08-ff6d53d113ca\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t56zb" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.671064 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4443d2fe-129a-41eb-b7b7-a7ee9bd3b6eb-service-ca-bundle\") pod \"authentication-operator-69f744f599-thckl\" (UID: \"4443d2fe-129a-41eb-b7b7-a7ee9bd3b6eb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-thckl" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.671217 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ff639bee-182b-4dd2-aabc-d396a3bd9ce5-auth-proxy-config\") pod \"machine-config-operator-74547568cd-9rvnv\" (UID: \"ff639bee-182b-4dd2-aabc-d396a3bd9ce5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9rvnv" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.671273 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f643d9e2-a7df-4aab-a2f5-4bf47363636b-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-jv24g\" (UID: \"f643d9e2-a7df-4aab-a2f5-4bf47363636b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jv24g" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.671817 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0fbd23d0-a6b6-4a4e-b418-07fc38e5f497-metrics-certs\") pod \"router-default-5444994796-rrhqx\" (UID: \"0fbd23d0-a6b6-4a4e-b418-07fc38e5f497\") " pod="openshift-ingress/router-default-5444994796-rrhqx" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.672011 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/b8455642-8313-4539-9729-fe658ef0e4d6-image-import-ca\") pod \"apiserver-76f77b778f-8xz4m\" (UID: \"b8455642-8313-4539-9729-fe658ef0e4d6\") " pod="openshift-apiserver/apiserver-76f77b778f-8xz4m" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.672414 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/0fbd23d0-a6b6-4a4e-b418-07fc38e5f497-stats-auth\") pod \"router-default-5444994796-rrhqx\" (UID: \"0fbd23d0-a6b6-4a4e-b418-07fc38e5f497\") " pod="openshift-ingress/router-default-5444994796-rrhqx" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.672772 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1051665a-7918-487f-ad08-ff6d53d113ca-etcd-client\") pod \"etcd-operator-b45778765-t56zb\" (UID: \"1051665a-7918-487f-ad08-ff6d53d113ca\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t56zb" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.673253 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a68fc0dd-997c-4d12-9a20-5345dfe98a41-client-ca\") pod \"controller-manager-879f6c89f-cgpdh\" (UID: \"a68fc0dd-997c-4d12-9a20-5345dfe98a41\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cgpdh" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.673708 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/eb15a464-d0e8-4e49-b018-14353cc9d6cb-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-9s5r8\" (UID: \"eb15a464-d0e8-4e49-b018-14353cc9d6cb\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9s5r8" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.673832 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a68fc0dd-997c-4d12-9a20-5345dfe98a41-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-cgpdh\" (UID: \"a68fc0dd-997c-4d12-9a20-5345dfe98a41\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cgpdh" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.674073 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/0fbd23d0-a6b6-4a4e-b418-07fc38e5f497-default-certificate\") pod \"router-default-5444994796-rrhqx\" (UID: \"0fbd23d0-a6b6-4a4e-b418-07fc38e5f497\") " pod="openshift-ingress/router-default-5444994796-rrhqx" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.674263 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b8455642-8313-4539-9729-fe658ef0e4d6-trusted-ca-bundle\") pod \"apiserver-76f77b778f-8xz4m\" (UID: \"b8455642-8313-4539-9729-fe658ef0e4d6\") " pod="openshift-apiserver/apiserver-76f77b778f-8xz4m" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.675321 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4443d2fe-129a-41eb-b7b7-a7ee9bd3b6eb-serving-cert\") pod \"authentication-operator-69f744f599-thckl\" (UID: \"4443d2fe-129a-41eb-b7b7-a7ee9bd3b6eb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-thckl" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.675444 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f643d9e2-a7df-4aab-a2f5-4bf47363636b-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-jv24g\" (UID: \"f643d9e2-a7df-4aab-a2f5-4bf47363636b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jv24g" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.676039 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/93ffc5bd-20a8-450c-861a-0aecc58d602b-machine-approver-tls\") pod \"machine-approver-56656f9798-fnpz5\" (UID: \"93ffc5bd-20a8-450c-861a-0aecc58d602b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fnpz5" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.676332 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/9172dd32-c65e-4637-ada8-d2c5b0a772ab-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-fqdzj\" (UID: \"9172dd32-c65e-4637-ada8-d2c5b0a772ab\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fqdzj" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.676675 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/b8455642-8313-4539-9729-fe658ef0e4d6-etcd-serving-ca\") pod \"apiserver-76f77b778f-8xz4m\" (UID: \"b8455642-8313-4539-9729-fe658ef0e4d6\") " pod="openshift-apiserver/apiserver-76f77b778f-8xz4m" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.676677 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/b8455642-8313-4539-9729-fe658ef0e4d6-encryption-config\") pod \"apiserver-76f77b778f-8xz4m\" (UID: \"b8455642-8313-4539-9729-fe658ef0e4d6\") " pod="openshift-apiserver/apiserver-76f77b778f-8xz4m" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.676792 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/b8455642-8313-4539-9729-fe658ef0e4d6-audit\") pod \"apiserver-76f77b778f-8xz4m\" (UID: \"b8455642-8313-4539-9729-fe658ef0e4d6\") " pod="openshift-apiserver/apiserver-76f77b778f-8xz4m" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.677112 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b8455642-8313-4539-9729-fe658ef0e4d6-etcd-client\") pod \"apiserver-76f77b778f-8xz4m\" (UID: \"b8455642-8313-4539-9729-fe658ef0e4d6\") " pod="openshift-apiserver/apiserver-76f77b778f-8xz4m" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.677708 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b8455642-8313-4539-9729-fe658ef0e4d6-serving-cert\") pod \"apiserver-76f77b778f-8xz4m\" (UID: \"b8455642-8313-4539-9729-fe658ef0e4d6\") " pod="openshift-apiserver/apiserver-76f77b778f-8xz4m" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.688976 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.708894 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.721609 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7abae297-572a-4c3c-8e7d-4e2157d838ad-metrics-tls\") pod \"dns-operator-744455d44c-gkpck\" (UID: \"7abae297-572a-4c3c-8e7d-4e2157d838ad\") " pod="openshift-dns-operator/dns-operator-744455d44c-gkpck" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.728320 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.748651 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.769364 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/7621a355-6a10-4143-b35e-f1f617cbf849-signing-cabundle\") pod \"service-ca-9c57cc56f-wrh75\" (UID: \"7621a355-6a10-4143-b35e-f1f617cbf849\") " pod="openshift-service-ca/service-ca-9c57cc56f-wrh75" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.769501 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ec27be85-6d9d-4d96-b3b8-b78a7a941acb-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ddpmx\" (UID: \"ec27be85-6d9d-4d96-b3b8-b78a7a941acb\") " pod="openshift-marketplace/marketplace-operator-79b997595-ddpmx" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.769574 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qc82p\" (UniqueName: \"kubernetes.io/projected/ec27be85-6d9d-4d96-b3b8-b78a7a941acb-kube-api-access-qc82p\") pod \"marketplace-operator-79b997595-ddpmx\" (UID: \"ec27be85-6d9d-4d96-b3b8-b78a7a941acb\") " pod="openshift-marketplace/marketplace-operator-79b997595-ddpmx" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.769653 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ec27be85-6d9d-4d96-b3b8-b78a7a941acb-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ddpmx\" (UID: \"ec27be85-6d9d-4d96-b3b8-b78a7a941acb\") " pod="openshift-marketplace/marketplace-operator-79b997595-ddpmx" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.769706 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d669k\" (UniqueName: \"kubernetes.io/projected/7621a355-6a10-4143-b35e-f1f617cbf849-kube-api-access-d669k\") pod \"service-ca-9c57cc56f-wrh75\" (UID: \"7621a355-6a10-4143-b35e-f1f617cbf849\") " pod="openshift-service-ca/service-ca-9c57cc56f-wrh75" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.769740 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/7621a355-6a10-4143-b35e-f1f617cbf849-signing-key\") pod \"service-ca-9c57cc56f-wrh75\" (UID: \"7621a355-6a10-4143-b35e-f1f617cbf849\") " pod="openshift-service-ca/service-ca-9c57cc56f-wrh75" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.771340 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.788945 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.793169 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f57cc683-84dc-4538-96fe-9a198a5b008f-serving-cert\") pod \"openshift-config-operator-7777fb866f-s5b86\" (UID: \"f57cc683-84dc-4538-96fe-9a198a5b008f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-s5b86" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.810122 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.830202 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.864457 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.871028 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.881644 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.882004 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6258654e-b524-4fdd-ba9b-7443b57c2585-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-5rwsv\" (UID: \"6258654e-b524-4fdd-ba9b-7443b57c2585\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5rwsv" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.889330 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.899871 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6258654e-b524-4fdd-ba9b-7443b57c2585-config\") pod \"kube-apiserver-operator-766d6c64bb-5rwsv\" (UID: \"6258654e-b524-4fdd-ba9b-7443b57c2585\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5rwsv" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.909514 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.929045 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.949585 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.963178 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9454f4b6-e1f9-4dc1-953c-442dbd4088ad-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-x6zsh\" (UID: \"9454f4b6-e1f9-4dc1-953c-442dbd4088ad\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x6zsh" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.968942 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.969725 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9454f4b6-e1f9-4dc1-953c-442dbd4088ad-config\") pod \"kube-controller-manager-operator-78b949d7b-x6zsh\" (UID: \"9454f4b6-e1f9-4dc1-953c-442dbd4088ad\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x6zsh" Dec 06 03:42:32 crc kubenswrapper[4802]: I1206 03:42:32.989234 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.002032 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e4f6934a-db7e-4031-baa8-2f41dc8c2c18-metrics-tls\") pod \"ingress-operator-5b745b69d9-qfdsg\" (UID: \"e4f6934a-db7e-4031-baa8-2f41dc8c2c18\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qfdsg" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.009464 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.032841 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.042711 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ff639bee-182b-4dd2-aabc-d396a3bd9ce5-proxy-tls\") pod \"machine-config-operator-74547568cd-9rvnv\" (UID: \"ff639bee-182b-4dd2-aabc-d396a3bd9ce5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9rvnv" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.049806 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.079779 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.089422 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.090976 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e4f6934a-db7e-4031-baa8-2f41dc8c2c18-trusted-ca\") pod \"ingress-operator-5b745b69d9-qfdsg\" (UID: \"e4f6934a-db7e-4031-baa8-2f41dc8c2c18\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qfdsg" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.109086 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.119136 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ff639bee-182b-4dd2-aabc-d396a3bd9ce5-images\") pod \"machine-config-operator-74547568cd-9rvnv\" (UID: \"ff639bee-182b-4dd2-aabc-d396a3bd9ce5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9rvnv" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.129265 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.149500 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.159682 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a92fd9bb-6d1b-4ab1-9f1a-bc9086ecb08e-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-ppzvl\" (UID: \"a92fd9bb-6d1b-4ab1-9f1a-bc9086ecb08e\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ppzvl" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.169722 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.189572 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.204222 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a92fd9bb-6d1b-4ab1-9f1a-bc9086ecb08e-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-ppzvl\" (UID: \"a92fd9bb-6d1b-4ab1-9f1a-bc9086ecb08e\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ppzvl" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.210896 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.229740 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.249255 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.261738 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/1a34c64b-194e-41a6-8b55-0a87187c4cfc-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-6d29n\" (UID: \"1a34c64b-194e-41a6-8b55-0a87187c4cfc\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6d29n" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.269799 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.310287 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.329494 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.348970 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.360500 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b2994e8f-6482-4dca-baf1-148754e78e85-profile-collector-cert\") pod \"olm-operator-6b444d44fb-9bfz8\" (UID: \"b2994e8f-6482-4dca-baf1-148754e78e85\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9bfz8" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.369319 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.389201 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.392730 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b2994e8f-6482-4dca-baf1-148754e78e85-srv-cert\") pod \"olm-operator-6b444d44fb-9bfz8\" (UID: \"b2994e8f-6482-4dca-baf1-148754e78e85\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9bfz8" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.409443 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.429021 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.441995 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e61f4275-03bf-4cb0-bb9a-e519c41f159a-proxy-tls\") pod \"machine-config-controller-84d6567774-jwc8v\" (UID: \"e61f4275-03bf-4cb0-bb9a-e519c41f159a\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jwc8v" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.450181 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.469575 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.489686 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.509334 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.530746 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.551014 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.568045 4802 request.go:700] Waited for 1.019161698s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-apiserver-operator/secrets?fieldSelector=metadata.name%3Dopenshift-apiserver-operator-dockercfg-xtcjv&limit=500&resourceVersion=0 Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.571016 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.589516 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.611076 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.630254 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.648903 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.670140 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.689157 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.710527 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.750115 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.759583 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-992nr\" (UniqueName: \"kubernetes.io/projected/3f3f6253-533f-4d89-ac10-f26eca1ac390-kube-api-access-992nr\") pod \"console-operator-58897d9998-4lqzs\" (UID: \"3f3f6253-533f-4d89-ac10-f26eca1ac390\") " pod="openshift-console-operator/console-operator-58897d9998-4lqzs" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.769547 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 06 03:42:33 crc kubenswrapper[4802]: E1206 03:42:33.769891 4802 configmap.go:193] Couldn't get configMap openshift-marketplace/marketplace-trusted-ca: failed to sync configmap cache: timed out waiting for the condition Dec 06 03:42:33 crc kubenswrapper[4802]: E1206 03:42:33.769967 4802 configmap.go:193] Couldn't get configMap openshift-service-ca/signing-cabundle: failed to sync configmap cache: timed out waiting for the condition Dec 06 03:42:33 crc kubenswrapper[4802]: E1206 03:42:33.769929 4802 secret.go:188] Couldn't get secret openshift-marketplace/marketplace-operator-metrics: failed to sync secret cache: timed out waiting for the condition Dec 06 03:42:33 crc kubenswrapper[4802]: E1206 03:42:33.770004 4802 secret.go:188] Couldn't get secret openshift-service-ca/signing-key: failed to sync secret cache: timed out waiting for the condition Dec 06 03:42:33 crc kubenswrapper[4802]: E1206 03:42:33.770033 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/ec27be85-6d9d-4d96-b3b8-b78a7a941acb-marketplace-trusted-ca podName:ec27be85-6d9d-4d96-b3b8-b78a7a941acb nodeName:}" failed. No retries permitted until 2025-12-06 03:42:34.270003156 +0000 UTC m=+147.141912348 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "marketplace-trusted-ca" (UniqueName: "kubernetes.io/configmap/ec27be85-6d9d-4d96-b3b8-b78a7a941acb-marketplace-trusted-ca") pod "marketplace-operator-79b997595-ddpmx" (UID: "ec27be85-6d9d-4d96-b3b8-b78a7a941acb") : failed to sync configmap cache: timed out waiting for the condition Dec 06 03:42:33 crc kubenswrapper[4802]: E1206 03:42:33.770173 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/7621a355-6a10-4143-b35e-f1f617cbf849-signing-cabundle podName:7621a355-6a10-4143-b35e-f1f617cbf849 nodeName:}" failed. No retries permitted until 2025-12-06 03:42:34.270130469 +0000 UTC m=+147.142039741 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "signing-cabundle" (UniqueName: "kubernetes.io/configmap/7621a355-6a10-4143-b35e-f1f617cbf849-signing-cabundle") pod "service-ca-9c57cc56f-wrh75" (UID: "7621a355-6a10-4143-b35e-f1f617cbf849") : failed to sync configmap cache: timed out waiting for the condition Dec 06 03:42:33 crc kubenswrapper[4802]: E1206 03:42:33.770222 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ec27be85-6d9d-4d96-b3b8-b78a7a941acb-marketplace-operator-metrics podName:ec27be85-6d9d-4d96-b3b8-b78a7a941acb nodeName:}" failed. No retries permitted until 2025-12-06 03:42:34.270203541 +0000 UTC m=+147.142112883 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "marketplace-operator-metrics" (UniqueName: "kubernetes.io/secret/ec27be85-6d9d-4d96-b3b8-b78a7a941acb-marketplace-operator-metrics") pod "marketplace-operator-79b997595-ddpmx" (UID: "ec27be85-6d9d-4d96-b3b8-b78a7a941acb") : failed to sync secret cache: timed out waiting for the condition Dec 06 03:42:33 crc kubenswrapper[4802]: E1206 03:42:33.770269 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7621a355-6a10-4143-b35e-f1f617cbf849-signing-key podName:7621a355-6a10-4143-b35e-f1f617cbf849 nodeName:}" failed. No retries permitted until 2025-12-06 03:42:34.270249603 +0000 UTC m=+147.142158925 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "signing-key" (UniqueName: "kubernetes.io/secret/7621a355-6a10-4143-b35e-f1f617cbf849-signing-key") pod "service-ca-9c57cc56f-wrh75" (UID: "7621a355-6a10-4143-b35e-f1f617cbf849") : failed to sync secret cache: timed out waiting for the condition Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.790456 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.834422 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgrht\" (UniqueName: \"kubernetes.io/projected/263721d5-fcca-411a-876e-0acfe0863d2d-kube-api-access-bgrht\") pod \"oauth-openshift-558db77b4-rqs9q\" (UID: \"263721d5-fcca-411a-876e-0acfe0863d2d\") " pod="openshift-authentication/oauth-openshift-558db77b4-rqs9q" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.857978 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpqvr\" (UniqueName: \"kubernetes.io/projected/db213df1-bfa2-4edd-a3fc-246e35e20585-kube-api-access-fpqvr\") pod \"route-controller-manager-6576b87f9c-9v855\" (UID: \"db213df1-bfa2-4edd-a3fc-246e35e20585\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9v855" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.878650 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfqp4\" (UniqueName: \"kubernetes.io/projected/dc4fc98a-badd-46dc-b700-6b01aff112b4-kube-api-access-lfqp4\") pod \"cluster-samples-operator-665b6dd947-ghcrj\" (UID: \"dc4fc98a-badd-46dc-b700-6b01aff112b4\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ghcrj" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.898582 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbrbz\" (UniqueName: \"kubernetes.io/projected/2362780e-93d1-4f97-bda0-138ed5180bb0-kube-api-access-qbrbz\") pod \"console-f9d7485db-7ckfj\" (UID: \"2362780e-93d1-4f97-bda0-138ed5180bb0\") " pod="openshift-console/console-f9d7485db-7ckfj" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.909047 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.912900 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6w6b\" (UniqueName: \"kubernetes.io/projected/bd64a08a-6920-4a0f-b299-aa5a009236a1-kube-api-access-g6w6b\") pod \"apiserver-7bbb656c7d-8br44\" (UID: \"bd64a08a-6920-4a0f-b299-aa5a009236a1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8br44" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.929994 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.946050 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-7ckfj" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.949450 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.969319 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.975821 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-4lqzs" Dec 06 03:42:33 crc kubenswrapper[4802]: I1206 03:42:33.990145 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.004551 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ghcrj" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.009361 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.029179 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.049569 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-rqs9q" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.050166 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.057687 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9v855" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.089224 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.092162 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.109473 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.113339 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8br44" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.131455 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.148893 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.169225 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.189583 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.208788 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.218499 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ghcrj"] Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.228715 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.249414 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.266575 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-rqs9q"] Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.269123 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.282356 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-9v855"] Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.290136 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.300117 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ec27be85-6d9d-4d96-b3b8-b78a7a941acb-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ddpmx\" (UID: \"ec27be85-6d9d-4d96-b3b8-b78a7a941acb\") " pod="openshift-marketplace/marketplace-operator-79b997595-ddpmx" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.300227 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ec27be85-6d9d-4d96-b3b8-b78a7a941acb-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ddpmx\" (UID: \"ec27be85-6d9d-4d96-b3b8-b78a7a941acb\") " pod="openshift-marketplace/marketplace-operator-79b997595-ddpmx" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.300266 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/7621a355-6a10-4143-b35e-f1f617cbf849-signing-key\") pod \"service-ca-9c57cc56f-wrh75\" (UID: \"7621a355-6a10-4143-b35e-f1f617cbf849\") " pod="openshift-service-ca/service-ca-9c57cc56f-wrh75" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.302271 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ec27be85-6d9d-4d96-b3b8-b78a7a941acb-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ddpmx\" (UID: \"ec27be85-6d9d-4d96-b3b8-b78a7a941acb\") " pod="openshift-marketplace/marketplace-operator-79b997595-ddpmx" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.303524 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/7621a355-6a10-4143-b35e-f1f617cbf849-signing-cabundle\") pod \"service-ca-9c57cc56f-wrh75\" (UID: \"7621a355-6a10-4143-b35e-f1f617cbf849\") " pod="openshift-service-ca/service-ca-9c57cc56f-wrh75" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.306385 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/7621a355-6a10-4143-b35e-f1f617cbf849-signing-key\") pod \"service-ca-9c57cc56f-wrh75\" (UID: \"7621a355-6a10-4143-b35e-f1f617cbf849\") " pod="openshift-service-ca/service-ca-9c57cc56f-wrh75" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.307020 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/7621a355-6a10-4143-b35e-f1f617cbf849-signing-cabundle\") pod \"service-ca-9c57cc56f-wrh75\" (UID: \"7621a355-6a10-4143-b35e-f1f617cbf849\") " pod="openshift-service-ca/service-ca-9c57cc56f-wrh75" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.307046 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ec27be85-6d9d-4d96-b3b8-b78a7a941acb-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ddpmx\" (UID: \"ec27be85-6d9d-4d96-b3b8-b78a7a941acb\") " pod="openshift-marketplace/marketplace-operator-79b997595-ddpmx" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.308613 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.329289 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.335475 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-8br44"] Dec 06 03:42:34 crc kubenswrapper[4802]: W1206 03:42:34.340587 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbd64a08a_6920_4a0f_b299_aa5a009236a1.slice/crio-6f74bef80421a37cf89bf1b82b8cd5846ae780be472b5d9490dac48128384a12 WatchSource:0}: Error finding container 6f74bef80421a37cf89bf1b82b8cd5846ae780be472b5d9490dac48128384a12: Status 404 returned error can't find the container with id 6f74bef80421a37cf89bf1b82b8cd5846ae780be472b5d9490dac48128384a12 Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.348388 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.369239 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.377695 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-4lqzs"] Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.379573 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-7ckfj"] Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.389428 4802 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.405207 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.409058 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.409978 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.428685 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.449287 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.469130 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.505720 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:42:34 crc kubenswrapper[4802]: E1206 03:42:34.505895 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:44:36.505865541 +0000 UTC m=+269.377774763 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.505965 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.506019 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.506145 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.508865 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.509341 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.509411 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.509460 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.529522 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.548800 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.570966 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.587167 4802 request.go:700] Waited for 1.921302233s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-storage-version-migrator-operator/serviceaccounts/kube-storage-version-migrator-operator/token Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.610275 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwm24\" (UniqueName: \"kubernetes.io/projected/a92fd9bb-6d1b-4ab1-9f1a-bc9086ecb08e-kube-api-access-fwm24\") pod \"kube-storage-version-migrator-operator-b67b599dd-ppzvl\" (UID: \"a92fd9bb-6d1b-4ab1-9f1a-bc9086ecb08e\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ppzvl" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.627628 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dvdr\" (UniqueName: \"kubernetes.io/projected/0057cff6-6085-4b41-bf4f-e1121bece99a-kube-api-access-7dvdr\") pod \"downloads-7954f5f757-7cg8t\" (UID: \"0057cff6-6085-4b41-bf4f-e1121bece99a\") " pod="openshift-console/downloads-7954f5f757-7cg8t" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.643361 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plz2r\" (UniqueName: \"kubernetes.io/projected/b8455642-8313-4539-9729-fe658ef0e4d6-kube-api-access-plz2r\") pod \"apiserver-76f77b778f-8xz4m\" (UID: \"b8455642-8313-4539-9729-fe658ef0e4d6\") " pod="openshift-apiserver/apiserver-76f77b778f-8xz4m" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.669956 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.672896 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9172dd32-c65e-4637-ada8-d2c5b0a772ab-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-fqdzj\" (UID: \"9172dd32-c65e-4637-ada8-d2c5b0a772ab\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fqdzj" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.674938 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.678888 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.699588 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxx8q\" (UniqueName: \"kubernetes.io/projected/9172dd32-c65e-4637-ada8-d2c5b0a772ab-kube-api-access-rxx8q\") pod \"cluster-image-registry-operator-dc59b4c8b-fqdzj\" (UID: \"9172dd32-c65e-4637-ada8-d2c5b0a772ab\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fqdzj" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.710565 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6cpv\" (UniqueName: \"kubernetes.io/projected/0fbd23d0-a6b6-4a4e-b418-07fc38e5f497-kube-api-access-c6cpv\") pod \"router-default-5444994796-rrhqx\" (UID: \"0fbd23d0-a6b6-4a4e-b418-07fc38e5f497\") " pod="openshift-ingress/router-default-5444994796-rrhqx" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.727190 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9bfx\" (UniqueName: \"kubernetes.io/projected/f643d9e2-a7df-4aab-a2f5-4bf47363636b-kube-api-access-m9bfx\") pod \"openshift-controller-manager-operator-756b6f6bc6-jv24g\" (UID: \"f643d9e2-a7df-4aab-a2f5-4bf47363636b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jv24g" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.733087 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-8xz4m" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.743476 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbwpp\" (UniqueName: \"kubernetes.io/projected/1051665a-7918-487f-ad08-ff6d53d113ca-kube-api-access-gbwpp\") pod \"etcd-operator-b45778765-t56zb\" (UID: \"1051665a-7918-487f-ad08-ff6d53d113ca\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t56zb" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.744068 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-7cg8t" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.753056 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jv24g" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.765668 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r765t\" (UniqueName: \"kubernetes.io/projected/4443d2fe-129a-41eb-b7b7-a7ee9bd3b6eb-kube-api-access-r765t\") pod \"authentication-operator-69f744f599-thckl\" (UID: \"4443d2fe-129a-41eb-b7b7-a7ee9bd3b6eb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-thckl" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.785462 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5t2sf\" (UniqueName: \"kubernetes.io/projected/e4f6934a-db7e-4031-baa8-2f41dc8c2c18-kube-api-access-5t2sf\") pod \"ingress-operator-5b745b69d9-qfdsg\" (UID: \"e4f6934a-db7e-4031-baa8-2f41dc8c2c18\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qfdsg" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.803914 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fqdzj" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.808134 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e4f6934a-db7e-4031-baa8-2f41dc8c2c18-bound-sa-token\") pod \"ingress-operator-5b745b69d9-qfdsg\" (UID: \"e4f6934a-db7e-4031-baa8-2f41dc8c2c18\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qfdsg" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.823132 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-t56zb" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.823231 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-rrhqx" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.827636 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cch2d\" (UniqueName: \"kubernetes.io/projected/93ffc5bd-20a8-450c-861a-0aecc58d602b-kube-api-access-cch2d\") pod \"machine-approver-56656f9798-fnpz5\" (UID: \"93ffc5bd-20a8-450c-861a-0aecc58d602b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fnpz5" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.848998 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2n5q\" (UniqueName: \"kubernetes.io/projected/b2994e8f-6482-4dca-baf1-148754e78e85-kube-api-access-l2n5q\") pod \"olm-operator-6b444d44fb-9bfz8\" (UID: \"b2994e8f-6482-4dca-baf1-148754e78e85\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9bfz8" Dec 06 03:42:34 crc kubenswrapper[4802]: W1206 03:42:34.849675 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0fbd23d0_a6b6_4a4e_b418_07fc38e5f497.slice/crio-742c61214d18c9e2cde5cb26ab7a28b543a6ed9d15b104c59ac000d401b30b6b WatchSource:0}: Error finding container 742c61214d18c9e2cde5cb26ab7a28b543a6ed9d15b104c59ac000d401b30b6b: Status 404 returned error can't find the container with id 742c61214d18c9e2cde5cb26ab7a28b543a6ed9d15b104c59ac000d401b30b6b Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.857529 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qfdsg" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.871664 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccnj9\" (UniqueName: \"kubernetes.io/projected/a68fc0dd-997c-4d12-9a20-5345dfe98a41-kube-api-access-ccnj9\") pod \"controller-manager-879f6c89f-cgpdh\" (UID: \"a68fc0dd-997c-4d12-9a20-5345dfe98a41\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cgpdh" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.873266 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ppzvl" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.888868 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9bfz8" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.896142 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnz55\" (UniqueName: \"kubernetes.io/projected/7abae297-572a-4c3c-8e7d-4e2157d838ad-kube-api-access-xnz55\") pod \"dns-operator-744455d44c-gkpck\" (UID: \"7abae297-572a-4c3c-8e7d-4e2157d838ad\") " pod="openshift-dns-operator/dns-operator-744455d44c-gkpck" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.906702 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwzv6\" (UniqueName: \"kubernetes.io/projected/eb15a464-d0e8-4e49-b018-14353cc9d6cb-kube-api-access-kwzv6\") pod \"machine-api-operator-5694c8668f-9s5r8\" (UID: \"eb15a464-d0e8-4e49-b018-14353cc9d6cb\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9s5r8" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.923286 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6258654e-b524-4fdd-ba9b-7443b57c2585-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-5rwsv\" (UID: \"6258654e-b524-4fdd-ba9b-7443b57c2585\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5rwsv" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.940719 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwgtl\" (UniqueName: \"kubernetes.io/projected/f57cc683-84dc-4538-96fe-9a198a5b008f-kube-api-access-zwgtl\") pod \"openshift-config-operator-7777fb866f-s5b86\" (UID: \"f57cc683-84dc-4538-96fe-9a198a5b008f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-s5b86" Dec 06 03:42:34 crc kubenswrapper[4802]: W1206 03:42:34.946391 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-e1159f5fe043cff6aa21d99bd07176091496b70fc5a3635e1907cb3743635dd0 WatchSource:0}: Error finding container e1159f5fe043cff6aa21d99bd07176091496b70fc5a3635e1907cb3743635dd0: Status 404 returned error can't find the container with id e1159f5fe043cff6aa21d99bd07176091496b70fc5a3635e1907cb3743635dd0 Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.965354 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88cj6\" (UniqueName: \"kubernetes.io/projected/1a34c64b-194e-41a6-8b55-0a87187c4cfc-kube-api-access-88cj6\") pod \"control-plane-machine-set-operator-78cbb6b69f-6d29n\" (UID: \"1a34c64b-194e-41a6-8b55-0a87187c4cfc\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6d29n" Dec 06 03:42:34 crc kubenswrapper[4802]: I1206 03:42:34.988673 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hxvf\" (UniqueName: \"kubernetes.io/projected/e61f4275-03bf-4cb0-bb9a-e519c41f159a-kube-api-access-4hxvf\") pod \"machine-config-controller-84d6567774-jwc8v\" (UID: \"e61f4275-03bf-4cb0-bb9a-e519c41f159a\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jwc8v" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.004173 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9454f4b6-e1f9-4dc1-953c-442dbd4088ad-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-x6zsh\" (UID: \"9454f4b6-e1f9-4dc1-953c-442dbd4088ad\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x6zsh" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.023005 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fnpz5" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.023711 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hrwfd\" (UniqueName: \"kubernetes.io/projected/ff639bee-182b-4dd2-aabc-d396a3bd9ce5-kube-api-access-hrwfd\") pod \"machine-config-operator-74547568cd-9rvnv\" (UID: \"ff639bee-182b-4dd2-aabc-d396a3bd9ce5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9rvnv" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.045812 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qc82p\" (UniqueName: \"kubernetes.io/projected/ec27be85-6d9d-4d96-b3b8-b78a7a941acb-kube-api-access-qc82p\") pod \"marketplace-operator-79b997595-ddpmx\" (UID: \"ec27be85-6d9d-4d96-b3b8-b78a7a941acb\") " pod="openshift-marketplace/marketplace-operator-79b997595-ddpmx" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.062209 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-thckl" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.068887 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d669k\" (UniqueName: \"kubernetes.io/projected/7621a355-6a10-4143-b35e-f1f617cbf849-kube-api-access-d669k\") pod \"service-ca-9c57cc56f-wrh75\" (UID: \"7621a355-6a10-4143-b35e-f1f617cbf849\") " pod="openshift-service-ca/service-ca-9c57cc56f-wrh75" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.092410 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-9s5r8" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.097548 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-cgpdh" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.113980 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mzbrn\" (UniqueName: \"kubernetes.io/projected/a668f1b9-b7e6-4c39-9fc2-9beb5e5e5bce-kube-api-access-mzbrn\") pod \"packageserver-d55dfcdfc-2nslt\" (UID: \"a668f1b9-b7e6-4c39-9fc2-9beb5e5e5bce\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2nslt" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.114023 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2610e572-bd43-4029-a846-b6694c1896f7-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-mpnf2\" (UID: \"2610e572-bd43-4029-a846-b6694c1896f7\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mpnf2" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.114041 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9znd\" (UniqueName: \"kubernetes.io/projected/18caf043-1bf2-4e67-bb3b-e4edf9fc8822-kube-api-access-h9znd\") pod \"package-server-manager-789f6589d5-wbjzw\" (UID: \"18caf043-1bf2-4e67-bb3b-e4edf9fc8822\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wbjzw" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.114120 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bj479\" (UniqueName: \"kubernetes.io/projected/8002be53-a1e5-4ef7-93b4-384d9b16a7e2-kube-api-access-bj479\") pod \"migrator-59844c95c7-zfjwx\" (UID: \"8002be53-a1e5-4ef7-93b4-384d9b16a7e2\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-zfjwx" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.114143 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c2adc97c-46b4-4de7-b49c-5f75a16c2c32-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-gmvnz\" (UID: \"c2adc97c-46b4-4de7-b49c-5f75a16c2c32\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gmvnz" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.114204 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a668f1b9-b7e6-4c39-9fc2-9beb5e5e5bce-webhook-cert\") pod \"packageserver-d55dfcdfc-2nslt\" (UID: \"a668f1b9-b7e6-4c39-9fc2-9beb5e5e5bce\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2nslt" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.114233 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d13a541e-f90d-4222-8486-2d47854ac8e5-installation-pull-secrets\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.114341 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/407d6ced-2671-40de-aa70-153cfda69fba-csi-data-dir\") pod \"csi-hostpathplugin-9sxg5\" (UID: \"407d6ced-2671-40de-aa70-153cfda69fba\") " pod="hostpath-provisioner/csi-hostpathplugin-9sxg5" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.114380 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbc8f\" (UniqueName: \"kubernetes.io/projected/19ade530-d0de-44d1-be83-96136753a35e-kube-api-access-pbc8f\") pod \"collect-profiles-29416530-88l5w\" (UID: \"19ade530-d0de-44d1-be83-96136753a35e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416530-88l5w" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.114395 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d13a541e-f90d-4222-8486-2d47854ac8e5-registry-tls\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.114411 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f7feaf6a-2ad5-4ef3-88d1-a2a483e5f0d9-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-ntmms\" (UID: \"f7feaf6a-2ad5-4ef3-88d1-a2a483e5f0d9\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ntmms" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.114428 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kcmz\" (UniqueName: \"kubernetes.io/projected/293e299a-d25d-4755-8d9d-4d8f224c8791-kube-api-access-4kcmz\") pod \"machine-config-server-khkwh\" (UID: \"293e299a-d25d-4755-8d9d-4d8f224c8791\") " pod="openshift-machine-config-operator/machine-config-server-khkwh" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.114443 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pflwz\" (UniqueName: \"kubernetes.io/projected/78eab4ec-5f1e-42cd-8ce7-48f76a55cc95-kube-api-access-pflwz\") pod \"catalog-operator-68c6474976-4gkmd\" (UID: \"78eab4ec-5f1e-42cd-8ce7-48f76a55cc95\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4gkmd" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.114460 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ddb4b244-e4e0-4ae4-a217-c6531fc90b6a-config-volume\") pod \"dns-default-whhs8\" (UID: \"ddb4b244-e4e0-4ae4-a217-c6531fc90b6a\") " pod="openshift-dns/dns-default-whhs8" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.114503 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/19ade530-d0de-44d1-be83-96136753a35e-secret-volume\") pod \"collect-profiles-29416530-88l5w\" (UID: \"19ade530-d0de-44d1-be83-96136753a35e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416530-88l5w" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.114525 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b78b4d0e-a83c-4339-b1c4-8998e69e7218-serving-cert\") pod \"service-ca-operator-777779d784-zb4m6\" (UID: \"b78b4d0e-a83c-4339-b1c4-8998e69e7218\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-zb4m6" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.114564 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/407d6ced-2671-40de-aa70-153cfda69fba-plugins-dir\") pod \"csi-hostpathplugin-9sxg5\" (UID: \"407d6ced-2671-40de-aa70-153cfda69fba\") " pod="hostpath-provisioner/csi-hostpathplugin-9sxg5" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.114621 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/407d6ced-2671-40de-aa70-153cfda69fba-registration-dir\") pod \"csi-hostpathplugin-9sxg5\" (UID: \"407d6ced-2671-40de-aa70-153cfda69fba\") " pod="hostpath-provisioner/csi-hostpathplugin-9sxg5" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.114640 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b78b4d0e-a83c-4339-b1c4-8998e69e7218-config\") pod \"service-ca-operator-777779d784-zb4m6\" (UID: \"b78b4d0e-a83c-4339-b1c4-8998e69e7218\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-zb4m6" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.114684 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzkg4\" (UniqueName: \"kubernetes.io/projected/d13a541e-f90d-4222-8486-2d47854ac8e5-kube-api-access-tzkg4\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.114726 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/78eab4ec-5f1e-42cd-8ce7-48f76a55cc95-srv-cert\") pod \"catalog-operator-68c6474976-4gkmd\" (UID: \"78eab4ec-5f1e-42cd-8ce7-48f76a55cc95\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4gkmd" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.114811 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgrhd\" (UniqueName: \"kubernetes.io/projected/ddb4b244-e4e0-4ae4-a217-c6531fc90b6a-kube-api-access-lgrhd\") pod \"dns-default-whhs8\" (UID: \"ddb4b244-e4e0-4ae4-a217-c6531fc90b6a\") " pod="openshift-dns/dns-default-whhs8" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.114843 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2610e572-bd43-4029-a846-b6694c1896f7-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-mpnf2\" (UID: \"2610e572-bd43-4029-a846-b6694c1896f7\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mpnf2" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.114862 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2610e572-bd43-4029-a846-b6694c1896f7-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-mpnf2\" (UID: \"2610e572-bd43-4029-a846-b6694c1896f7\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mpnf2" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.114883 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wl6sm\" (UniqueName: \"kubernetes.io/projected/b78b4d0e-a83c-4339-b1c4-8998e69e7218-kube-api-access-wl6sm\") pod \"service-ca-operator-777779d784-zb4m6\" (UID: \"b78b4d0e-a83c-4339-b1c4-8998e69e7218\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-zb4m6" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.114916 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a668f1b9-b7e6-4c39-9fc2-9beb5e5e5bce-apiservice-cert\") pod \"packageserver-d55dfcdfc-2nslt\" (UID: \"a668f1b9-b7e6-4c39-9fc2-9beb5e5e5bce\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2nslt" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.114989 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/293e299a-d25d-4755-8d9d-4d8f224c8791-node-bootstrap-token\") pod \"machine-config-server-khkwh\" (UID: \"293e299a-d25d-4755-8d9d-4d8f224c8791\") " pod="openshift-machine-config-operator/machine-config-server-khkwh" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.115007 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/78eab4ec-5f1e-42cd-8ce7-48f76a55cc95-profile-collector-cert\") pod \"catalog-operator-68c6474976-4gkmd\" (UID: \"78eab4ec-5f1e-42cd-8ce7-48f76a55cc95\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4gkmd" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.115054 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ddb4b244-e4e0-4ae4-a217-c6531fc90b6a-metrics-tls\") pod \"dns-default-whhs8\" (UID: \"ddb4b244-e4e0-4ae4-a217-c6531fc90b6a\") " pod="openshift-dns/dns-default-whhs8" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.115071 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d13a541e-f90d-4222-8486-2d47854ac8e5-registry-certificates\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.115103 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d13a541e-f90d-4222-8486-2d47854ac8e5-ca-trust-extracted\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.115164 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dlgh\" (UniqueName: \"kubernetes.io/projected/c2adc97c-46b4-4de7-b49c-5f75a16c2c32-kube-api-access-5dlgh\") pod \"openshift-apiserver-operator-796bbdcf4f-gmvnz\" (UID: \"c2adc97c-46b4-4de7-b49c-5f75a16c2c32\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gmvnz" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.115186 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2adc97c-46b4-4de7-b49c-5f75a16c2c32-config\") pod \"openshift-apiserver-operator-796bbdcf4f-gmvnz\" (UID: \"c2adc97c-46b4-4de7-b49c-5f75a16c2c32\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gmvnz" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.115222 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/293e299a-d25d-4755-8d9d-4d8f224c8791-certs\") pod \"machine-config-server-khkwh\" (UID: \"293e299a-d25d-4755-8d9d-4d8f224c8791\") " pod="openshift-machine-config-operator/machine-config-server-khkwh" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.115301 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/407d6ced-2671-40de-aa70-153cfda69fba-mountpoint-dir\") pod \"csi-hostpathplugin-9sxg5\" (UID: \"407d6ced-2671-40de-aa70-153cfda69fba\") " pod="hostpath-provisioner/csi-hostpathplugin-9sxg5" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.115376 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/19ade530-d0de-44d1-be83-96136753a35e-config-volume\") pod \"collect-profiles-29416530-88l5w\" (UID: \"19ade530-d0de-44d1-be83-96136753a35e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416530-88l5w" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.115474 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.115521 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d13a541e-f90d-4222-8486-2d47854ac8e5-bound-sa-token\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.115618 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a668f1b9-b7e6-4c39-9fc2-9beb5e5e5bce-tmpfs\") pod \"packageserver-d55dfcdfc-2nslt\" (UID: \"a668f1b9-b7e6-4c39-9fc2-9beb5e5e5bce\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2nslt" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.115671 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/407d6ced-2671-40de-aa70-153cfda69fba-socket-dir\") pod \"csi-hostpathplugin-9sxg5\" (UID: \"407d6ced-2671-40de-aa70-153cfda69fba\") " pod="hostpath-provisioner/csi-hostpathplugin-9sxg5" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.115735 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h72cf\" (UniqueName: \"kubernetes.io/projected/f7feaf6a-2ad5-4ef3-88d1-a2a483e5f0d9-kube-api-access-h72cf\") pod \"multus-admission-controller-857f4d67dd-ntmms\" (UID: \"f7feaf6a-2ad5-4ef3-88d1-a2a483e5f0d9\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ntmms" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.115800 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/18caf043-1bf2-4e67-bb3b-e4edf9fc8822-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-wbjzw\" (UID: \"18caf043-1bf2-4e67-bb3b-e4edf9fc8822\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wbjzw" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.115847 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mcb9\" (UniqueName: \"kubernetes.io/projected/407d6ced-2671-40de-aa70-153cfda69fba-kube-api-access-9mcb9\") pod \"csi-hostpathplugin-9sxg5\" (UID: \"407d6ced-2671-40de-aa70-153cfda69fba\") " pod="hostpath-provisioner/csi-hostpathplugin-9sxg5" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.115872 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d13a541e-f90d-4222-8486-2d47854ac8e5-trusted-ca\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.142423 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5rwsv" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.145181 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-s5b86" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.145788 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-gkpck" Dec 06 03:42:35 crc kubenswrapper[4802]: E1206 03:42:35.145913 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:42:35.645895122 +0000 UTC m=+148.517804274 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-blwnp" (UID: "d13a541e-f90d-4222-8486-2d47854ac8e5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.184395 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x6zsh" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.185505 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6d29n" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.185590 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9rvnv" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.185975 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-t56zb"] Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.195052 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jwc8v" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.216556 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.216860 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d13a541e-f90d-4222-8486-2d47854ac8e5-installation-pull-secrets\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.216887 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a668f1b9-b7e6-4c39-9fc2-9beb5e5e5bce-webhook-cert\") pod \"packageserver-d55dfcdfc-2nslt\" (UID: \"a668f1b9-b7e6-4c39-9fc2-9beb5e5e5bce\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2nslt" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.216913 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkqj8\" (UniqueName: \"kubernetes.io/projected/a8dc72d5-549b-40ad-9343-4f402a8e8572-kube-api-access-pkqj8\") pod \"ingress-canary-5kgv6\" (UID: \"a8dc72d5-549b-40ad-9343-4f402a8e8572\") " pod="openshift-ingress-canary/ingress-canary-5kgv6" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.216935 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/407d6ced-2671-40de-aa70-153cfda69fba-csi-data-dir\") pod \"csi-hostpathplugin-9sxg5\" (UID: \"407d6ced-2671-40de-aa70-153cfda69fba\") " pod="hostpath-provisioner/csi-hostpathplugin-9sxg5" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.216953 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbc8f\" (UniqueName: \"kubernetes.io/projected/19ade530-d0de-44d1-be83-96136753a35e-kube-api-access-pbc8f\") pod \"collect-profiles-29416530-88l5w\" (UID: \"19ade530-d0de-44d1-be83-96136753a35e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416530-88l5w" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.216970 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pflwz\" (UniqueName: \"kubernetes.io/projected/78eab4ec-5f1e-42cd-8ce7-48f76a55cc95-kube-api-access-pflwz\") pod \"catalog-operator-68c6474976-4gkmd\" (UID: \"78eab4ec-5f1e-42cd-8ce7-48f76a55cc95\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4gkmd" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.216987 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ddb4b244-e4e0-4ae4-a217-c6531fc90b6a-config-volume\") pod \"dns-default-whhs8\" (UID: \"ddb4b244-e4e0-4ae4-a217-c6531fc90b6a\") " pod="openshift-dns/dns-default-whhs8" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.217005 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d13a541e-f90d-4222-8486-2d47854ac8e5-registry-tls\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.217021 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f7feaf6a-2ad5-4ef3-88d1-a2a483e5f0d9-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-ntmms\" (UID: \"f7feaf6a-2ad5-4ef3-88d1-a2a483e5f0d9\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ntmms" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.217038 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kcmz\" (UniqueName: \"kubernetes.io/projected/293e299a-d25d-4755-8d9d-4d8f224c8791-kube-api-access-4kcmz\") pod \"machine-config-server-khkwh\" (UID: \"293e299a-d25d-4755-8d9d-4d8f224c8791\") " pod="openshift-machine-config-operator/machine-config-server-khkwh" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.217053 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/19ade530-d0de-44d1-be83-96136753a35e-secret-volume\") pod \"collect-profiles-29416530-88l5w\" (UID: \"19ade530-d0de-44d1-be83-96136753a35e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416530-88l5w" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.217068 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b78b4d0e-a83c-4339-b1c4-8998e69e7218-serving-cert\") pod \"service-ca-operator-777779d784-zb4m6\" (UID: \"b78b4d0e-a83c-4339-b1c4-8998e69e7218\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-zb4m6" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.217085 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/407d6ced-2671-40de-aa70-153cfda69fba-plugins-dir\") pod \"csi-hostpathplugin-9sxg5\" (UID: \"407d6ced-2671-40de-aa70-153cfda69fba\") " pod="hostpath-provisioner/csi-hostpathplugin-9sxg5" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.217121 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/407d6ced-2671-40de-aa70-153cfda69fba-registration-dir\") pod \"csi-hostpathplugin-9sxg5\" (UID: \"407d6ced-2671-40de-aa70-153cfda69fba\") " pod="hostpath-provisioner/csi-hostpathplugin-9sxg5" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.217139 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b78b4d0e-a83c-4339-b1c4-8998e69e7218-config\") pod \"service-ca-operator-777779d784-zb4m6\" (UID: \"b78b4d0e-a83c-4339-b1c4-8998e69e7218\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-zb4m6" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.217178 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzkg4\" (UniqueName: \"kubernetes.io/projected/d13a541e-f90d-4222-8486-2d47854ac8e5-kube-api-access-tzkg4\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.217210 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/78eab4ec-5f1e-42cd-8ce7-48f76a55cc95-srv-cert\") pod \"catalog-operator-68c6474976-4gkmd\" (UID: \"78eab4ec-5f1e-42cd-8ce7-48f76a55cc95\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4gkmd" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.217230 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgrhd\" (UniqueName: \"kubernetes.io/projected/ddb4b244-e4e0-4ae4-a217-c6531fc90b6a-kube-api-access-lgrhd\") pod \"dns-default-whhs8\" (UID: \"ddb4b244-e4e0-4ae4-a217-c6531fc90b6a\") " pod="openshift-dns/dns-default-whhs8" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.217247 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2610e572-bd43-4029-a846-b6694c1896f7-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-mpnf2\" (UID: \"2610e572-bd43-4029-a846-b6694c1896f7\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mpnf2" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.217264 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wl6sm\" (UniqueName: \"kubernetes.io/projected/b78b4d0e-a83c-4339-b1c4-8998e69e7218-kube-api-access-wl6sm\") pod \"service-ca-operator-777779d784-zb4m6\" (UID: \"b78b4d0e-a83c-4339-b1c4-8998e69e7218\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-zb4m6" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.217281 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2610e572-bd43-4029-a846-b6694c1896f7-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-mpnf2\" (UID: \"2610e572-bd43-4029-a846-b6694c1896f7\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mpnf2" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.217297 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a668f1b9-b7e6-4c39-9fc2-9beb5e5e5bce-apiservice-cert\") pod \"packageserver-d55dfcdfc-2nslt\" (UID: \"a668f1b9-b7e6-4c39-9fc2-9beb5e5e5bce\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2nslt" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.217317 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/293e299a-d25d-4755-8d9d-4d8f224c8791-node-bootstrap-token\") pod \"machine-config-server-khkwh\" (UID: \"293e299a-d25d-4755-8d9d-4d8f224c8791\") " pod="openshift-machine-config-operator/machine-config-server-khkwh" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.217332 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/78eab4ec-5f1e-42cd-8ce7-48f76a55cc95-profile-collector-cert\") pod \"catalog-operator-68c6474976-4gkmd\" (UID: \"78eab4ec-5f1e-42cd-8ce7-48f76a55cc95\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4gkmd" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.217353 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a8dc72d5-549b-40ad-9343-4f402a8e8572-cert\") pod \"ingress-canary-5kgv6\" (UID: \"a8dc72d5-549b-40ad-9343-4f402a8e8572\") " pod="openshift-ingress-canary/ingress-canary-5kgv6" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.217387 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ddb4b244-e4e0-4ae4-a217-c6531fc90b6a-metrics-tls\") pod \"dns-default-whhs8\" (UID: \"ddb4b244-e4e0-4ae4-a217-c6531fc90b6a\") " pod="openshift-dns/dns-default-whhs8" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.217416 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d13a541e-f90d-4222-8486-2d47854ac8e5-registry-certificates\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.217434 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d13a541e-f90d-4222-8486-2d47854ac8e5-ca-trust-extracted\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.217452 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dlgh\" (UniqueName: \"kubernetes.io/projected/c2adc97c-46b4-4de7-b49c-5f75a16c2c32-kube-api-access-5dlgh\") pod \"openshift-apiserver-operator-796bbdcf4f-gmvnz\" (UID: \"c2adc97c-46b4-4de7-b49c-5f75a16c2c32\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gmvnz" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.217469 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2adc97c-46b4-4de7-b49c-5f75a16c2c32-config\") pod \"openshift-apiserver-operator-796bbdcf4f-gmvnz\" (UID: \"c2adc97c-46b4-4de7-b49c-5f75a16c2c32\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gmvnz" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.217486 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/407d6ced-2671-40de-aa70-153cfda69fba-mountpoint-dir\") pod \"csi-hostpathplugin-9sxg5\" (UID: \"407d6ced-2671-40de-aa70-153cfda69fba\") " pod="hostpath-provisioner/csi-hostpathplugin-9sxg5" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.217499 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/293e299a-d25d-4755-8d9d-4d8f224c8791-certs\") pod \"machine-config-server-khkwh\" (UID: \"293e299a-d25d-4755-8d9d-4d8f224c8791\") " pod="openshift-machine-config-operator/machine-config-server-khkwh" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.217526 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/19ade530-d0de-44d1-be83-96136753a35e-config-volume\") pod \"collect-profiles-29416530-88l5w\" (UID: \"19ade530-d0de-44d1-be83-96136753a35e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416530-88l5w" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.217551 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d13a541e-f90d-4222-8486-2d47854ac8e5-bound-sa-token\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.217578 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a668f1b9-b7e6-4c39-9fc2-9beb5e5e5bce-tmpfs\") pod \"packageserver-d55dfcdfc-2nslt\" (UID: \"a668f1b9-b7e6-4c39-9fc2-9beb5e5e5bce\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2nslt" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.217616 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/407d6ced-2671-40de-aa70-153cfda69fba-socket-dir\") pod \"csi-hostpathplugin-9sxg5\" (UID: \"407d6ced-2671-40de-aa70-153cfda69fba\") " pod="hostpath-provisioner/csi-hostpathplugin-9sxg5" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.217633 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h72cf\" (UniqueName: \"kubernetes.io/projected/f7feaf6a-2ad5-4ef3-88d1-a2a483e5f0d9-kube-api-access-h72cf\") pod \"multus-admission-controller-857f4d67dd-ntmms\" (UID: \"f7feaf6a-2ad5-4ef3-88d1-a2a483e5f0d9\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ntmms" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.217663 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/18caf043-1bf2-4e67-bb3b-e4edf9fc8822-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-wbjzw\" (UID: \"18caf043-1bf2-4e67-bb3b-e4edf9fc8822\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wbjzw" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.217679 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mcb9\" (UniqueName: \"kubernetes.io/projected/407d6ced-2671-40de-aa70-153cfda69fba-kube-api-access-9mcb9\") pod \"csi-hostpathplugin-9sxg5\" (UID: \"407d6ced-2671-40de-aa70-153cfda69fba\") " pod="hostpath-provisioner/csi-hostpathplugin-9sxg5" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.217694 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d13a541e-f90d-4222-8486-2d47854ac8e5-trusted-ca\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.217732 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mzbrn\" (UniqueName: \"kubernetes.io/projected/a668f1b9-b7e6-4c39-9fc2-9beb5e5e5bce-kube-api-access-mzbrn\") pod \"packageserver-d55dfcdfc-2nslt\" (UID: \"a668f1b9-b7e6-4c39-9fc2-9beb5e5e5bce\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2nslt" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.217768 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2610e572-bd43-4029-a846-b6694c1896f7-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-mpnf2\" (UID: \"2610e572-bd43-4029-a846-b6694c1896f7\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mpnf2" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.217784 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9znd\" (UniqueName: \"kubernetes.io/projected/18caf043-1bf2-4e67-bb3b-e4edf9fc8822-kube-api-access-h9znd\") pod \"package-server-manager-789f6589d5-wbjzw\" (UID: \"18caf043-1bf2-4e67-bb3b-e4edf9fc8822\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wbjzw" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.217810 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bj479\" (UniqueName: \"kubernetes.io/projected/8002be53-a1e5-4ef7-93b4-384d9b16a7e2-kube-api-access-bj479\") pod \"migrator-59844c95c7-zfjwx\" (UID: \"8002be53-a1e5-4ef7-93b4-384d9b16a7e2\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-zfjwx" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.217827 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c2adc97c-46b4-4de7-b49c-5f75a16c2c32-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-gmvnz\" (UID: \"c2adc97c-46b4-4de7-b49c-5f75a16c2c32\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gmvnz" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.222019 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d13a541e-f90d-4222-8486-2d47854ac8e5-installation-pull-secrets\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:35 crc kubenswrapper[4802]: E1206 03:42:35.222173 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:42:35.722147181 +0000 UTC m=+148.594056333 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.223721 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b78b4d0e-a83c-4339-b1c4-8998e69e7218-config\") pod \"service-ca-operator-777779d784-zb4m6\" (UID: \"b78b4d0e-a83c-4339-b1c4-8998e69e7218\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-zb4m6" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.223933 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/407d6ced-2671-40de-aa70-153cfda69fba-csi-data-dir\") pod \"csi-hostpathplugin-9sxg5\" (UID: \"407d6ced-2671-40de-aa70-153cfda69fba\") " pod="hostpath-provisioner/csi-hostpathplugin-9sxg5" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.224665 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ddb4b244-e4e0-4ae4-a217-c6531fc90b6a-config-volume\") pod \"dns-default-whhs8\" (UID: \"ddb4b244-e4e0-4ae4-a217-c6531fc90b6a\") " pod="openshift-dns/dns-default-whhs8" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.229189 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/407d6ced-2671-40de-aa70-153cfda69fba-mountpoint-dir\") pod \"csi-hostpathplugin-9sxg5\" (UID: \"407d6ced-2671-40de-aa70-153cfda69fba\") " pod="hostpath-provisioner/csi-hostpathplugin-9sxg5" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.229953 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a668f1b9-b7e6-4c39-9fc2-9beb5e5e5bce-webhook-cert\") pod \"packageserver-d55dfcdfc-2nslt\" (UID: \"a668f1b9-b7e6-4c39-9fc2-9beb5e5e5bce\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2nslt" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.230146 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/19ade530-d0de-44d1-be83-96136753a35e-config-volume\") pod \"collect-profiles-29416530-88l5w\" (UID: \"19ade530-d0de-44d1-be83-96136753a35e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416530-88l5w" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.231873 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a668f1b9-b7e6-4c39-9fc2-9beb5e5e5bce-tmpfs\") pod \"packageserver-d55dfcdfc-2nslt\" (UID: \"a668f1b9-b7e6-4c39-9fc2-9beb5e5e5bce\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2nslt" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.232545 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/407d6ced-2671-40de-aa70-153cfda69fba-socket-dir\") pod \"csi-hostpathplugin-9sxg5\" (UID: \"407d6ced-2671-40de-aa70-153cfda69fba\") " pod="hostpath-provisioner/csi-hostpathplugin-9sxg5" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.235492 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2610e572-bd43-4029-a846-b6694c1896f7-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-mpnf2\" (UID: \"2610e572-bd43-4029-a846-b6694c1896f7\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mpnf2" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.237288 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/18caf043-1bf2-4e67-bb3b-e4edf9fc8822-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-wbjzw\" (UID: \"18caf043-1bf2-4e67-bb3b-e4edf9fc8822\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wbjzw" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.238632 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d13a541e-f90d-4222-8486-2d47854ac8e5-trusted-ca\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.238930 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c2adc97c-46b4-4de7-b49c-5f75a16c2c32-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-gmvnz\" (UID: \"c2adc97c-46b4-4de7-b49c-5f75a16c2c32\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gmvnz" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.241102 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/78eab4ec-5f1e-42cd-8ce7-48f76a55cc95-profile-collector-cert\") pod \"catalog-operator-68c6474976-4gkmd\" (UID: \"78eab4ec-5f1e-42cd-8ce7-48f76a55cc95\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4gkmd" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.241487 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/407d6ced-2671-40de-aa70-153cfda69fba-registration-dir\") pod \"csi-hostpathplugin-9sxg5\" (UID: \"407d6ced-2671-40de-aa70-153cfda69fba\") " pod="hostpath-provisioner/csi-hostpathplugin-9sxg5" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.241595 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/407d6ced-2671-40de-aa70-153cfda69fba-plugins-dir\") pod \"csi-hostpathplugin-9sxg5\" (UID: \"407d6ced-2671-40de-aa70-153cfda69fba\") " pod="hostpath-provisioner/csi-hostpathplugin-9sxg5" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.242638 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d13a541e-f90d-4222-8486-2d47854ac8e5-registry-tls\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.242984 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d13a541e-f90d-4222-8486-2d47854ac8e5-ca-trust-extracted\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.244538 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/293e299a-d25d-4755-8d9d-4d8f224c8791-node-bootstrap-token\") pod \"machine-config-server-khkwh\" (UID: \"293e299a-d25d-4755-8d9d-4d8f224c8791\") " pod="openshift-machine-config-operator/machine-config-server-khkwh" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.245715 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d13a541e-f90d-4222-8486-2d47854ac8e5-registry-certificates\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.245739 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/293e299a-d25d-4755-8d9d-4d8f224c8791-certs\") pod \"machine-config-server-khkwh\" (UID: \"293e299a-d25d-4755-8d9d-4d8f224c8791\") " pod="openshift-machine-config-operator/machine-config-server-khkwh" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.246547 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2adc97c-46b4-4de7-b49c-5f75a16c2c32-config\") pod \"openshift-apiserver-operator-796bbdcf4f-gmvnz\" (UID: \"c2adc97c-46b4-4de7-b49c-5f75a16c2c32\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gmvnz" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.247259 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2610e572-bd43-4029-a846-b6694c1896f7-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-mpnf2\" (UID: \"2610e572-bd43-4029-a846-b6694c1896f7\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mpnf2" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.247678 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ddb4b244-e4e0-4ae4-a217-c6531fc90b6a-metrics-tls\") pod \"dns-default-whhs8\" (UID: \"ddb4b244-e4e0-4ae4-a217-c6531fc90b6a\") " pod="openshift-dns/dns-default-whhs8" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.247966 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/78eab4ec-5f1e-42cd-8ce7-48f76a55cc95-srv-cert\") pod \"catalog-operator-68c6474976-4gkmd\" (UID: \"78eab4ec-5f1e-42cd-8ce7-48f76a55cc95\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4gkmd" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.260925 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f7feaf6a-2ad5-4ef3-88d1-a2a483e5f0d9-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-ntmms\" (UID: \"f7feaf6a-2ad5-4ef3-88d1-a2a483e5f0d9\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ntmms" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.260939 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a668f1b9-b7e6-4c39-9fc2-9beb5e5e5bce-apiservice-cert\") pod \"packageserver-d55dfcdfc-2nslt\" (UID: \"a668f1b9-b7e6-4c39-9fc2-9beb5e5e5bce\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2nslt" Dec 06 03:42:35 crc kubenswrapper[4802]: W1206 03:42:35.263128 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-3acf17c85df1ca0ac1074c2412b3b34c1efb1fdead9a28fcad6aa971265dcfda WatchSource:0}: Error finding container 3acf17c85df1ca0ac1074c2412b3b34c1efb1fdead9a28fcad6aa971265dcfda: Status 404 returned error can't find the container with id 3acf17c85df1ca0ac1074c2412b3b34c1efb1fdead9a28fcad6aa971265dcfda Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.265446 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/19ade530-d0de-44d1-be83-96136753a35e-secret-volume\") pod \"collect-profiles-29416530-88l5w\" (UID: \"19ade530-d0de-44d1-be83-96136753a35e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416530-88l5w" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.268502 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fqdzj"] Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.274262 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b78b4d0e-a83c-4339-b1c4-8998e69e7218-serving-cert\") pod \"service-ca-operator-777779d784-zb4m6\" (UID: \"b78b4d0e-a83c-4339-b1c4-8998e69e7218\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-zb4m6" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.286377 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pflwz\" (UniqueName: \"kubernetes.io/projected/78eab4ec-5f1e-42cd-8ce7-48f76a55cc95-kube-api-access-pflwz\") pod \"catalog-operator-68c6474976-4gkmd\" (UID: \"78eab4ec-5f1e-42cd-8ce7-48f76a55cc95\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4gkmd" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.286466 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbc8f\" (UniqueName: \"kubernetes.io/projected/19ade530-d0de-44d1-be83-96136753a35e-kube-api-access-pbc8f\") pod \"collect-profiles-29416530-88l5w\" (UID: \"19ade530-d0de-44d1-be83-96136753a35e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416530-88l5w" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.287941 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ddpmx" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.291339 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-7cg8t"] Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.292687 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-8xz4m"] Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.297593 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-wrh75" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.312292 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jv24g"] Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.315719 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d13a541e-f90d-4222-8486-2d47854ac8e5-bound-sa-token\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.319091 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416530-88l5w" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.319365 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pkqj8\" (UniqueName: \"kubernetes.io/projected/a8dc72d5-549b-40ad-9343-4f402a8e8572-kube-api-access-pkqj8\") pod \"ingress-canary-5kgv6\" (UID: \"a8dc72d5-549b-40ad-9343-4f402a8e8572\") " pod="openshift-ingress-canary/ingress-canary-5kgv6" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.319446 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a8dc72d5-549b-40ad-9343-4f402a8e8572-cert\") pod \"ingress-canary-5kgv6\" (UID: \"a8dc72d5-549b-40ad-9343-4f402a8e8572\") " pod="openshift-ingress-canary/ingress-canary-5kgv6" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.319491 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:35 crc kubenswrapper[4802]: E1206 03:42:35.319826 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:42:35.819814236 +0000 UTC m=+148.691723388 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-blwnp" (UID: "d13a541e-f90d-4222-8486-2d47854ac8e5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.338919 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzkg4\" (UniqueName: \"kubernetes.io/projected/d13a541e-f90d-4222-8486-2d47854ac8e5-kube-api-access-tzkg4\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.344395 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-7ckfj" event={"ID":"2362780e-93d1-4f97-bda0-138ed5180bb0","Type":"ContainerStarted","Data":"db66933a746dd27be52693eeaeb393a614c2a8afafa38bc15082419ffb470d35"} Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.344434 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-7ckfj" event={"ID":"2362780e-93d1-4f97-bda0-138ed5180bb0","Type":"ContainerStarted","Data":"d3153a3a6e26c8493c7e1c5ad9c40b8a38751e3962d1244ed255328bae75eedc"} Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.347991 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fnpz5" event={"ID":"93ffc5bd-20a8-450c-861a-0aecc58d602b","Type":"ContainerStarted","Data":"ef7e571fec1c0425f317bf0ca649350593fa07e21293baf045b07289c2b0fe09"} Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.349412 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h72cf\" (UniqueName: \"kubernetes.io/projected/f7feaf6a-2ad5-4ef3-88d1-a2a483e5f0d9-kube-api-access-h72cf\") pod \"multus-admission-controller-857f4d67dd-ntmms\" (UID: \"f7feaf6a-2ad5-4ef3-88d1-a2a483e5f0d9\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ntmms" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.356668 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-rrhqx" event={"ID":"0fbd23d0-a6b6-4a4e-b418-07fc38e5f497","Type":"ContainerStarted","Data":"049ff6c0b0f58a6aa1fd62c4d9263f96e90bd3c9b21fd25e108bc9e1f857aace"} Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.356714 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-rrhqx" event={"ID":"0fbd23d0-a6b6-4a4e-b418-07fc38e5f497","Type":"ContainerStarted","Data":"742c61214d18c9e2cde5cb26ab7a28b543a6ed9d15b104c59ac000d401b30b6b"} Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.363193 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a8dc72d5-549b-40ad-9343-4f402a8e8572-cert\") pod \"ingress-canary-5kgv6\" (UID: \"a8dc72d5-549b-40ad-9343-4f402a8e8572\") " pod="openshift-ingress-canary/ingress-canary-5kgv6" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.393306 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mcb9\" (UniqueName: \"kubernetes.io/projected/407d6ced-2671-40de-aa70-153cfda69fba-kube-api-access-9mcb9\") pod \"csi-hostpathplugin-9sxg5\" (UID: \"407d6ced-2671-40de-aa70-153cfda69fba\") " pod="hostpath-provisioner/csi-hostpathplugin-9sxg5" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.406658 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9bfz8"] Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.421637 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:42:35 crc kubenswrapper[4802]: E1206 03:42:35.422528 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:42:35.922512075 +0000 UTC m=+148.794421217 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.433271 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kcmz\" (UniqueName: \"kubernetes.io/projected/293e299a-d25d-4755-8d9d-4d8f224c8791-kube-api-access-4kcmz\") pod \"machine-config-server-khkwh\" (UID: \"293e299a-d25d-4755-8d9d-4d8f224c8791\") " pod="openshift-machine-config-operator/machine-config-server-khkwh" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.438088 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mzbrn\" (UniqueName: \"kubernetes.io/projected/a668f1b9-b7e6-4c39-9fc2-9beb5e5e5bce-kube-api-access-mzbrn\") pod \"packageserver-d55dfcdfc-2nslt\" (UID: \"a668f1b9-b7e6-4c39-9fc2-9beb5e5e5bce\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2nslt" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.438649 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9znd\" (UniqueName: \"kubernetes.io/projected/18caf043-1bf2-4e67-bb3b-e4edf9fc8822-kube-api-access-h9znd\") pod \"package-server-manager-789f6589d5-wbjzw\" (UID: \"18caf043-1bf2-4e67-bb3b-e4edf9fc8822\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wbjzw" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.460197 4802 generic.go:334] "Generic (PLEG): container finished" podID="bd64a08a-6920-4a0f-b299-aa5a009236a1" containerID="ddd513efdc7dc72e31a789cbede9cae5168f3852739d6beb11806bcaa2041f35" exitCode=0 Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.460573 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bj479\" (UniqueName: \"kubernetes.io/projected/8002be53-a1e5-4ef7-93b4-384d9b16a7e2-kube-api-access-bj479\") pod \"migrator-59844c95c7-zfjwx\" (UID: \"8002be53-a1e5-4ef7-93b4-384d9b16a7e2\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-zfjwx" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.470043 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9v855" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.470077 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-qfdsg"] Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.470092 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8br44" event={"ID":"bd64a08a-6920-4a0f-b299-aa5a009236a1","Type":"ContainerDied","Data":"ddd513efdc7dc72e31a789cbede9cae5168f3852739d6beb11806bcaa2041f35"} Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.470112 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8br44" event={"ID":"bd64a08a-6920-4a0f-b299-aa5a009236a1","Type":"ContainerStarted","Data":"6f74bef80421a37cf89bf1b82b8cd5846ae780be472b5d9490dac48128384a12"} Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.470124 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ppzvl"] Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.470133 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9v855" event={"ID":"db213df1-bfa2-4edd-a3fc-246e35e20585","Type":"ContainerStarted","Data":"941b27608e31e3ffeb9e31e27d6baeca5f2247a255b2a9d98a41883c2d37c3b0"} Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.470144 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9v855" event={"ID":"db213df1-bfa2-4edd-a3fc-246e35e20585","Type":"ContainerStarted","Data":"01a3cf31ddd14c1ad48b68874e13595b105e4c77bfdcd50de7a9d301d58afad2"} Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.494781 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2610e572-bd43-4029-a846-b6694c1896f7-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-mpnf2\" (UID: \"2610e572-bd43-4029-a846-b6694c1896f7\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mpnf2" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.501208 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9v855" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.501232 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgrhd\" (UniqueName: \"kubernetes.io/projected/ddb4b244-e4e0-4ae4-a217-c6531fc90b6a-kube-api-access-lgrhd\") pod \"dns-default-whhs8\" (UID: \"ddb4b244-e4e0-4ae4-a217-c6531fc90b6a\") " pod="openshift-dns/dns-default-whhs8" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.501321 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ghcrj" event={"ID":"dc4fc98a-badd-46dc-b700-6b01aff112b4","Type":"ContainerStarted","Data":"0b377027a53cbe37ea10d636d545a2a874b5bef326afedf7920d58613a0a9edf"} Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.501349 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ghcrj" event={"ID":"dc4fc98a-badd-46dc-b700-6b01aff112b4","Type":"ContainerStarted","Data":"b2b5abefcb565d7a880a8ab2e5f92e5e54a150b26d0c37d99d125392df1a1cc2"} Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.501360 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ghcrj" event={"ID":"dc4fc98a-badd-46dc-b700-6b01aff112b4","Type":"ContainerStarted","Data":"38227979c15450079727098d7ff6b374e0d1d009be0ad81d7050010434dbc41d"} Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.501725 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4gkmd" Dec 06 03:42:35 crc kubenswrapper[4802]: W1206 03:42:35.503927 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb2994e8f_6482_4dca_baf1_148754e78e85.slice/crio-c914c9d63a12be44e678868fe213f492b0a5cc78b857debba5e4772a1ac410b2 WatchSource:0}: Error finding container c914c9d63a12be44e678868fe213f492b0a5cc78b857debba5e4772a1ac410b2: Status 404 returned error can't find the container with id c914c9d63a12be44e678868fe213f492b0a5cc78b857debba5e4772a1ac410b2 Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.516499 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-4lqzs" event={"ID":"3f3f6253-533f-4d89-ac10-f26eca1ac390","Type":"ContainerStarted","Data":"b52b5edab50a1c5373678e14a38dd28647490c691229f279bf994e8e2d4bf59b"} Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.516533 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-4lqzs" event={"ID":"3f3f6253-533f-4d89-ac10-f26eca1ac390","Type":"ContainerStarted","Data":"a989df68068fb514cb86d365c6a1c02b79b59bf090a67da3156af8ff813f1830"} Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.517134 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-4lqzs" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.517223 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wbjzw" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.522048 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"dc82727a1ba162112e5b82c11a1c605a6ea79ffefc05762e9be4e7a6e60258b4"} Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.522079 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"e1159f5fe043cff6aa21d99bd07176091496b70fc5a3635e1907cb3743635dd0"} Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.523789 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wl6sm\" (UniqueName: \"kubernetes.io/projected/b78b4d0e-a83c-4339-b1c4-8998e69e7218-kube-api-access-wl6sm\") pod \"service-ca-operator-777779d784-zb4m6\" (UID: \"b78b4d0e-a83c-4339-b1c4-8998e69e7218\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-zb4m6" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.525438 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.526583 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-ntmms" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.535058 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2nslt" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.535545 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-rqs9q" event={"ID":"263721d5-fcca-411a-876e-0acfe0863d2d","Type":"ContainerStarted","Data":"f6333b5fe7dbcf85d5cef40f1644c3754ad08959a1f3230bb14eaa606e8db655"} Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.535577 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-rqs9q" event={"ID":"263721d5-fcca-411a-876e-0acfe0863d2d","Type":"ContainerStarted","Data":"19eb0b14b80bc6311964b5a7503f525c2e7d8fe43f506b84a274e59d8cbf29e5"} Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.536185 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-rqs9q" Dec 06 03:42:35 crc kubenswrapper[4802]: E1206 03:42:35.537958 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:42:36.037940688 +0000 UTC m=+148.909849840 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-blwnp" (UID: "d13a541e-f90d-4222-8486-2d47854ac8e5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.540127 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mpnf2" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.541483 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dlgh\" (UniqueName: \"kubernetes.io/projected/c2adc97c-46b4-4de7-b49c-5f75a16c2c32-kube-api-access-5dlgh\") pod \"openshift-apiserver-operator-796bbdcf4f-gmvnz\" (UID: \"c2adc97c-46b4-4de7-b49c-5f75a16c2c32\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gmvnz" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.550113 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-thckl"] Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.557937 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-t56zb" event={"ID":"1051665a-7918-487f-ad08-ff6d53d113ca","Type":"ContainerStarted","Data":"2947d90cecdae048594cde6a9e27df25f4c0bde15e5aa8ccc294e36f45dd4ea4"} Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.569711 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-zfjwx" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.574415 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkqj8\" (UniqueName: \"kubernetes.io/projected/a8dc72d5-549b-40ad-9343-4f402a8e8572-kube-api-access-pkqj8\") pod \"ingress-canary-5kgv6\" (UID: \"a8dc72d5-549b-40ad-9343-4f402a8e8572\") " pod="openshift-ingress-canary/ingress-canary-5kgv6" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.574927 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-4lqzs" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.579849 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"5b294481a81b85eb354f84eb0396a6da2c9b251923b2092205973b39890e0b0a"} Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.580346 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-zb4m6" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.630334 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.633341 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-whhs8" Dec 06 03:42:35 crc kubenswrapper[4802]: E1206 03:42:35.635191 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:42:36.13516933 +0000 UTC m=+149.007078472 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.654608 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-9sxg5" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.664485 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-khkwh" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.666052 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-5kgv6" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.732147 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:35 crc kubenswrapper[4802]: E1206 03:42:35.732485 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:42:36.232472365 +0000 UTC m=+149.104381517 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-blwnp" (UID: "d13a541e-f90d-4222-8486-2d47854ac8e5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.732838 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-rqs9q" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.808947 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-cgpdh"] Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.810002 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gmvnz" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.826443 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-rrhqx" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.829496 4802 patch_prober.go:28] interesting pod/router-default-5444994796-rrhqx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 03:42:35 crc kubenswrapper[4802]: [-]has-synced failed: reason withheld Dec 06 03:42:35 crc kubenswrapper[4802]: [+]process-running ok Dec 06 03:42:35 crc kubenswrapper[4802]: healthz check failed Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.829533 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-rrhqx" podUID="0fbd23d0-a6b6-4a4e-b418-07fc38e5f497" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.834066 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:42:35 crc kubenswrapper[4802]: E1206 03:42:35.834469 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:42:36.334447506 +0000 UTC m=+149.206356658 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:35 crc kubenswrapper[4802]: I1206 03:42:35.935533 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:35 crc kubenswrapper[4802]: E1206 03:42:35.935973 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:42:36.435956904 +0000 UTC m=+149.307866056 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-blwnp" (UID: "d13a541e-f90d-4222-8486-2d47854ac8e5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.040185 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:42:36 crc kubenswrapper[4802]: E1206 03:42:36.040661 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:42:36.540635397 +0000 UTC m=+149.412544559 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.081509 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-gkpck"] Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.094126 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-s5b86"] Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.095741 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-9rvnv"] Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.105037 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-jwc8v"] Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.109360 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5rwsv"] Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.114023 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6d29n"] Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.124886 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-9s5r8"] Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.140834 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x6zsh"] Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.141573 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:36 crc kubenswrapper[4802]: E1206 03:42:36.141932 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:42:36.641915148 +0000 UTC m=+149.513824300 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-blwnp" (UID: "d13a541e-f90d-4222-8486-2d47854ac8e5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:36 crc kubenswrapper[4802]: W1206 03:42:36.240004 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6258654e_b524_4fdd_ba9b_7443b57c2585.slice/crio-a438cb6a7702b7ea62daf962ab9b9bca815591237353e9972a54ea3a5c4d182e WatchSource:0}: Error finding container a438cb6a7702b7ea62daf962ab9b9bca815591237353e9972a54ea3a5c4d182e: Status 404 returned error can't find the container with id a438cb6a7702b7ea62daf962ab9b9bca815591237353e9972a54ea3a5c4d182e Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.242831 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:42:36 crc kubenswrapper[4802]: E1206 03:42:36.243081 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:42:36.743065567 +0000 UTC m=+149.614974719 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:36 crc kubenswrapper[4802]: W1206 03:42:36.262107 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode61f4275_03bf_4cb0_bb9a_e519c41f159a.slice/crio-935ea5241886e3d5dd733649673dc9dfec080fa6fa0c1c5332b2e3c566c075a8 WatchSource:0}: Error finding container 935ea5241886e3d5dd733649673dc9dfec080fa6fa0c1c5332b2e3c566c075a8: Status 404 returned error can't find the container with id 935ea5241886e3d5dd733649673dc9dfec080fa6fa0c1c5332b2e3c566c075a8 Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.353089 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:36 crc kubenswrapper[4802]: E1206 03:42:36.353384 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:42:36.853371492 +0000 UTC m=+149.725280644 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-blwnp" (UID: "d13a541e-f90d-4222-8486-2d47854ac8e5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.417978 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-wrh75"] Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.465171 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:42:36 crc kubenswrapper[4802]: E1206 03:42:36.465947 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:42:36.965925816 +0000 UTC m=+149.837834968 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.491618 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mpnf2"] Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.516057 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416530-88l5w"] Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.567434 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.568149 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ghcrj" podStartSLOduration=130.568127762 podStartE2EDuration="2m10.568127762s" podCreationTimestamp="2025-12-06 03:40:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:42:36.566602892 +0000 UTC m=+149.438512054" watchObservedRunningTime="2025-12-06 03:42:36.568127762 +0000 UTC m=+149.440036914" Dec 06 03:42:36 crc kubenswrapper[4802]: E1206 03:42:36.569661 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:42:37.069646054 +0000 UTC m=+149.941555216 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-blwnp" (UID: "d13a541e-f90d-4222-8486-2d47854ac8e5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:36 crc kubenswrapper[4802]: W1206 03:42:36.570971 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7621a355_6a10_4143_b35e_f1f617cbf849.slice/crio-1668b1a8561eaf402740ff3723c65f536540cfd0e8888f4e988bee2e7341f93a WatchSource:0}: Error finding container 1668b1a8561eaf402740ff3723c65f536540cfd0e8888f4e988bee2e7341f93a: Status 404 returned error can't find the container with id 1668b1a8561eaf402740ff3723c65f536540cfd0e8888f4e988bee2e7341f93a Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.619148 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6d29n" event={"ID":"1a34c64b-194e-41a6-8b55-0a87187c4cfc","Type":"ContainerStarted","Data":"c033e3e92c460d3b7c0d7dfdfba59d24d8057a94fc54ff053b03bda831aca911"} Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.621151 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9bfz8" event={"ID":"b2994e8f-6482-4dca-baf1-148754e78e85","Type":"ContainerStarted","Data":"c914c9d63a12be44e678868fe213f492b0a5cc78b857debba5e4772a1ac410b2"} Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.639803 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fqdzj" event={"ID":"9172dd32-c65e-4637-ada8-d2c5b0a772ab","Type":"ContainerStarted","Data":"5d4c25e4af35a5f14cf10721a6545a5aa38be12c955a318601c4416d94942e84"} Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.639848 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fqdzj" event={"ID":"9172dd32-c65e-4637-ada8-d2c5b0a772ab","Type":"ContainerStarted","Data":"28e767d2ae0f5c14224f5538c104717418c767e15385ebab18abffbe8733be86"} Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.666281 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-7cg8t" event={"ID":"0057cff6-6085-4b41-bf4f-e1121bece99a","Type":"ContainerStarted","Data":"90de0f40500f1fa173719aef10f25bb7b5be364a95099c98c4bd7d410e054c22"} Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.666326 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-7cg8t" event={"ID":"0057cff6-6085-4b41-bf4f-e1121bece99a","Type":"ContainerStarted","Data":"99869a184264df8597502c20e233a206070af3d741fed6a54a2115de5faaf788"} Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.666821 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-7cg8t" Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.669147 4802 patch_prober.go:28] interesting pod/downloads-7954f5f757-7cg8t container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.669177 4802 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-7cg8t" podUID="0057cff6-6085-4b41-bf4f-e1121bece99a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.670604 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:42:36 crc kubenswrapper[4802]: E1206 03:42:36.670887 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:42:37.170872554 +0000 UTC m=+150.042781706 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.673642 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qfdsg" event={"ID":"e4f6934a-db7e-4031-baa8-2f41dc8c2c18","Type":"ContainerStarted","Data":"272cfd1156342792251cce111dd55257c7ecd4cb59012408b22c8671489eb6bf"} Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.673699 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qfdsg" event={"ID":"e4f6934a-db7e-4031-baa8-2f41dc8c2c18","Type":"ContainerStarted","Data":"17f756db14bad50f6eda1ef1d872551bb0370f476398c71b9e0d6160cdf57ae9"} Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.681781 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9rvnv" event={"ID":"ff639bee-182b-4dd2-aabc-d396a3bd9ce5","Type":"ContainerStarted","Data":"504f0956f4e92cddfe545312dbcde2956776fca6b516756f3b88892cf3b89fe6"} Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.685324 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-zfjwx"] Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.688957 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-9s5r8" event={"ID":"eb15a464-d0e8-4e49-b018-14353cc9d6cb","Type":"ContainerStarted","Data":"27354689eda8951e4af4450ea9629effaacb5877fa085587698bce5918526d7c"} Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.702968 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4gkmd"] Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.709634 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fnpz5" event={"ID":"93ffc5bd-20a8-450c-861a-0aecc58d602b","Type":"ContainerStarted","Data":"4c0ca344b4228e438c2b1b1f1820af7e1b8b8f9a879e3d7e32eab986945723e6"} Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.711948 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ddpmx"] Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.729311 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-wrh75" event={"ID":"7621a355-6a10-4143-b35e-f1f617cbf849","Type":"ContainerStarted","Data":"1668b1a8561eaf402740ff3723c65f536540cfd0e8888f4e988bee2e7341f93a"} Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.732196 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-4lqzs" podStartSLOduration=130.732177531 podStartE2EDuration="2m10.732177531s" podCreationTimestamp="2025-12-06 03:40:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:42:36.701430395 +0000 UTC m=+149.573339547" watchObservedRunningTime="2025-12-06 03:42:36.732177531 +0000 UTC m=+149.604086683" Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.732888 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-zb4m6"] Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.734607 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5rwsv" event={"ID":"6258654e-b524-4fdd-ba9b-7443b57c2585","Type":"ContainerStarted","Data":"a438cb6a7702b7ea62daf962ab9b9bca815591237353e9972a54ea3a5c4d182e"} Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.739386 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"dce152b26e25f9dbf4a50e5f9f72703f3e725e6d9c264bd623b9644dd19cac5d"} Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.740177 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.748289 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-khkwh" event={"ID":"293e299a-d25d-4755-8d9d-4d8f224c8791","Type":"ContainerStarted","Data":"89ef1a81bc4a2a9cc7b267900f9a3787bf6abec65ac1b7a58ce523f6cc9960c1"} Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.758033 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-cgpdh" event={"ID":"a68fc0dd-997c-4d12-9a20-5345dfe98a41","Type":"ContainerStarted","Data":"f9fc0a9a62f254e815cc5ad499da3ac6e027b333f613cfd4d44e4a88fe48a39a"} Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.758064 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-cgpdh" event={"ID":"a68fc0dd-997c-4d12-9a20-5345dfe98a41","Type":"ContainerStarted","Data":"197163bad5e7b4ac89d2ca3dc0e289138a45c852df8c982644c9ef597896be89"} Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.758540 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-cgpdh" Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.773710 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:36 crc kubenswrapper[4802]: E1206 03:42:36.776190 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:42:37.276174914 +0000 UTC m=+150.148084086 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-blwnp" (UID: "d13a541e-f90d-4222-8486-2d47854ac8e5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.781778 4802 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-cgpdh container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.12:8443/healthz\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.781848 4802 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-cgpdh" podUID="a68fc0dd-997c-4d12-9a20-5345dfe98a41" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.12:8443/healthz\": dial tcp 10.217.0.12:8443: connect: connection refused" Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.786193 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jv24g" event={"ID":"f643d9e2-a7df-4aab-a2f5-4bf47363636b","Type":"ContainerStarted","Data":"e526ca421136de2c5f12cfb888517c1ac99c0e6d298c178dee9db074460ea65a"} Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.786226 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jv24g" event={"ID":"f643d9e2-a7df-4aab-a2f5-4bf47363636b","Type":"ContainerStarted","Data":"08b404ebca67054634423ffd8a483f73a76b99fccc2f031995d800a448ab6b45"} Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.804693 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9v855" podStartSLOduration=129.80467667 podStartE2EDuration="2m9.80467667s" podCreationTimestamp="2025-12-06 03:40:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:42:36.803448776 +0000 UTC m=+149.675357928" watchObservedRunningTime="2025-12-06 03:42:36.80467667 +0000 UTC m=+149.676585822" Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.806595 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-gkpck" event={"ID":"7abae297-572a-4c3c-8e7d-4e2157d838ad","Type":"ContainerStarted","Data":"bd562c189e2334b01e80027ef13e6ebe795ff9f950c66ba105ad1ed176a03f15"} Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.817880 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ppzvl" event={"ID":"a92fd9bb-6d1b-4ab1-9f1a-bc9086ecb08e","Type":"ContainerStarted","Data":"512448b8c4a3adaf751ffebdc30cd7edb63acf20ac0393e21dabc76ddf1e1ab0"} Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.832945 4802 patch_prober.go:28] interesting pod/router-default-5444994796-rrhqx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 03:42:36 crc kubenswrapper[4802]: [-]has-synced failed: reason withheld Dec 06 03:42:36 crc kubenswrapper[4802]: [+]process-running ok Dec 06 03:42:36 crc kubenswrapper[4802]: healthz check failed Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.832993 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-rrhqx" podUID="0fbd23d0-a6b6-4a4e-b418-07fc38e5f497" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.852648 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-t56zb" event={"ID":"1051665a-7918-487f-ad08-ff6d53d113ca","Type":"ContainerStarted","Data":"285396aa0a627994dfa8dc79da6f0172519b198dbbd992855440f79f5aa66cb5"} Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.872463 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-ntmms"] Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.897087 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:42:36 crc kubenswrapper[4802]: E1206 03:42:36.897460 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:42:37.397439193 +0000 UTC m=+150.269348345 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.897914 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.898231 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-s5b86" event={"ID":"f57cc683-84dc-4538-96fe-9a198a5b008f","Type":"ContainerStarted","Data":"b0a3144c1fd1d280aa419880d78ee2a0ee2e8065892a3810d392e9dedda21c14"} Dec 06 03:42:36 crc kubenswrapper[4802]: E1206 03:42:36.898277 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:42:37.398263595 +0000 UTC m=+150.270172747 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-blwnp" (UID: "d13a541e-f90d-4222-8486-2d47854ac8e5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.909512 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-9sxg5"] Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.941360 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jwc8v" event={"ID":"e61f4275-03bf-4cb0-bb9a-e519c41f159a","Type":"ContainerStarted","Data":"935ea5241886e3d5dd733649673dc9dfec080fa6fa0c1c5332b2e3c566c075a8"} Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.953173 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-thckl" event={"ID":"4443d2fe-129a-41eb-b7b7-a7ee9bd3b6eb","Type":"ContainerStarted","Data":"475a185e71a45bf483eb92ed2732934c8e422747dad873408b3e22bb06bf4a0c"} Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.953217 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-thckl" event={"ID":"4443d2fe-129a-41eb-b7b7-a7ee9bd3b6eb","Type":"ContainerStarted","Data":"b153e4f8ac9a048cf080d7f0704c26237aaa9b950381007e62b1d6da289bb414"} Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.967422 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2nslt"] Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.972437 4802 generic.go:334] "Generic (PLEG): container finished" podID="b8455642-8313-4539-9729-fe658ef0e4d6" containerID="db3cc7b319c869abb88b7d2dff3c18149b827c25ec9fdfb2c4d00a3f0683a59f" exitCode=0 Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.972725 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-8xz4m" event={"ID":"b8455642-8313-4539-9729-fe658ef0e4d6","Type":"ContainerDied","Data":"db3cc7b319c869abb88b7d2dff3c18149b827c25ec9fdfb2c4d00a3f0683a59f"} Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.972764 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-8xz4m" event={"ID":"b8455642-8313-4539-9729-fe658ef0e4d6","Type":"ContainerStarted","Data":"f19624032ac90a317f09a04b6f5c41d6b117b6ad7caf18a9f26ac4b3fb2d30c3"} Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.989813 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"ce7807905d18be82446ae9507fbf01562f38ad9da6274e0b282e4b1346a3ee8f"} Dec 06 03:42:36 crc kubenswrapper[4802]: I1206 03:42:36.989857 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"3acf17c85df1ca0ac1074c2412b3b34c1efb1fdead9a28fcad6aa971265dcfda"} Dec 06 03:42:37 crc kubenswrapper[4802]: I1206 03:42:37.003426 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:42:37 crc kubenswrapper[4802]: E1206 03:42:37.004430 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:42:37.504414167 +0000 UTC m=+150.376323319 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:37 crc kubenswrapper[4802]: I1206 03:42:37.016409 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x6zsh" event={"ID":"9454f4b6-e1f9-4dc1-953c-442dbd4088ad","Type":"ContainerStarted","Data":"5ed77a39aa6bb5ce067c730077adc1e1f51f51969b2103016a03e60e85ef3c6f"} Dec 06 03:42:37 crc kubenswrapper[4802]: I1206 03:42:37.019039 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-7ckfj" podStartSLOduration=131.01902561 podStartE2EDuration="2m11.01902561s" podCreationTimestamp="2025-12-06 03:40:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:42:37.007926122 +0000 UTC m=+149.879835274" watchObservedRunningTime="2025-12-06 03:42:37.01902561 +0000 UTC m=+149.890934762" Dec 06 03:42:37 crc kubenswrapper[4802]: I1206 03:42:37.053077 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gmvnz"] Dec 06 03:42:37 crc kubenswrapper[4802]: I1206 03:42:37.075341 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-whhs8"] Dec 06 03:42:37 crc kubenswrapper[4802]: I1206 03:42:37.085156 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wbjzw"] Dec 06 03:42:37 crc kubenswrapper[4802]: I1206 03:42:37.106360 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:37 crc kubenswrapper[4802]: E1206 03:42:37.107957 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:42:37.6079427 +0000 UTC m=+150.479851852 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-blwnp" (UID: "d13a541e-f90d-4222-8486-2d47854ac8e5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:37 crc kubenswrapper[4802]: I1206 03:42:37.115641 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-5kgv6"] Dec 06 03:42:37 crc kubenswrapper[4802]: I1206 03:42:37.207797 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:42:37 crc kubenswrapper[4802]: E1206 03:42:37.208702 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:42:37.708685507 +0000 UTC m=+150.580594659 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:37 crc kubenswrapper[4802]: I1206 03:42:37.217937 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-rrhqx" podStartSLOduration=130.217919085 podStartE2EDuration="2m10.217919085s" podCreationTimestamp="2025-12-06 03:40:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:42:37.215284304 +0000 UTC m=+150.087193456" watchObservedRunningTime="2025-12-06 03:42:37.217919085 +0000 UTC m=+150.089828237" Dec 06 03:42:37 crc kubenswrapper[4802]: W1206 03:42:37.264493 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda8dc72d5_549b_40ad_9343_4f402a8e8572.slice/crio-a8fe1e82894dae8d3d723718dad709d42ca014769b6fb46fad5e6bca9ab6d3d2 WatchSource:0}: Error finding container a8fe1e82894dae8d3d723718dad709d42ca014769b6fb46fad5e6bca9ab6d3d2: Status 404 returned error can't find the container with id a8fe1e82894dae8d3d723718dad709d42ca014769b6fb46fad5e6bca9ab6d3d2 Dec 06 03:42:37 crc kubenswrapper[4802]: I1206 03:42:37.311469 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:37 crc kubenswrapper[4802]: E1206 03:42:37.311731 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:42:37.811719826 +0000 UTC m=+150.683628978 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-blwnp" (UID: "d13a541e-f90d-4222-8486-2d47854ac8e5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:37 crc kubenswrapper[4802]: I1206 03:42:37.349218 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-rqs9q" podStartSLOduration=131.349202453 podStartE2EDuration="2m11.349202453s" podCreationTimestamp="2025-12-06 03:40:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:42:37.3472009 +0000 UTC m=+150.219110052" watchObservedRunningTime="2025-12-06 03:42:37.349202453 +0000 UTC m=+150.221111605" Dec 06 03:42:37 crc kubenswrapper[4802]: I1206 03:42:37.417653 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:42:37 crc kubenswrapper[4802]: E1206 03:42:37.418113 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:42:37.918097295 +0000 UTC m=+150.790006447 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:37 crc kubenswrapper[4802]: I1206 03:42:37.524593 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:37 crc kubenswrapper[4802]: E1206 03:42:37.525205 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:42:38.025190553 +0000 UTC m=+150.897099705 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-blwnp" (UID: "d13a541e-f90d-4222-8486-2d47854ac8e5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:37 crc kubenswrapper[4802]: I1206 03:42:37.628438 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:42:37 crc kubenswrapper[4802]: E1206 03:42:37.628614 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:42:38.128586862 +0000 UTC m=+151.000496014 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:37 crc kubenswrapper[4802]: I1206 03:42:37.628905 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:37 crc kubenswrapper[4802]: E1206 03:42:37.629173 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:42:38.129165357 +0000 UTC m=+151.001074509 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-blwnp" (UID: "d13a541e-f90d-4222-8486-2d47854ac8e5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:37 crc kubenswrapper[4802]: I1206 03:42:37.714479 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-7cg8t" podStartSLOduration=131.7144645 podStartE2EDuration="2m11.7144645s" podCreationTimestamp="2025-12-06 03:40:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:42:37.668967667 +0000 UTC m=+150.540876819" watchObservedRunningTime="2025-12-06 03:42:37.7144645 +0000 UTC m=+150.586373652" Dec 06 03:42:37 crc kubenswrapper[4802]: I1206 03:42:37.731285 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:42:37 crc kubenswrapper[4802]: E1206 03:42:37.731627 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:42:38.231610731 +0000 UTC m=+151.103519883 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:37 crc kubenswrapper[4802]: I1206 03:42:37.760081 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-t56zb" podStartSLOduration=131.760065985 podStartE2EDuration="2m11.760065985s" podCreationTimestamp="2025-12-06 03:40:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:42:37.75690725 +0000 UTC m=+150.628816402" watchObservedRunningTime="2025-12-06 03:42:37.760065985 +0000 UTC m=+150.631975127" Dec 06 03:42:37 crc kubenswrapper[4802]: I1206 03:42:37.761327 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fqdzj" podStartSLOduration=131.761317648 podStartE2EDuration="2m11.761317648s" podCreationTimestamp="2025-12-06 03:40:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:42:37.715147158 +0000 UTC m=+150.587056310" watchObservedRunningTime="2025-12-06 03:42:37.761317648 +0000 UTC m=+150.633226800" Dec 06 03:42:37 crc kubenswrapper[4802]: I1206 03:42:37.804038 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ppzvl" podStartSLOduration=130.804023817 podStartE2EDuration="2m10.804023817s" podCreationTimestamp="2025-12-06 03:40:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:42:37.781894402 +0000 UTC m=+150.653803554" watchObservedRunningTime="2025-12-06 03:42:37.804023817 +0000 UTC m=+150.675932969" Dec 06 03:42:37 crc kubenswrapper[4802]: I1206 03:42:37.806130 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-thckl" podStartSLOduration=131.806118772 podStartE2EDuration="2m11.806118772s" podCreationTimestamp="2025-12-06 03:40:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:42:37.802214428 +0000 UTC m=+150.674123570" watchObservedRunningTime="2025-12-06 03:42:37.806118772 +0000 UTC m=+150.678027944" Dec 06 03:42:37 crc kubenswrapper[4802]: I1206 03:42:37.832913 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:37 crc kubenswrapper[4802]: E1206 03:42:37.833334 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:42:38.333318694 +0000 UTC m=+151.205227856 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-blwnp" (UID: "d13a541e-f90d-4222-8486-2d47854ac8e5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:37 crc kubenswrapper[4802]: I1206 03:42:37.840980 4802 patch_prober.go:28] interesting pod/router-default-5444994796-rrhqx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 03:42:37 crc kubenswrapper[4802]: [-]has-synced failed: reason withheld Dec 06 03:42:37 crc kubenswrapper[4802]: [+]process-running ok Dec 06 03:42:37 crc kubenswrapper[4802]: healthz check failed Dec 06 03:42:37 crc kubenswrapper[4802]: I1206 03:42:37.841036 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-rrhqx" podUID="0fbd23d0-a6b6-4a4e-b418-07fc38e5f497" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 03:42:37 crc kubenswrapper[4802]: I1206 03:42:37.933271 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:42:37 crc kubenswrapper[4802]: E1206 03:42:37.933631 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:42:38.433618309 +0000 UTC m=+151.305527451 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:38 crc kubenswrapper[4802]: I1206 03:42:37.993608 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-cgpdh" podStartSLOduration=131.993589551 podStartE2EDuration="2m11.993589551s" podCreationTimestamp="2025-12-06 03:40:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:42:37.985347599 +0000 UTC m=+150.857256751" watchObservedRunningTime="2025-12-06 03:42:37.993589551 +0000 UTC m=+150.865498693" Dec 06 03:42:38 crc kubenswrapper[4802]: I1206 03:42:38.005936 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jv24g" podStartSLOduration=132.005918012 podStartE2EDuration="2m12.005918012s" podCreationTimestamp="2025-12-06 03:40:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:42:37.957326566 +0000 UTC m=+150.829235718" watchObservedRunningTime="2025-12-06 03:42:38.005918012 +0000 UTC m=+150.877827164" Dec 06 03:42:38 crc kubenswrapper[4802]: I1206 03:42:38.036034 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:38 crc kubenswrapper[4802]: E1206 03:42:38.036333 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:42:38.53631901 +0000 UTC m=+151.408228162 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-blwnp" (UID: "d13a541e-f90d-4222-8486-2d47854ac8e5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:38 crc kubenswrapper[4802]: I1206 03:42:38.129361 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-zb4m6" event={"ID":"b78b4d0e-a83c-4339-b1c4-8998e69e7218","Type":"ContainerStarted","Data":"202220e0213c9b942458ff044a2bd893be05faf607966815aa388132230cfe9b"} Dec 06 03:42:38 crc kubenswrapper[4802]: I1206 03:42:38.129410 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-zb4m6" event={"ID":"b78b4d0e-a83c-4339-b1c4-8998e69e7218","Type":"ContainerStarted","Data":"3df37580ff171e5298a63ec8b8ac789c6a201c4178e32022111ee43ae3401b25"} Dec 06 03:42:38 crc kubenswrapper[4802]: I1206 03:42:38.138124 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:42:38 crc kubenswrapper[4802]: E1206 03:42:38.139133 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:42:38.639119332 +0000 UTC m=+151.511028484 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:38 crc kubenswrapper[4802]: I1206 03:42:38.154239 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416530-88l5w" event={"ID":"19ade530-d0de-44d1-be83-96136753a35e","Type":"ContainerStarted","Data":"6f0a1e014d7145ba9b9f1a0e3e33e00d6baa3616ed4444a2de7bf58db8cda08e"} Dec 06 03:42:38 crc kubenswrapper[4802]: I1206 03:42:38.187674 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qfdsg" event={"ID":"e4f6934a-db7e-4031-baa8-2f41dc8c2c18","Type":"ContainerStarted","Data":"c1c25e946ba23e5bdcc96c11585acdf3ec3db41bc4ee819794796fd11be12f8c"} Dec 06 03:42:38 crc kubenswrapper[4802]: I1206 03:42:38.241606 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:38 crc kubenswrapper[4802]: E1206 03:42:38.242180 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:42:38.742169501 +0000 UTC m=+151.614078653 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-blwnp" (UID: "d13a541e-f90d-4222-8486-2d47854ac8e5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:38 crc kubenswrapper[4802]: I1206 03:42:38.341386 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jwc8v" event={"ID":"e61f4275-03bf-4cb0-bb9a-e519c41f159a","Type":"ContainerStarted","Data":"7b40ccd5497f5611e7c0c4c0d7c24c48403c060fac7029d414e2e1654bc87eb7"} Dec 06 03:42:38 crc kubenswrapper[4802]: I1206 03:42:38.341449 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jwc8v" event={"ID":"e61f4275-03bf-4cb0-bb9a-e519c41f159a","Type":"ContainerStarted","Data":"14a161445cc8237a03a7dfb360abacd06a8e869ff5c84cd25537812355d6e237"} Dec 06 03:42:38 crc kubenswrapper[4802]: I1206 03:42:38.344570 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:42:38 crc kubenswrapper[4802]: E1206 03:42:38.345717 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:42:38.845702834 +0000 UTC m=+151.717611986 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:38 crc kubenswrapper[4802]: I1206 03:42:38.425584 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2nslt" event={"ID":"a668f1b9-b7e6-4c39-9fc2-9beb5e5e5bce","Type":"ContainerStarted","Data":"5e8d89628c515ed48fde67a58f891fc9f36595e4c2be34f9581e7df8c4903f86"} Dec 06 03:42:38 crc kubenswrapper[4802]: I1206 03:42:38.428629 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2nslt" Dec 06 03:42:38 crc kubenswrapper[4802]: I1206 03:42:38.432211 4802 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-2nslt container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.26:5443/healthz\": dial tcp 10.217.0.26:5443: connect: connection refused" start-of-body= Dec 06 03:42:38 crc kubenswrapper[4802]: I1206 03:42:38.432258 4802 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2nslt" podUID="a668f1b9-b7e6-4c39-9fc2-9beb5e5e5bce" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.26:5443/healthz\": dial tcp 10.217.0.26:5443: connect: connection refused" Dec 06 03:42:38 crc kubenswrapper[4802]: I1206 03:42:38.446509 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-ntmms" event={"ID":"f7feaf6a-2ad5-4ef3-88d1-a2a483e5f0d9","Type":"ContainerStarted","Data":"87493b2790b6bd952fa47cc019ea74030c02a88278b7b247823046ef3f939e82"} Dec 06 03:42:38 crc kubenswrapper[4802]: I1206 03:42:38.446764 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-ntmms" event={"ID":"f7feaf6a-2ad5-4ef3-88d1-a2a483e5f0d9","Type":"ContainerStarted","Data":"4a86359bb18361486cc0ae26820830f25ef500b68c9a224471f04251580a7948"} Dec 06 03:42:38 crc kubenswrapper[4802]: I1206 03:42:38.447375 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:38 crc kubenswrapper[4802]: E1206 03:42:38.447651 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:42:38.947639033 +0000 UTC m=+151.819548185 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-blwnp" (UID: "d13a541e-f90d-4222-8486-2d47854ac8e5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:38 crc kubenswrapper[4802]: I1206 03:42:38.463280 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-5kgv6" event={"ID":"a8dc72d5-549b-40ad-9343-4f402a8e8572","Type":"ContainerStarted","Data":"a8fe1e82894dae8d3d723718dad709d42ca014769b6fb46fad5e6bca9ab6d3d2"} Dec 06 03:42:38 crc kubenswrapper[4802]: I1206 03:42:38.481951 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5rwsv" event={"ID":"6258654e-b524-4fdd-ba9b-7443b57c2585","Type":"ContainerStarted","Data":"b0b21d45a39ad34df8be48e978ad1317197758620cd3f2095f7e36980db38207"} Dec 06 03:42:38 crc kubenswrapper[4802]: I1206 03:42:38.488778 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-wrh75" event={"ID":"7621a355-6a10-4143-b35e-f1f617cbf849","Type":"ContainerStarted","Data":"ea22133c68f130939b317293a80e1947baef38a8be766306ecea7426ac192e95"} Dec 06 03:42:38 crc kubenswrapper[4802]: I1206 03:42:38.517453 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9bfz8" event={"ID":"b2994e8f-6482-4dca-baf1-148754e78e85","Type":"ContainerStarted","Data":"203847e667addaa0d742bd7616257e1f43b8a93c33b7b6226e02453ac95eb242"} Dec 06 03:42:38 crc kubenswrapper[4802]: I1206 03:42:38.518533 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9bfz8" Dec 06 03:42:38 crc kubenswrapper[4802]: I1206 03:42:38.523246 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mpnf2" event={"ID":"2610e572-bd43-4029-a846-b6694c1896f7","Type":"ContainerStarted","Data":"1e1c210df9ce69f02eda8f9785e38491b2789f80287a4e94906d5446ad12c57e"} Dec 06 03:42:38 crc kubenswrapper[4802]: I1206 03:42:38.523273 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mpnf2" event={"ID":"2610e572-bd43-4029-a846-b6694c1896f7","Type":"ContainerStarted","Data":"9e88839ec358df358175f387a0c67e284216bc660e1193ec9717038436df593b"} Dec 06 03:42:38 crc kubenswrapper[4802]: I1206 03:42:38.549830 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:42:38 crc kubenswrapper[4802]: E1206 03:42:38.550921 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:42:39.050904318 +0000 UTC m=+151.922813470 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:38 crc kubenswrapper[4802]: I1206 03:42:38.552507 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9bfz8" Dec 06 03:42:38 crc kubenswrapper[4802]: I1206 03:42:38.559275 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ppzvl" event={"ID":"a92fd9bb-6d1b-4ab1-9f1a-bc9086ecb08e","Type":"ContainerStarted","Data":"82f196d484982c9178c46cbb1d1638c8339a7d2b39db5c838eb6e10bfe257d65"} Dec 06 03:42:38 crc kubenswrapper[4802]: I1206 03:42:38.564167 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4gkmd" event={"ID":"78eab4ec-5f1e-42cd-8ce7-48f76a55cc95","Type":"ContainerStarted","Data":"017f041763dc35d37e587e623123ec56443f9c531ddee2f929cfe8b43c3128c7"} Dec 06 03:42:38 crc kubenswrapper[4802]: I1206 03:42:38.564196 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4gkmd" event={"ID":"78eab4ec-5f1e-42cd-8ce7-48f76a55cc95","Type":"ContainerStarted","Data":"4924b71402f3ac6d2f3df3c01a67a238ac3da4d49d46218ad0f5a4b9036c85fe"} Dec 06 03:42:38 crc kubenswrapper[4802]: I1206 03:42:38.564724 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4gkmd" Dec 06 03:42:38 crc kubenswrapper[4802]: I1206 03:42:38.568927 4802 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-4gkmd container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.19:8443/healthz\": dial tcp 10.217.0.19:8443: connect: connection refused" start-of-body= Dec 06 03:42:38 crc kubenswrapper[4802]: I1206 03:42:38.568969 4802 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4gkmd" podUID="78eab4ec-5f1e-42cd-8ce7-48f76a55cc95" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.19:8443/healthz\": dial tcp 10.217.0.19:8443: connect: connection refused" Dec 06 03:42:38 crc kubenswrapper[4802]: I1206 03:42:38.609716 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x6zsh" event={"ID":"9454f4b6-e1f9-4dc1-953c-442dbd4088ad","Type":"ContainerStarted","Data":"ae9975479744ee84bd691f537388950c1c2ddd6098bdf59deae3d07d596666ba"} Dec 06 03:42:38 crc kubenswrapper[4802]: I1206 03:42:38.637699 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-whhs8" event={"ID":"ddb4b244-e4e0-4ae4-a217-c6531fc90b6a","Type":"ContainerStarted","Data":"dfc2c34baf8cb30f7a6888e5aa5153e98841553fe0e4fc3d6e043940543b99d0"} Dec 06 03:42:38 crc kubenswrapper[4802]: I1206 03:42:38.651653 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:38 crc kubenswrapper[4802]: E1206 03:42:38.664087 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:42:39.16407093 +0000 UTC m=+152.035980082 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-blwnp" (UID: "d13a541e-f90d-4222-8486-2d47854ac8e5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:38 crc kubenswrapper[4802]: I1206 03:42:38.719057 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-9sxg5" event={"ID":"407d6ced-2671-40de-aa70-153cfda69fba","Type":"ContainerStarted","Data":"4b2222cfa7cdfcf20bd90eb838307f8d8df071349b5ba7f0c0ce4fca6b075a54"} Dec 06 03:42:38 crc kubenswrapper[4802]: I1206 03:42:38.753820 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5rwsv" podStartSLOduration=131.753802411 podStartE2EDuration="2m11.753802411s" podCreationTimestamp="2025-12-06 03:40:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:42:38.740484983 +0000 UTC m=+151.612394135" watchObservedRunningTime="2025-12-06 03:42:38.753802411 +0000 UTC m=+151.625711563" Dec 06 03:42:38 crc kubenswrapper[4802]: I1206 03:42:38.755320 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mpnf2" podStartSLOduration=131.755313381 podStartE2EDuration="2m11.755313381s" podCreationTimestamp="2025-12-06 03:40:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:42:38.661729697 +0000 UTC m=+151.533638849" watchObservedRunningTime="2025-12-06 03:42:38.755313381 +0000 UTC m=+151.627222533" Dec 06 03:42:38 crc kubenswrapper[4802]: I1206 03:42:38.756057 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:42:38 crc kubenswrapper[4802]: E1206 03:42:38.756319 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:42:39.256308239 +0000 UTC m=+152.128217391 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:38 crc kubenswrapper[4802]: I1206 03:42:38.785263 4802 generic.go:334] "Generic (PLEG): container finished" podID="f57cc683-84dc-4538-96fe-9a198a5b008f" containerID="442cb2781fdd4a7f9197bbb996173d2877cd37038d8eae8e14a4a9bc4095b73d" exitCode=0 Dec 06 03:42:38 crc kubenswrapper[4802]: I1206 03:42:38.785322 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-s5b86" event={"ID":"f57cc683-84dc-4538-96fe-9a198a5b008f","Type":"ContainerDied","Data":"442cb2781fdd4a7f9197bbb996173d2877cd37038d8eae8e14a4a9bc4095b73d"} Dec 06 03:42:38 crc kubenswrapper[4802]: I1206 03:42:38.799509 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-zfjwx" event={"ID":"8002be53-a1e5-4ef7-93b4-384d9b16a7e2","Type":"ContainerStarted","Data":"356cb1441926040e0a44a617835d67c37ca437a48c568881ea8c739f4e825326"} Dec 06 03:42:38 crc kubenswrapper[4802]: I1206 03:42:38.799565 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-zfjwx" event={"ID":"8002be53-a1e5-4ef7-93b4-384d9b16a7e2","Type":"ContainerStarted","Data":"3520510526edab6aa582720104e8604e6bf748319b9a31a752ef2425840336f2"} Dec 06 03:42:38 crc kubenswrapper[4802]: I1206 03:42:38.826372 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fnpz5" event={"ID":"93ffc5bd-20a8-450c-861a-0aecc58d602b","Type":"ContainerStarted","Data":"beb2ad71cf9dc4ffb91496324b100743d9fd3cea13b8c03e058996837a427901"} Dec 06 03:42:38 crc kubenswrapper[4802]: I1206 03:42:38.834135 4802 patch_prober.go:28] interesting pod/router-default-5444994796-rrhqx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 03:42:38 crc kubenswrapper[4802]: [-]has-synced failed: reason withheld Dec 06 03:42:38 crc kubenswrapper[4802]: [+]process-running ok Dec 06 03:42:38 crc kubenswrapper[4802]: healthz check failed Dec 06 03:42:38 crc kubenswrapper[4802]: I1206 03:42:38.834198 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-rrhqx" podUID="0fbd23d0-a6b6-4a4e-b418-07fc38e5f497" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 03:42:38 crc kubenswrapper[4802]: I1206 03:42:38.848937 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wbjzw" event={"ID":"18caf043-1bf2-4e67-bb3b-e4edf9fc8822","Type":"ContainerStarted","Data":"e16bf1b3c275c0645eb8b9ae8182b10670d5de965d49eb53c8532e447f5d3b77"} Dec 06 03:42:38 crc kubenswrapper[4802]: I1206 03:42:38.858787 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:38 crc kubenswrapper[4802]: E1206 03:42:38.859766 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:42:39.359733487 +0000 UTC m=+152.231642639 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-blwnp" (UID: "d13a541e-f90d-4222-8486-2d47854ac8e5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:38 crc kubenswrapper[4802]: I1206 03:42:38.926161 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9rvnv" event={"ID":"ff639bee-182b-4dd2-aabc-d396a3bd9ce5","Type":"ContainerStarted","Data":"c56cb007336c706550fdf398a80960ea5fd1c4788d3a9f63cfb156306e6a77af"} Dec 06 03:42:38 crc kubenswrapper[4802]: I1206 03:42:38.955155 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gmvnz" event={"ID":"c2adc97c-46b4-4de7-b49c-5f75a16c2c32","Type":"ContainerStarted","Data":"dd7ff208a25790188a79be343157bdf8fd20edf329f1f8ba18ef1b048e7923ee"} Dec 06 03:42:38 crc kubenswrapper[4802]: I1206 03:42:38.961404 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jwc8v" podStartSLOduration=131.96138932 podStartE2EDuration="2m11.96138932s" podCreationTimestamp="2025-12-06 03:40:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:42:38.951402441 +0000 UTC m=+151.823311593" watchObservedRunningTime="2025-12-06 03:42:38.96138932 +0000 UTC m=+151.833298472" Dec 06 03:42:38 crc kubenswrapper[4802]: I1206 03:42:38.961446 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:42:38 crc kubenswrapper[4802]: E1206 03:42:38.961836 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:42:39.461824571 +0000 UTC m=+152.333733723 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:38 crc kubenswrapper[4802]: I1206 03:42:38.972547 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-9s5r8" event={"ID":"eb15a464-d0e8-4e49-b018-14353cc9d6cb","Type":"ContainerStarted","Data":"b97cea68d77f3a16f099b3348de1e075e19c124bdb8a26a54f11b517a4678d50"} Dec 06 03:42:38 crc kubenswrapper[4802]: I1206 03:42:38.989581 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-gkpck" event={"ID":"7abae297-572a-4c3c-8e7d-4e2157d838ad","Type":"ContainerStarted","Data":"af6ea2a9ae3febbf48417e31bb710ddc29c1f38c1b1b26adf0d9afa5ec056743"} Dec 06 03:42:39 crc kubenswrapper[4802]: I1206 03:42:39.005665 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-wrh75" podStartSLOduration=132.005647179 podStartE2EDuration="2m12.005647179s" podCreationTimestamp="2025-12-06 03:40:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:42:39.005179876 +0000 UTC m=+151.877089028" watchObservedRunningTime="2025-12-06 03:42:39.005647179 +0000 UTC m=+151.877556351" Dec 06 03:42:39 crc kubenswrapper[4802]: I1206 03:42:39.013528 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8br44" event={"ID":"bd64a08a-6920-4a0f-b299-aa5a009236a1","Type":"ContainerStarted","Data":"23ebe414de2ba610e832640a9d7d3955f64a011198f0c675baae13334f5fb24a"} Dec 06 03:42:39 crc kubenswrapper[4802]: I1206 03:42:39.036464 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6d29n" event={"ID":"1a34c64b-194e-41a6-8b55-0a87187c4cfc","Type":"ContainerStarted","Data":"3f43ae33150400528ea8f961b52510bbc58619d9a74b26fa637b05fd074d4d73"} Dec 06 03:42:39 crc kubenswrapper[4802]: I1206 03:42:39.069583 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:39 crc kubenswrapper[4802]: E1206 03:42:39.071002 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:42:39.570976065 +0000 UTC m=+152.442885227 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-blwnp" (UID: "d13a541e-f90d-4222-8486-2d47854ac8e5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:39 crc kubenswrapper[4802]: I1206 03:42:39.075234 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9bfz8" podStartSLOduration=132.075219018 podStartE2EDuration="2m12.075219018s" podCreationTimestamp="2025-12-06 03:40:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:42:39.07416294 +0000 UTC m=+151.946072082" watchObservedRunningTime="2025-12-06 03:42:39.075219018 +0000 UTC m=+151.947128170" Dec 06 03:42:39 crc kubenswrapper[4802]: I1206 03:42:39.079568 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ddpmx" event={"ID":"ec27be85-6d9d-4d96-b3b8-b78a7a941acb","Type":"ContainerStarted","Data":"bf69b7afe30f3da1ad1bf2efa2ab232935a02b768403437dfc405d6d91f019da"} Dec 06 03:42:39 crc kubenswrapper[4802]: I1206 03:42:39.079644 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ddpmx" event={"ID":"ec27be85-6d9d-4d96-b3b8-b78a7a941acb","Type":"ContainerStarted","Data":"105affda22265051b61aca4aeaf6bc4c0dbb6cba1492e6b2f5c6a830948846bb"} Dec 06 03:42:39 crc kubenswrapper[4802]: I1206 03:42:39.080522 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-ddpmx" Dec 06 03:42:39 crc kubenswrapper[4802]: I1206 03:42:39.087875 4802 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-ddpmx container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.41:8080/healthz\": dial tcp 10.217.0.41:8080: connect: connection refused" start-of-body= Dec 06 03:42:39 crc kubenswrapper[4802]: I1206 03:42:39.087925 4802 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-ddpmx" podUID="ec27be85-6d9d-4d96-b3b8-b78a7a941acb" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.41:8080/healthz\": dial tcp 10.217.0.41:8080: connect: connection refused" Dec 06 03:42:39 crc kubenswrapper[4802]: I1206 03:42:39.089801 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-khkwh" event={"ID":"293e299a-d25d-4755-8d9d-4d8f224c8791","Type":"ContainerStarted","Data":"81764adbd9ca14167c1d7471a5494e8dbf043ef071acb59e2cecd39bd5b9b152"} Dec 06 03:42:39 crc kubenswrapper[4802]: I1206 03:42:39.092257 4802 patch_prober.go:28] interesting pod/downloads-7954f5f757-7cg8t container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Dec 06 03:42:39 crc kubenswrapper[4802]: I1206 03:42:39.092305 4802 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-7cg8t" podUID="0057cff6-6085-4b41-bf4f-e1121bece99a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Dec 06 03:42:39 crc kubenswrapper[4802]: I1206 03:42:39.110629 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-cgpdh" Dec 06 03:42:39 crc kubenswrapper[4802]: I1206 03:42:39.115217 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8br44" Dec 06 03:42:39 crc kubenswrapper[4802]: I1206 03:42:39.115477 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8br44" Dec 06 03:42:39 crc kubenswrapper[4802]: I1206 03:42:39.134377 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8br44" Dec 06 03:42:39 crc kubenswrapper[4802]: I1206 03:42:39.175000 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:42:39 crc kubenswrapper[4802]: E1206 03:42:39.175912 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:42:39.675897274 +0000 UTC m=+152.547806426 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:39 crc kubenswrapper[4802]: I1206 03:42:39.195049 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-5kgv6" podStartSLOduration=7.195031158 podStartE2EDuration="7.195031158s" podCreationTimestamp="2025-12-06 03:42:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:42:39.160189172 +0000 UTC m=+152.032098324" watchObservedRunningTime="2025-12-06 03:42:39.195031158 +0000 UTC m=+152.066940310" Dec 06 03:42:39 crc kubenswrapper[4802]: I1206 03:42:39.231641 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-zb4m6" podStartSLOduration=132.231622752 podStartE2EDuration="2m12.231622752s" podCreationTimestamp="2025-12-06 03:40:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:42:39.196798476 +0000 UTC m=+152.068707628" watchObservedRunningTime="2025-12-06 03:42:39.231622752 +0000 UTC m=+152.103531904" Dec 06 03:42:39 crc kubenswrapper[4802]: I1206 03:42:39.232684 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qfdsg" podStartSLOduration=132.232677881 podStartE2EDuration="2m12.232677881s" podCreationTimestamp="2025-12-06 03:40:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:42:39.226912036 +0000 UTC m=+152.098821178" watchObservedRunningTime="2025-12-06 03:42:39.232677881 +0000 UTC m=+152.104587043" Dec 06 03:42:39 crc kubenswrapper[4802]: I1206 03:42:39.277316 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:39 crc kubenswrapper[4802]: E1206 03:42:39.280016 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:42:39.779996581 +0000 UTC m=+152.651905813 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-blwnp" (UID: "d13a541e-f90d-4222-8486-2d47854ac8e5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:39 crc kubenswrapper[4802]: I1206 03:42:39.308713 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29416530-88l5w" podStartSLOduration=133.308695973 podStartE2EDuration="2m13.308695973s" podCreationTimestamp="2025-12-06 03:40:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:42:39.261877595 +0000 UTC m=+152.133786747" watchObservedRunningTime="2025-12-06 03:42:39.308695973 +0000 UTC m=+152.180605125" Dec 06 03:42:39 crc kubenswrapper[4802]: I1206 03:42:39.342323 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2nslt" podStartSLOduration=132.342301786 podStartE2EDuration="2m12.342301786s" podCreationTimestamp="2025-12-06 03:40:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:42:39.310351507 +0000 UTC m=+152.182260669" watchObservedRunningTime="2025-12-06 03:42:39.342301786 +0000 UTC m=+152.214210939" Dec 06 03:42:39 crc kubenswrapper[4802]: I1206 03:42:39.370966 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gmvnz" podStartSLOduration=133.370950406 podStartE2EDuration="2m13.370950406s" podCreationTimestamp="2025-12-06 03:40:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:42:39.343400696 +0000 UTC m=+152.215309848" watchObservedRunningTime="2025-12-06 03:42:39.370950406 +0000 UTC m=+152.242859558" Dec 06 03:42:39 crc kubenswrapper[4802]: I1206 03:42:39.372571 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fnpz5" podStartSLOduration=134.372559529 podStartE2EDuration="2m14.372559529s" podCreationTimestamp="2025-12-06 03:40:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:42:39.370650088 +0000 UTC m=+152.242559240" watchObservedRunningTime="2025-12-06 03:42:39.372559529 +0000 UTC m=+152.244468681" Dec 06 03:42:39 crc kubenswrapper[4802]: I1206 03:42:39.380280 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:42:39 crc kubenswrapper[4802]: E1206 03:42:39.380887 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:42:39.880868372 +0000 UTC m=+152.752777534 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:39 crc kubenswrapper[4802]: I1206 03:42:39.455227 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9rvnv" podStartSLOduration=132.455215691 podStartE2EDuration="2m12.455215691s" podCreationTimestamp="2025-12-06 03:40:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:42:39.403855101 +0000 UTC m=+152.275764253" watchObservedRunningTime="2025-12-06 03:42:39.455215691 +0000 UTC m=+152.327124843" Dec 06 03:42:39 crc kubenswrapper[4802]: I1206 03:42:39.457116 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6d29n" podStartSLOduration=132.457106912 podStartE2EDuration="2m12.457106912s" podCreationTimestamp="2025-12-06 03:40:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:42:39.454363958 +0000 UTC m=+152.326273110" watchObservedRunningTime="2025-12-06 03:42:39.457106912 +0000 UTC m=+152.329016064" Dec 06 03:42:39 crc kubenswrapper[4802]: I1206 03:42:39.482155 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:39 crc kubenswrapper[4802]: E1206 03:42:39.482494 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:42:39.982478764 +0000 UTC m=+152.854387906 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-blwnp" (UID: "d13a541e-f90d-4222-8486-2d47854ac8e5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:39 crc kubenswrapper[4802]: I1206 03:42:39.557646 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-9s5r8" podStartSLOduration=132.557625633 podStartE2EDuration="2m12.557625633s" podCreationTimestamp="2025-12-06 03:40:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:42:39.552250249 +0000 UTC m=+152.424159401" watchObservedRunningTime="2025-12-06 03:42:39.557625633 +0000 UTC m=+152.429534785" Dec 06 03:42:39 crc kubenswrapper[4802]: I1206 03:42:39.583206 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:42:39 crc kubenswrapper[4802]: E1206 03:42:39.583681 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:42:40.083633192 +0000 UTC m=+152.955542344 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:39 crc kubenswrapper[4802]: I1206 03:42:39.617384 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4gkmd" podStartSLOduration=132.617363198 podStartE2EDuration="2m12.617363198s" podCreationTimestamp="2025-12-06 03:40:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:42:39.616260578 +0000 UTC m=+152.488169730" watchObservedRunningTime="2025-12-06 03:42:39.617363198 +0000 UTC m=+152.489272350" Dec 06 03:42:39 crc kubenswrapper[4802]: I1206 03:42:39.688137 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:39 crc kubenswrapper[4802]: E1206 03:42:39.688392 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:42:40.188380557 +0000 UTC m=+153.060289709 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-blwnp" (UID: "d13a541e-f90d-4222-8486-2d47854ac8e5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:39 crc kubenswrapper[4802]: I1206 03:42:39.690700 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-zfjwx" podStartSLOduration=132.690688899 podStartE2EDuration="2m12.690688899s" podCreationTimestamp="2025-12-06 03:40:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:42:39.688735336 +0000 UTC m=+152.560644488" watchObservedRunningTime="2025-12-06 03:42:39.690688899 +0000 UTC m=+152.562598051" Dec 06 03:42:39 crc kubenswrapper[4802]: I1206 03:42:39.718200 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-gkpck" podStartSLOduration=133.718182198 podStartE2EDuration="2m13.718182198s" podCreationTimestamp="2025-12-06 03:40:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:42:39.717615243 +0000 UTC m=+152.589524385" watchObservedRunningTime="2025-12-06 03:42:39.718182198 +0000 UTC m=+152.590091350" Dec 06 03:42:39 crc kubenswrapper[4802]: I1206 03:42:39.756369 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-khkwh" podStartSLOduration=7.756355233 podStartE2EDuration="7.756355233s" podCreationTimestamp="2025-12-06 03:42:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:42:39.755129981 +0000 UTC m=+152.627039133" watchObservedRunningTime="2025-12-06 03:42:39.756355233 +0000 UTC m=+152.628264385" Dec 06 03:42:39 crc kubenswrapper[4802]: I1206 03:42:39.789064 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:42:39 crc kubenswrapper[4802]: E1206 03:42:39.789393 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:42:40.289378471 +0000 UTC m=+153.161287613 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:39 crc kubenswrapper[4802]: I1206 03:42:39.824319 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x6zsh" podStartSLOduration=132.82430721 podStartE2EDuration="2m12.82430721s" podCreationTimestamp="2025-12-06 03:40:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:42:39.82320376 +0000 UTC m=+152.695112912" watchObservedRunningTime="2025-12-06 03:42:39.82430721 +0000 UTC m=+152.696216362" Dec 06 03:42:39 crc kubenswrapper[4802]: I1206 03:42:39.830341 4802 patch_prober.go:28] interesting pod/router-default-5444994796-rrhqx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 03:42:39 crc kubenswrapper[4802]: [-]has-synced failed: reason withheld Dec 06 03:42:39 crc kubenswrapper[4802]: [+]process-running ok Dec 06 03:42:39 crc kubenswrapper[4802]: healthz check failed Dec 06 03:42:39 crc kubenswrapper[4802]: I1206 03:42:39.830389 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-rrhqx" podUID="0fbd23d0-a6b6-4a4e-b418-07fc38e5f497" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 03:42:39 crc kubenswrapper[4802]: I1206 03:42:39.890616 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:39 crc kubenswrapper[4802]: E1206 03:42:39.890933 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:42:40.39092175 +0000 UTC m=+153.262830902 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-blwnp" (UID: "d13a541e-f90d-4222-8486-2d47854ac8e5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:39 crc kubenswrapper[4802]: I1206 03:42:39.931620 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-ddpmx" podStartSLOduration=132.931605883 podStartE2EDuration="2m12.931605883s" podCreationTimestamp="2025-12-06 03:40:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:42:39.879971776 +0000 UTC m=+152.751880928" watchObservedRunningTime="2025-12-06 03:42:39.931605883 +0000 UTC m=+152.803515035" Dec 06 03:42:39 crc kubenswrapper[4802]: I1206 03:42:39.992048 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:42:39 crc kubenswrapper[4802]: E1206 03:42:39.992238 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:42:40.492198082 +0000 UTC m=+153.364107234 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:39 crc kubenswrapper[4802]: I1206 03:42:39.992360 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:39 crc kubenswrapper[4802]: E1206 03:42:39.992676 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:42:40.492668774 +0000 UTC m=+153.364577926 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-blwnp" (UID: "d13a541e-f90d-4222-8486-2d47854ac8e5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.093579 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:42:40 crc kubenswrapper[4802]: E1206 03:42:40.093765 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:42:40.59372863 +0000 UTC m=+153.465637782 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.094129 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:40 crc kubenswrapper[4802]: E1206 03:42:40.094443 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:42:40.594435779 +0000 UTC m=+153.466344931 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-blwnp" (UID: "d13a541e-f90d-4222-8486-2d47854ac8e5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.098827 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gmvnz" event={"ID":"c2adc97c-46b4-4de7-b49c-5f75a16c2c32","Type":"ContainerStarted","Data":"bc3623be719031332fe40a3491f104fef24e5a4e7aa621bcccfc3b4d8eff6aaa"} Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.105217 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-8xz4m" event={"ID":"b8455642-8313-4539-9729-fe658ef0e4d6","Type":"ContainerStarted","Data":"5278ad56bc8fccb9588e4f99f2d42397432a4252ef127e15f8af2deedd1b6542"} Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.105262 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-8xz4m" event={"ID":"b8455642-8313-4539-9729-fe658ef0e4d6","Type":"ContainerStarted","Data":"d2ce8886737affc7b7ff1f561aedfd5e5627dcb5f4382cc3bc5903350a70b6b4"} Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.139161 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-9sxg5" event={"ID":"407d6ced-2671-40de-aa70-153cfda69fba","Type":"ContainerStarted","Data":"a6205825e92c6ba20be95c5ec6a4fed064bea3fda5461a02e8a7182525bbaa26"} Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.139225 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-9sxg5" event={"ID":"407d6ced-2671-40de-aa70-153cfda69fba","Type":"ContainerStarted","Data":"220af3fbc22a604f97e9e811de2fc9db9e89d155e16b4e6655da8b551bb758b3"} Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.147087 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-s5b86" event={"ID":"f57cc683-84dc-4538-96fe-9a198a5b008f","Type":"ContainerStarted","Data":"dd5344c16b9d34a130f5060a618aab02f8073a4dcd6cbdd0dbace0a0b279e567"} Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.147201 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-s5b86" Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.152698 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-5kgv6" event={"ID":"a8dc72d5-549b-40ad-9343-4f402a8e8572","Type":"ContainerStarted","Data":"2221855bfcaa99f5e70faa243f71d66c48e4e17932ba437309bc0a74d3ab10cc"} Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.153533 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8br44" podStartSLOduration=133.153514897 podStartE2EDuration="2m13.153514897s" podCreationTimestamp="2025-12-06 03:40:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:42:39.939254879 +0000 UTC m=+152.811164031" watchObservedRunningTime="2025-12-06 03:42:40.153514897 +0000 UTC m=+153.025424039" Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.154695 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-8xz4m" podStartSLOduration=134.154687519 podStartE2EDuration="2m14.154687519s" podCreationTimestamp="2025-12-06 03:40:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:42:40.152056698 +0000 UTC m=+153.023965850" watchObservedRunningTime="2025-12-06 03:42:40.154687519 +0000 UTC m=+153.026596671" Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.155035 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2nslt" event={"ID":"a668f1b9-b7e6-4c39-9fc2-9beb5e5e5bce","Type":"ContainerStarted","Data":"5199ded21d628a24cec3e973eb272c52498a6151482f8d162fe94d39dd56cdf8"} Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.160154 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416530-88l5w" event={"ID":"19ade530-d0de-44d1-be83-96136753a35e","Type":"ContainerStarted","Data":"017399694bb75b5df9ba1d00a8d0ea5dd6ffc293c970cfffc5e1bdf4bc7e000a"} Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.168200 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-zfjwx" event={"ID":"8002be53-a1e5-4ef7-93b4-384d9b16a7e2","Type":"ContainerStarted","Data":"de83177f7fed02bf4d72617031a0e5c545cbf607b5f9fbfa27c5e427b3b617be"} Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.191200 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-gkpck" event={"ID":"7abae297-572a-4c3c-8e7d-4e2157d838ad","Type":"ContainerStarted","Data":"4c5eb956d086d19e30d149fe84271eb840886b47534e7fdc3e20f47127644a6d"} Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.195083 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:42:40 crc kubenswrapper[4802]: E1206 03:42:40.195245 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:42:40.695220357 +0000 UTC m=+153.567129509 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.195658 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:40 crc kubenswrapper[4802]: E1206 03:42:40.196703 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:42:40.696694958 +0000 UTC m=+153.568604100 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-blwnp" (UID: "d13a541e-f90d-4222-8486-2d47854ac8e5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.201178 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-whhs8" event={"ID":"ddb4b244-e4e0-4ae4-a217-c6531fc90b6a","Type":"ContainerStarted","Data":"666b9ac45b6f2b89e73e4f2f9502467b902eaa417eb25b15e31d7eb3b1ec0573"} Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.201226 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-whhs8" event={"ID":"ddb4b244-e4e0-4ae4-a217-c6531fc90b6a","Type":"ContainerStarted","Data":"bad9e572d836e1fe9dd7223817dc7ca15c7ab7c99c2e9de6a8ae1f48fd20a358"} Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.201287 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-whhs8" Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.207080 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-ntmms" event={"ID":"f7feaf6a-2ad5-4ef3-88d1-a2a483e5f0d9","Type":"ContainerStarted","Data":"3e774db7c183b990c4932b9a6cd1dedc68f2157c55e0065be47c1b6dbcac4c58"} Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.212343 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-9s5r8" event={"ID":"eb15a464-d0e8-4e49-b018-14353cc9d6cb","Type":"ContainerStarted","Data":"93145508ca970a4ff65cf7bca99dbd9cee7a8f0787d97a55fc26c069e51347c8"} Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.214218 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wbjzw" event={"ID":"18caf043-1bf2-4e67-bb3b-e4edf9fc8822","Type":"ContainerStarted","Data":"14c592c9dfb9acdf01fbaace76434d6972e7842c008c4594986543e3d51e067c"} Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.214341 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wbjzw" event={"ID":"18caf043-1bf2-4e67-bb3b-e4edf9fc8822","Type":"ContainerStarted","Data":"5baf6d89cd0141830b538d9feefc9d9ba7e142b7e6b3a11da5ffd592412b0009"} Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.214530 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wbjzw" Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.218757 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9rvnv" event={"ID":"ff639bee-182b-4dd2-aabc-d396a3bd9ce5","Type":"ContainerStarted","Data":"19cfb8d3a291e389a1d7cd636accd3c89a4d8e4d59bf43e5e33ecfbb495dfd12"} Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.219468 4802 patch_prober.go:28] interesting pod/downloads-7954f5f757-7cg8t container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.219502 4802 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-7cg8t" podUID="0057cff6-6085-4b41-bf4f-e1121bece99a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.220155 4802 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-ddpmx container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.41:8080/healthz\": dial tcp 10.217.0.41:8080: connect: connection refused" start-of-body= Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.220184 4802 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-ddpmx" podUID="ec27be85-6d9d-4d96-b3b8-b78a7a941acb" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.41:8080/healthz\": dial tcp 10.217.0.41:8080: connect: connection refused" Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.233658 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8br44" Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.238692 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4gkmd" Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.256059 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-s5b86" podStartSLOduration=134.256043182 podStartE2EDuration="2m14.256043182s" podCreationTimestamp="2025-12-06 03:40:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:42:40.205320709 +0000 UTC m=+153.077229861" watchObservedRunningTime="2025-12-06 03:42:40.256043182 +0000 UTC m=+153.127952334" Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.297254 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:42:40 crc kubenswrapper[4802]: E1206 03:42:40.297442 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:42:40.797417784 +0000 UTC m=+153.669326936 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.297592 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:40 crc kubenswrapper[4802]: E1206 03:42:40.299677 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:42:40.799667784 +0000 UTC m=+153.671576936 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-blwnp" (UID: "d13a541e-f90d-4222-8486-2d47854ac8e5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.319488 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-whhs8" podStartSLOduration=8.319469097 podStartE2EDuration="8.319469097s" podCreationTimestamp="2025-12-06 03:42:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:42:40.255082537 +0000 UTC m=+153.126991689" watchObservedRunningTime="2025-12-06 03:42:40.319469097 +0000 UTC m=+153.191378249" Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.398646 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:42:40 crc kubenswrapper[4802]: E1206 03:42:40.398814 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:42:40.898780138 +0000 UTC m=+153.770689290 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.398954 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:40 crc kubenswrapper[4802]: E1206 03:42:40.399256 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:42:40.899246051 +0000 UTC m=+153.771155203 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-blwnp" (UID: "d13a541e-f90d-4222-8486-2d47854ac8e5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.423708 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wbjzw" podStartSLOduration=133.423685718 podStartE2EDuration="2m13.423685718s" podCreationTimestamp="2025-12-06 03:40:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:42:40.389980562 +0000 UTC m=+153.261889714" watchObservedRunningTime="2025-12-06 03:42:40.423685718 +0000 UTC m=+153.295594870" Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.503363 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:42:40 crc kubenswrapper[4802]: E1206 03:42:40.503863 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:42:41.003847601 +0000 UTC m=+153.875756753 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.526852 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:40 crc kubenswrapper[4802]: E1206 03:42:40.527375 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:42:41.027362894 +0000 UTC m=+153.899272046 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-blwnp" (UID: "d13a541e-f90d-4222-8486-2d47854ac8e5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.595487 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-ntmms" podStartSLOduration=133.595472324 podStartE2EDuration="2m13.595472324s" podCreationTimestamp="2025-12-06 03:40:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:42:40.503001559 +0000 UTC m=+153.374910711" watchObservedRunningTime="2025-12-06 03:42:40.595472324 +0000 UTC m=+153.467381476" Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.597188 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zzqgx"] Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.598178 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zzqgx" Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.603099 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.629134 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.629293 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wv7n5\" (UniqueName: \"kubernetes.io/projected/03be8283-9ddf-4ecb-a1e7-579c322ca5d6-kube-api-access-wv7n5\") pod \"community-operators-zzqgx\" (UID: \"03be8283-9ddf-4ecb-a1e7-579c322ca5d6\") " pod="openshift-marketplace/community-operators-zzqgx" Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.629317 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03be8283-9ddf-4ecb-a1e7-579c322ca5d6-catalog-content\") pod \"community-operators-zzqgx\" (UID: \"03be8283-9ddf-4ecb-a1e7-579c322ca5d6\") " pod="openshift-marketplace/community-operators-zzqgx" Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.629404 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03be8283-9ddf-4ecb-a1e7-579c322ca5d6-utilities\") pod \"community-operators-zzqgx\" (UID: \"03be8283-9ddf-4ecb-a1e7-579c322ca5d6\") " pod="openshift-marketplace/community-operators-zzqgx" Dec 06 03:42:40 crc kubenswrapper[4802]: E1206 03:42:40.629491 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:42:41.129477668 +0000 UTC m=+154.001386820 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.631535 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zzqgx"] Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.730819 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03be8283-9ddf-4ecb-a1e7-579c322ca5d6-utilities\") pod \"community-operators-zzqgx\" (UID: \"03be8283-9ddf-4ecb-a1e7-579c322ca5d6\") " pod="openshift-marketplace/community-operators-zzqgx" Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.730862 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wv7n5\" (UniqueName: \"kubernetes.io/projected/03be8283-9ddf-4ecb-a1e7-579c322ca5d6-kube-api-access-wv7n5\") pod \"community-operators-zzqgx\" (UID: \"03be8283-9ddf-4ecb-a1e7-579c322ca5d6\") " pod="openshift-marketplace/community-operators-zzqgx" Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.730904 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03be8283-9ddf-4ecb-a1e7-579c322ca5d6-catalog-content\") pod \"community-operators-zzqgx\" (UID: \"03be8283-9ddf-4ecb-a1e7-579c322ca5d6\") " pod="openshift-marketplace/community-operators-zzqgx" Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.730930 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:40 crc kubenswrapper[4802]: E1206 03:42:40.731213 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:42:41.231201361 +0000 UTC m=+154.103110513 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-blwnp" (UID: "d13a541e-f90d-4222-8486-2d47854ac8e5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.731247 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03be8283-9ddf-4ecb-a1e7-579c322ca5d6-utilities\") pod \"community-operators-zzqgx\" (UID: \"03be8283-9ddf-4ecb-a1e7-579c322ca5d6\") " pod="openshift-marketplace/community-operators-zzqgx" Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.731624 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03be8283-9ddf-4ecb-a1e7-579c322ca5d6-catalog-content\") pod \"community-operators-zzqgx\" (UID: \"03be8283-9ddf-4ecb-a1e7-579c322ca5d6\") " pod="openshift-marketplace/community-operators-zzqgx" Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.760633 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wv7n5\" (UniqueName: \"kubernetes.io/projected/03be8283-9ddf-4ecb-a1e7-579c322ca5d6-kube-api-access-wv7n5\") pod \"community-operators-zzqgx\" (UID: \"03be8283-9ddf-4ecb-a1e7-579c322ca5d6\") " pod="openshift-marketplace/community-operators-zzqgx" Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.772725 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-z29bw"] Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.773605 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z29bw" Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.780551 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.795114 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-z29bw"] Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.832912 4802 patch_prober.go:28] interesting pod/router-default-5444994796-rrhqx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 03:42:40 crc kubenswrapper[4802]: [-]has-synced failed: reason withheld Dec 06 03:42:40 crc kubenswrapper[4802]: [+]process-running ok Dec 06 03:42:40 crc kubenswrapper[4802]: healthz check failed Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.834126 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-rrhqx" podUID="0fbd23d0-a6b6-4a4e-b418-07fc38e5f497" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.834039 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:42:40 crc kubenswrapper[4802]: E1206 03:42:40.834090 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:42:41.334076486 +0000 UTC m=+154.205985638 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.834514 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b68d090-3526-4778-ad64-51da5e2191a4-utilities\") pod \"certified-operators-z29bw\" (UID: \"9b68d090-3526-4778-ad64-51da5e2191a4\") " pod="openshift-marketplace/certified-operators-z29bw" Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.834620 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.834716 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5n77\" (UniqueName: \"kubernetes.io/projected/9b68d090-3526-4778-ad64-51da5e2191a4-kube-api-access-t5n77\") pod \"certified-operators-z29bw\" (UID: \"9b68d090-3526-4778-ad64-51da5e2191a4\") " pod="openshift-marketplace/certified-operators-z29bw" Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.834866 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b68d090-3526-4778-ad64-51da5e2191a4-catalog-content\") pod \"certified-operators-z29bw\" (UID: \"9b68d090-3526-4778-ad64-51da5e2191a4\") " pod="openshift-marketplace/certified-operators-z29bw" Dec 06 03:42:40 crc kubenswrapper[4802]: E1206 03:42:40.835078 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:42:41.335061473 +0000 UTC m=+154.206970625 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-blwnp" (UID: "d13a541e-f90d-4222-8486-2d47854ac8e5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.843863 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2nslt" Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.931300 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zzqgx" Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.935472 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:42:40 crc kubenswrapper[4802]: E1206 03:42:40.935650 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:42:41.435627865 +0000 UTC m=+154.307537017 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.935532 4802 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.935896 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b68d090-3526-4778-ad64-51da5e2191a4-catalog-content\") pod \"certified-operators-z29bw\" (UID: \"9b68d090-3526-4778-ad64-51da5e2191a4\") " pod="openshift-marketplace/certified-operators-z29bw" Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.936026 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b68d090-3526-4778-ad64-51da5e2191a4-utilities\") pod \"certified-operators-z29bw\" (UID: \"9b68d090-3526-4778-ad64-51da5e2191a4\") " pod="openshift-marketplace/certified-operators-z29bw" Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.936139 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.936254 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5n77\" (UniqueName: \"kubernetes.io/projected/9b68d090-3526-4778-ad64-51da5e2191a4-kube-api-access-t5n77\") pod \"certified-operators-z29bw\" (UID: \"9b68d090-3526-4778-ad64-51da5e2191a4\") " pod="openshift-marketplace/certified-operators-z29bw" Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.937064 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b68d090-3526-4778-ad64-51da5e2191a4-catalog-content\") pod \"certified-operators-z29bw\" (UID: \"9b68d090-3526-4778-ad64-51da5e2191a4\") " pod="openshift-marketplace/certified-operators-z29bw" Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.937378 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b68d090-3526-4778-ad64-51da5e2191a4-utilities\") pod \"certified-operators-z29bw\" (UID: \"9b68d090-3526-4778-ad64-51da5e2191a4\") " pod="openshift-marketplace/certified-operators-z29bw" Dec 06 03:42:40 crc kubenswrapper[4802]: E1206 03:42:40.937700 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:42:41.437689721 +0000 UTC m=+154.309598873 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-blwnp" (UID: "d13a541e-f90d-4222-8486-2d47854ac8e5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.974490 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5n77\" (UniqueName: \"kubernetes.io/projected/9b68d090-3526-4778-ad64-51da5e2191a4-kube-api-access-t5n77\") pod \"certified-operators-z29bw\" (UID: \"9b68d090-3526-4778-ad64-51da5e2191a4\") " pod="openshift-marketplace/certified-operators-z29bw" Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.984255 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hqh24"] Dec 06 03:42:40 crc kubenswrapper[4802]: I1206 03:42:40.985390 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hqh24" Dec 06 03:42:41 crc kubenswrapper[4802]: I1206 03:42:41.004947 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hqh24"] Dec 06 03:42:41 crc kubenswrapper[4802]: I1206 03:42:41.036802 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:42:41 crc kubenswrapper[4802]: I1206 03:42:41.037014 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sstvn\" (UniqueName: \"kubernetes.io/projected/9ef50a75-3fb0-4a65-ae6a-45f981d5c979-kube-api-access-sstvn\") pod \"community-operators-hqh24\" (UID: \"9ef50a75-3fb0-4a65-ae6a-45f981d5c979\") " pod="openshift-marketplace/community-operators-hqh24" Dec 06 03:42:41 crc kubenswrapper[4802]: I1206 03:42:41.037119 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ef50a75-3fb0-4a65-ae6a-45f981d5c979-utilities\") pod \"community-operators-hqh24\" (UID: \"9ef50a75-3fb0-4a65-ae6a-45f981d5c979\") " pod="openshift-marketplace/community-operators-hqh24" Dec 06 03:42:41 crc kubenswrapper[4802]: I1206 03:42:41.037141 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ef50a75-3fb0-4a65-ae6a-45f981d5c979-catalog-content\") pod \"community-operators-hqh24\" (UID: \"9ef50a75-3fb0-4a65-ae6a-45f981d5c979\") " pod="openshift-marketplace/community-operators-hqh24" Dec 06 03:42:41 crc kubenswrapper[4802]: E1206 03:42:41.037244 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:42:41.537228646 +0000 UTC m=+154.409137798 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:41 crc kubenswrapper[4802]: I1206 03:42:41.100008 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z29bw" Dec 06 03:42:41 crc kubenswrapper[4802]: I1206 03:42:41.138622 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:41 crc kubenswrapper[4802]: I1206 03:42:41.138938 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ef50a75-3fb0-4a65-ae6a-45f981d5c979-utilities\") pod \"community-operators-hqh24\" (UID: \"9ef50a75-3fb0-4a65-ae6a-45f981d5c979\") " pod="openshift-marketplace/community-operators-hqh24" Dec 06 03:42:41 crc kubenswrapper[4802]: I1206 03:42:41.138957 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ef50a75-3fb0-4a65-ae6a-45f981d5c979-catalog-content\") pod \"community-operators-hqh24\" (UID: \"9ef50a75-3fb0-4a65-ae6a-45f981d5c979\") " pod="openshift-marketplace/community-operators-hqh24" Dec 06 03:42:41 crc kubenswrapper[4802]: I1206 03:42:41.138983 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sstvn\" (UniqueName: \"kubernetes.io/projected/9ef50a75-3fb0-4a65-ae6a-45f981d5c979-kube-api-access-sstvn\") pod \"community-operators-hqh24\" (UID: \"9ef50a75-3fb0-4a65-ae6a-45f981d5c979\") " pod="openshift-marketplace/community-operators-hqh24" Dec 06 03:42:41 crc kubenswrapper[4802]: E1206 03:42:41.139540 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:42:41.639523025 +0000 UTC m=+154.511432177 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-blwnp" (UID: "d13a541e-f90d-4222-8486-2d47854ac8e5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:41 crc kubenswrapper[4802]: I1206 03:42:41.139805 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ef50a75-3fb0-4a65-ae6a-45f981d5c979-catalog-content\") pod \"community-operators-hqh24\" (UID: \"9ef50a75-3fb0-4a65-ae6a-45f981d5c979\") " pod="openshift-marketplace/community-operators-hqh24" Dec 06 03:42:41 crc kubenswrapper[4802]: I1206 03:42:41.140025 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ef50a75-3fb0-4a65-ae6a-45f981d5c979-utilities\") pod \"community-operators-hqh24\" (UID: \"9ef50a75-3fb0-4a65-ae6a-45f981d5c979\") " pod="openshift-marketplace/community-operators-hqh24" Dec 06 03:42:41 crc kubenswrapper[4802]: I1206 03:42:41.177707 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sstvn\" (UniqueName: \"kubernetes.io/projected/9ef50a75-3fb0-4a65-ae6a-45f981d5c979-kube-api-access-sstvn\") pod \"community-operators-hqh24\" (UID: \"9ef50a75-3fb0-4a65-ae6a-45f981d5c979\") " pod="openshift-marketplace/community-operators-hqh24" Dec 06 03:42:41 crc kubenswrapper[4802]: I1206 03:42:41.177851 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-gq4cr"] Dec 06 03:42:41 crc kubenswrapper[4802]: I1206 03:42:41.178707 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gq4cr" Dec 06 03:42:41 crc kubenswrapper[4802]: I1206 03:42:41.213885 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gq4cr"] Dec 06 03:42:41 crc kubenswrapper[4802]: I1206 03:42:41.241302 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:42:41 crc kubenswrapper[4802]: I1206 03:42:41.241460 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e6be98f-f5d2-4158-b4b8-eab47ad91564-utilities\") pod \"certified-operators-gq4cr\" (UID: \"4e6be98f-f5d2-4158-b4b8-eab47ad91564\") " pod="openshift-marketplace/certified-operators-gq4cr" Dec 06 03:42:41 crc kubenswrapper[4802]: I1206 03:42:41.241481 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78lrv\" (UniqueName: \"kubernetes.io/projected/4e6be98f-f5d2-4158-b4b8-eab47ad91564-kube-api-access-78lrv\") pod \"certified-operators-gq4cr\" (UID: \"4e6be98f-f5d2-4158-b4b8-eab47ad91564\") " pod="openshift-marketplace/certified-operators-gq4cr" Dec 06 03:42:41 crc kubenswrapper[4802]: I1206 03:42:41.241508 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e6be98f-f5d2-4158-b4b8-eab47ad91564-catalog-content\") pod \"certified-operators-gq4cr\" (UID: \"4e6be98f-f5d2-4158-b4b8-eab47ad91564\") " pod="openshift-marketplace/certified-operators-gq4cr" Dec 06 03:42:41 crc kubenswrapper[4802]: E1206 03:42:41.241674 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:42:41.74166026 +0000 UTC m=+154.613569412 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:41 crc kubenswrapper[4802]: I1206 03:42:41.269086 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-9sxg5" event={"ID":"407d6ced-2671-40de-aa70-153cfda69fba","Type":"ContainerStarted","Data":"f1d8ad33b52ddcd5bcd028b7c0429023d85306787e55836815ae7641894db363"} Dec 06 03:42:41 crc kubenswrapper[4802]: I1206 03:42:41.269121 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-9sxg5" event={"ID":"407d6ced-2671-40de-aa70-153cfda69fba","Type":"ContainerStarted","Data":"410060121d6a03948225fc61de8a2a5a89c2f4b1d8485fbdbd5e05569353a353"} Dec 06 03:42:41 crc kubenswrapper[4802]: I1206 03:42:41.310112 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-ddpmx" Dec 06 03:42:41 crc kubenswrapper[4802]: I1206 03:42:41.310348 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hqh24" Dec 06 03:42:41 crc kubenswrapper[4802]: I1206 03:42:41.342354 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e6be98f-f5d2-4158-b4b8-eab47ad91564-utilities\") pod \"certified-operators-gq4cr\" (UID: \"4e6be98f-f5d2-4158-b4b8-eab47ad91564\") " pod="openshift-marketplace/certified-operators-gq4cr" Dec 06 03:42:41 crc kubenswrapper[4802]: I1206 03:42:41.342391 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78lrv\" (UniqueName: \"kubernetes.io/projected/4e6be98f-f5d2-4158-b4b8-eab47ad91564-kube-api-access-78lrv\") pod \"certified-operators-gq4cr\" (UID: \"4e6be98f-f5d2-4158-b4b8-eab47ad91564\") " pod="openshift-marketplace/certified-operators-gq4cr" Dec 06 03:42:41 crc kubenswrapper[4802]: I1206 03:42:41.342467 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e6be98f-f5d2-4158-b4b8-eab47ad91564-catalog-content\") pod \"certified-operators-gq4cr\" (UID: \"4e6be98f-f5d2-4158-b4b8-eab47ad91564\") " pod="openshift-marketplace/certified-operators-gq4cr" Dec 06 03:42:41 crc kubenswrapper[4802]: I1206 03:42:41.342776 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:41 crc kubenswrapper[4802]: I1206 03:42:41.353151 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e6be98f-f5d2-4158-b4b8-eab47ad91564-utilities\") pod \"certified-operators-gq4cr\" (UID: \"4e6be98f-f5d2-4158-b4b8-eab47ad91564\") " pod="openshift-marketplace/certified-operators-gq4cr" Dec 06 03:42:41 crc kubenswrapper[4802]: I1206 03:42:41.366422 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e6be98f-f5d2-4158-b4b8-eab47ad91564-catalog-content\") pod \"certified-operators-gq4cr\" (UID: \"4e6be98f-f5d2-4158-b4b8-eab47ad91564\") " pod="openshift-marketplace/certified-operators-gq4cr" Dec 06 03:42:41 crc kubenswrapper[4802]: E1206 03:42:41.374309 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:42:41.874288394 +0000 UTC m=+154.746197546 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-blwnp" (UID: "d13a541e-f90d-4222-8486-2d47854ac8e5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:41 crc kubenswrapper[4802]: I1206 03:42:41.385645 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-9sxg5" podStartSLOduration=9.385624989 podStartE2EDuration="9.385624989s" podCreationTimestamp="2025-12-06 03:42:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:42:41.322082891 +0000 UTC m=+154.193992033" watchObservedRunningTime="2025-12-06 03:42:41.385624989 +0000 UTC m=+154.257534141" Dec 06 03:42:41 crc kubenswrapper[4802]: I1206 03:42:41.410969 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78lrv\" (UniqueName: \"kubernetes.io/projected/4e6be98f-f5d2-4158-b4b8-eab47ad91564-kube-api-access-78lrv\") pod \"certified-operators-gq4cr\" (UID: \"4e6be98f-f5d2-4158-b4b8-eab47ad91564\") " pod="openshift-marketplace/certified-operators-gq4cr" Dec 06 03:42:41 crc kubenswrapper[4802]: I1206 03:42:41.444167 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:42:41 crc kubenswrapper[4802]: E1206 03:42:41.444684 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:42:41.944661365 +0000 UTC m=+154.816570517 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:41 crc kubenswrapper[4802]: I1206 03:42:41.516083 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gq4cr" Dec 06 03:42:41 crc kubenswrapper[4802]: I1206 03:42:41.518006 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zzqgx"] Dec 06 03:42:41 crc kubenswrapper[4802]: I1206 03:42:41.545429 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:41 crc kubenswrapper[4802]: E1206 03:42:41.545897 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:42:42.045866675 +0000 UTC m=+154.917775827 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-blwnp" (UID: "d13a541e-f90d-4222-8486-2d47854ac8e5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:41 crc kubenswrapper[4802]: I1206 03:42:41.626300 4802 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-06T03:42:40.93576353Z","Handler":null,"Name":""} Dec 06 03:42:41 crc kubenswrapper[4802]: I1206 03:42:41.650243 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:42:41 crc kubenswrapper[4802]: E1206 03:42:41.650503 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:42:42.150487707 +0000 UTC m=+155.022396859 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:42:41 crc kubenswrapper[4802]: I1206 03:42:41.656967 4802 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 06 03:42:41 crc kubenswrapper[4802]: I1206 03:42:41.657001 4802 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 06 03:42:41 crc kubenswrapper[4802]: I1206 03:42:41.704601 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-z29bw"] Dec 06 03:42:41 crc kubenswrapper[4802]: I1206 03:42:41.751667 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:41 crc kubenswrapper[4802]: I1206 03:42:41.776608 4802 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 06 03:42:41 crc kubenswrapper[4802]: I1206 03:42:41.776900 4802 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:41 crc kubenswrapper[4802]: I1206 03:42:41.820692 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hqh24"] Dec 06 03:42:41 crc kubenswrapper[4802]: I1206 03:42:41.837701 4802 patch_prober.go:28] interesting pod/router-default-5444994796-rrhqx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 03:42:41 crc kubenswrapper[4802]: [-]has-synced failed: reason withheld Dec 06 03:42:41 crc kubenswrapper[4802]: [+]process-running ok Dec 06 03:42:41 crc kubenswrapper[4802]: healthz check failed Dec 06 03:42:41 crc kubenswrapper[4802]: I1206 03:42:41.837773 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-rrhqx" podUID="0fbd23d0-a6b6-4a4e-b418-07fc38e5f497" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 03:42:41 crc kubenswrapper[4802]: I1206 03:42:41.916702 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gq4cr"] Dec 06 03:42:42 crc kubenswrapper[4802]: I1206 03:42:42.275557 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hqh24" event={"ID":"9ef50a75-3fb0-4a65-ae6a-45f981d5c979","Type":"ContainerStarted","Data":"40a9127061e1d069851a74b67d1d9f7910ae3b9c533659f012828ec807360493"} Dec 06 03:42:42 crc kubenswrapper[4802]: I1206 03:42:42.277113 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zzqgx" event={"ID":"03be8283-9ddf-4ecb-a1e7-579c322ca5d6","Type":"ContainerStarted","Data":"c33bb4b3a3a9a725b1c15f3ef15e32e69d08dc350ddee82abc5c3c961451a763"} Dec 06 03:42:42 crc kubenswrapper[4802]: I1206 03:42:42.278389 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z29bw" event={"ID":"9b68d090-3526-4778-ad64-51da5e2191a4","Type":"ContainerStarted","Data":"84396c83a7f184dacab7a87871a7a312d9938d7d51249d9ac392afdc1edf8da5"} Dec 06 03:42:42 crc kubenswrapper[4802]: W1206 03:42:42.322769 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4e6be98f_f5d2_4158_b4b8_eab47ad91564.slice/crio-afa20aa2013c5eda74a9c65ef47f3598d8e6a2395f06742ef522a880728662ba WatchSource:0}: Error finding container afa20aa2013c5eda74a9c65ef47f3598d8e6a2395f06742ef522a880728662ba: Status 404 returned error can't find the container with id afa20aa2013c5eda74a9c65ef47f3598d8e6a2395f06742ef522a880728662ba Dec 06 03:42:42 crc kubenswrapper[4802]: I1206 03:42:42.387072 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-blwnp\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:42 crc kubenswrapper[4802]: I1206 03:42:42.461857 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:42:42 crc kubenswrapper[4802]: I1206 03:42:42.473595 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 06 03:42:42 crc kubenswrapper[4802]: I1206 03:42:42.507855 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:42 crc kubenswrapper[4802]: I1206 03:42:42.772898 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5smmc"] Dec 06 03:42:42 crc kubenswrapper[4802]: I1206 03:42:42.774039 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5smmc" Dec 06 03:42:42 crc kubenswrapper[4802]: I1206 03:42:42.776190 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 06 03:42:42 crc kubenswrapper[4802]: I1206 03:42:42.786256 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-blwnp"] Dec 06 03:42:42 crc kubenswrapper[4802]: I1206 03:42:42.789933 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5smmc"] Dec 06 03:42:42 crc kubenswrapper[4802]: W1206 03:42:42.792385 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd13a541e_f90d_4222_8486_2d47854ac8e5.slice/crio-b86b5a19d5941c6c13bce1346c809fb07700a39b5a1c7efabde6bc0a19a338ab WatchSource:0}: Error finding container b86b5a19d5941c6c13bce1346c809fb07700a39b5a1c7efabde6bc0a19a338ab: Status 404 returned error can't find the container with id b86b5a19d5941c6c13bce1346c809fb07700a39b5a1c7efabde6bc0a19a338ab Dec 06 03:42:42 crc kubenswrapper[4802]: I1206 03:42:42.827220 4802 patch_prober.go:28] interesting pod/router-default-5444994796-rrhqx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 03:42:42 crc kubenswrapper[4802]: [-]has-synced failed: reason withheld Dec 06 03:42:42 crc kubenswrapper[4802]: [+]process-running ok Dec 06 03:42:42 crc kubenswrapper[4802]: healthz check failed Dec 06 03:42:42 crc kubenswrapper[4802]: I1206 03:42:42.827472 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-rrhqx" podUID="0fbd23d0-a6b6-4a4e-b418-07fc38e5f497" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 03:42:42 crc kubenswrapper[4802]: I1206 03:42:42.865721 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wr5tq\" (UniqueName: \"kubernetes.io/projected/f3b86ac0-0a60-46f1-b82c-4bea67124a88-kube-api-access-wr5tq\") pod \"redhat-marketplace-5smmc\" (UID: \"f3b86ac0-0a60-46f1-b82c-4bea67124a88\") " pod="openshift-marketplace/redhat-marketplace-5smmc" Dec 06 03:42:42 crc kubenswrapper[4802]: I1206 03:42:42.865785 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3b86ac0-0a60-46f1-b82c-4bea67124a88-utilities\") pod \"redhat-marketplace-5smmc\" (UID: \"f3b86ac0-0a60-46f1-b82c-4bea67124a88\") " pod="openshift-marketplace/redhat-marketplace-5smmc" Dec 06 03:42:42 crc kubenswrapper[4802]: I1206 03:42:42.865825 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3b86ac0-0a60-46f1-b82c-4bea67124a88-catalog-content\") pod \"redhat-marketplace-5smmc\" (UID: \"f3b86ac0-0a60-46f1-b82c-4bea67124a88\") " pod="openshift-marketplace/redhat-marketplace-5smmc" Dec 06 03:42:42 crc kubenswrapper[4802]: I1206 03:42:42.967097 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3b86ac0-0a60-46f1-b82c-4bea67124a88-catalog-content\") pod \"redhat-marketplace-5smmc\" (UID: \"f3b86ac0-0a60-46f1-b82c-4bea67124a88\") " pod="openshift-marketplace/redhat-marketplace-5smmc" Dec 06 03:42:42 crc kubenswrapper[4802]: I1206 03:42:42.967222 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wr5tq\" (UniqueName: \"kubernetes.io/projected/f3b86ac0-0a60-46f1-b82c-4bea67124a88-kube-api-access-wr5tq\") pod \"redhat-marketplace-5smmc\" (UID: \"f3b86ac0-0a60-46f1-b82c-4bea67124a88\") " pod="openshift-marketplace/redhat-marketplace-5smmc" Dec 06 03:42:42 crc kubenswrapper[4802]: I1206 03:42:42.967254 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3b86ac0-0a60-46f1-b82c-4bea67124a88-utilities\") pod \"redhat-marketplace-5smmc\" (UID: \"f3b86ac0-0a60-46f1-b82c-4bea67124a88\") " pod="openshift-marketplace/redhat-marketplace-5smmc" Dec 06 03:42:42 crc kubenswrapper[4802]: I1206 03:42:42.967842 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3b86ac0-0a60-46f1-b82c-4bea67124a88-utilities\") pod \"redhat-marketplace-5smmc\" (UID: \"f3b86ac0-0a60-46f1-b82c-4bea67124a88\") " pod="openshift-marketplace/redhat-marketplace-5smmc" Dec 06 03:42:42 crc kubenswrapper[4802]: I1206 03:42:42.968144 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3b86ac0-0a60-46f1-b82c-4bea67124a88-catalog-content\") pod \"redhat-marketplace-5smmc\" (UID: \"f3b86ac0-0a60-46f1-b82c-4bea67124a88\") " pod="openshift-marketplace/redhat-marketplace-5smmc" Dec 06 03:42:42 crc kubenswrapper[4802]: I1206 03:42:42.997615 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wr5tq\" (UniqueName: \"kubernetes.io/projected/f3b86ac0-0a60-46f1-b82c-4bea67124a88-kube-api-access-wr5tq\") pod \"redhat-marketplace-5smmc\" (UID: \"f3b86ac0-0a60-46f1-b82c-4bea67124a88\") " pod="openshift-marketplace/redhat-marketplace-5smmc" Dec 06 03:42:43 crc kubenswrapper[4802]: I1206 03:42:43.097290 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5smmc" Dec 06 03:42:43 crc kubenswrapper[4802]: I1206 03:42:43.181461 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-sxwkm"] Dec 06 03:42:43 crc kubenswrapper[4802]: I1206 03:42:43.182706 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sxwkm" Dec 06 03:42:43 crc kubenswrapper[4802]: I1206 03:42:43.207478 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sxwkm"] Dec 06 03:42:43 crc kubenswrapper[4802]: I1206 03:42:43.270611 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a20041d-53aa-4179-b92e-9e7204c800ce-utilities\") pod \"redhat-marketplace-sxwkm\" (UID: \"4a20041d-53aa-4179-b92e-9e7204c800ce\") " pod="openshift-marketplace/redhat-marketplace-sxwkm" Dec 06 03:42:43 crc kubenswrapper[4802]: I1206 03:42:43.270700 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a20041d-53aa-4179-b92e-9e7204c800ce-catalog-content\") pod \"redhat-marketplace-sxwkm\" (UID: \"4a20041d-53aa-4179-b92e-9e7204c800ce\") " pod="openshift-marketplace/redhat-marketplace-sxwkm" Dec 06 03:42:43 crc kubenswrapper[4802]: I1206 03:42:43.270735 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-222z2\" (UniqueName: \"kubernetes.io/projected/4a20041d-53aa-4179-b92e-9e7204c800ce-kube-api-access-222z2\") pod \"redhat-marketplace-sxwkm\" (UID: \"4a20041d-53aa-4179-b92e-9e7204c800ce\") " pod="openshift-marketplace/redhat-marketplace-sxwkm" Dec 06 03:42:43 crc kubenswrapper[4802]: I1206 03:42:43.283686 4802 patch_prober.go:28] interesting pod/machine-config-daemon-zpxxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 03:42:43 crc kubenswrapper[4802]: I1206 03:42:43.283990 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 03:42:43 crc kubenswrapper[4802]: I1206 03:42:43.294024 4802 generic.go:334] "Generic (PLEG): container finished" podID="03be8283-9ddf-4ecb-a1e7-579c322ca5d6" containerID="9907ef4a9dc245ca170757969279ed86ad2bab18ef94fdcbd36bc3561c07b90d" exitCode=0 Dec 06 03:42:43 crc kubenswrapper[4802]: I1206 03:42:43.294083 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zzqgx" event={"ID":"03be8283-9ddf-4ecb-a1e7-579c322ca5d6","Type":"ContainerDied","Data":"9907ef4a9dc245ca170757969279ed86ad2bab18ef94fdcbd36bc3561c07b90d"} Dec 06 03:42:43 crc kubenswrapper[4802]: I1206 03:42:43.295774 4802 generic.go:334] "Generic (PLEG): container finished" podID="4e6be98f-f5d2-4158-b4b8-eab47ad91564" containerID="030dcc3082c01a924aff1c1d291341852b7b09c6ccb9682fff26368e3879f542" exitCode=0 Dec 06 03:42:43 crc kubenswrapper[4802]: I1206 03:42:43.295862 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gq4cr" event={"ID":"4e6be98f-f5d2-4158-b4b8-eab47ad91564","Type":"ContainerDied","Data":"030dcc3082c01a924aff1c1d291341852b7b09c6ccb9682fff26368e3879f542"} Dec 06 03:42:43 crc kubenswrapper[4802]: I1206 03:42:43.295883 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gq4cr" event={"ID":"4e6be98f-f5d2-4158-b4b8-eab47ad91564","Type":"ContainerStarted","Data":"afa20aa2013c5eda74a9c65ef47f3598d8e6a2395f06742ef522a880728662ba"} Dec 06 03:42:43 crc kubenswrapper[4802]: I1206 03:42:43.296765 4802 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 03:42:43 crc kubenswrapper[4802]: I1206 03:42:43.299436 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" event={"ID":"d13a541e-f90d-4222-8486-2d47854ac8e5","Type":"ContainerStarted","Data":"ed30078e684092fe648afb023cfb75ac7a68756ac6e18783bdb323fe369bec1f"} Dec 06 03:42:43 crc kubenswrapper[4802]: I1206 03:42:43.299463 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" event={"ID":"d13a541e-f90d-4222-8486-2d47854ac8e5","Type":"ContainerStarted","Data":"b86b5a19d5941c6c13bce1346c809fb07700a39b5a1c7efabde6bc0a19a338ab"} Dec 06 03:42:43 crc kubenswrapper[4802]: I1206 03:42:43.299710 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:42:43 crc kubenswrapper[4802]: I1206 03:42:43.301250 4802 generic.go:334] "Generic (PLEG): container finished" podID="9b68d090-3526-4778-ad64-51da5e2191a4" containerID="397b292060e8e62c253746a3759e3dfa9f6e6b591b42c77fa0b30757636c00a7" exitCode=0 Dec 06 03:42:43 crc kubenswrapper[4802]: I1206 03:42:43.301277 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z29bw" event={"ID":"9b68d090-3526-4778-ad64-51da5e2191a4","Type":"ContainerDied","Data":"397b292060e8e62c253746a3759e3dfa9f6e6b591b42c77fa0b30757636c00a7"} Dec 06 03:42:43 crc kubenswrapper[4802]: I1206 03:42:43.304087 4802 generic.go:334] "Generic (PLEG): container finished" podID="9ef50a75-3fb0-4a65-ae6a-45f981d5c979" containerID="37ffe03bba19bb6171c5d04d99f9a3638b6e17e277bf6bf20a39cfe3baf7414d" exitCode=0 Dec 06 03:42:43 crc kubenswrapper[4802]: I1206 03:42:43.304116 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hqh24" event={"ID":"9ef50a75-3fb0-4a65-ae6a-45f981d5c979","Type":"ContainerDied","Data":"37ffe03bba19bb6171c5d04d99f9a3638b6e17e277bf6bf20a39cfe3baf7414d"} Dec 06 03:42:43 crc kubenswrapper[4802]: I1206 03:42:43.370329 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5smmc"] Dec 06 03:42:43 crc kubenswrapper[4802]: I1206 03:42:43.372318 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a20041d-53aa-4179-b92e-9e7204c800ce-utilities\") pod \"redhat-marketplace-sxwkm\" (UID: \"4a20041d-53aa-4179-b92e-9e7204c800ce\") " pod="openshift-marketplace/redhat-marketplace-sxwkm" Dec 06 03:42:43 crc kubenswrapper[4802]: I1206 03:42:43.372960 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a20041d-53aa-4179-b92e-9e7204c800ce-utilities\") pod \"redhat-marketplace-sxwkm\" (UID: \"4a20041d-53aa-4179-b92e-9e7204c800ce\") " pod="openshift-marketplace/redhat-marketplace-sxwkm" Dec 06 03:42:43 crc kubenswrapper[4802]: I1206 03:42:43.373313 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a20041d-53aa-4179-b92e-9e7204c800ce-catalog-content\") pod \"redhat-marketplace-sxwkm\" (UID: \"4a20041d-53aa-4179-b92e-9e7204c800ce\") " pod="openshift-marketplace/redhat-marketplace-sxwkm" Dec 06 03:42:43 crc kubenswrapper[4802]: I1206 03:42:43.373347 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-222z2\" (UniqueName: \"kubernetes.io/projected/4a20041d-53aa-4179-b92e-9e7204c800ce-kube-api-access-222z2\") pod \"redhat-marketplace-sxwkm\" (UID: \"4a20041d-53aa-4179-b92e-9e7204c800ce\") " pod="openshift-marketplace/redhat-marketplace-sxwkm" Dec 06 03:42:43 crc kubenswrapper[4802]: I1206 03:42:43.373685 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a20041d-53aa-4179-b92e-9e7204c800ce-catalog-content\") pod \"redhat-marketplace-sxwkm\" (UID: \"4a20041d-53aa-4179-b92e-9e7204c800ce\") " pod="openshift-marketplace/redhat-marketplace-sxwkm" Dec 06 03:42:43 crc kubenswrapper[4802]: I1206 03:42:43.392696 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-222z2\" (UniqueName: \"kubernetes.io/projected/4a20041d-53aa-4179-b92e-9e7204c800ce-kube-api-access-222z2\") pod \"redhat-marketplace-sxwkm\" (UID: \"4a20041d-53aa-4179-b92e-9e7204c800ce\") " pod="openshift-marketplace/redhat-marketplace-sxwkm" Dec 06 03:42:43 crc kubenswrapper[4802]: I1206 03:42:43.425488 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" podStartSLOduration=137.425468498 podStartE2EDuration="2m17.425468498s" podCreationTimestamp="2025-12-06 03:40:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:42:43.403048626 +0000 UTC m=+156.274957778" watchObservedRunningTime="2025-12-06 03:42:43.425468498 +0000 UTC m=+156.297377640" Dec 06 03:42:43 crc kubenswrapper[4802]: I1206 03:42:43.456512 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 06 03:42:43 crc kubenswrapper[4802]: I1206 03:42:43.499363 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sxwkm" Dec 06 03:42:43 crc kubenswrapper[4802]: I1206 03:42:43.594005 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 06 03:42:43 crc kubenswrapper[4802]: I1206 03:42:43.594606 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 03:42:43 crc kubenswrapper[4802]: I1206 03:42:43.596430 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 06 03:42:43 crc kubenswrapper[4802]: I1206 03:42:43.597188 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 06 03:42:43 crc kubenswrapper[4802]: I1206 03:42:43.605063 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 06 03:42:43 crc kubenswrapper[4802]: I1206 03:42:43.678390 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1ef5f0e2-fd47-4f85-aa9d-3a95fc157eed-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"1ef5f0e2-fd47-4f85-aa9d-3a95fc157eed\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 03:42:43 crc kubenswrapper[4802]: I1206 03:42:43.678824 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1ef5f0e2-fd47-4f85-aa9d-3a95fc157eed-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"1ef5f0e2-fd47-4f85-aa9d-3a95fc157eed\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 03:42:43 crc kubenswrapper[4802]: I1206 03:42:43.706074 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sxwkm"] Dec 06 03:42:43 crc kubenswrapper[4802]: W1206 03:42:43.711675 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4a20041d_53aa_4179_b92e_9e7204c800ce.slice/crio-1fe6950a93a2dd7bc463ee39b919bfdfd9b0431781316a96fe69e2867eb54bfa WatchSource:0}: Error finding container 1fe6950a93a2dd7bc463ee39b919bfdfd9b0431781316a96fe69e2867eb54bfa: Status 404 returned error can't find the container with id 1fe6950a93a2dd7bc463ee39b919bfdfd9b0431781316a96fe69e2867eb54bfa Dec 06 03:42:43 crc kubenswrapper[4802]: I1206 03:42:43.781187 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1ef5f0e2-fd47-4f85-aa9d-3a95fc157eed-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"1ef5f0e2-fd47-4f85-aa9d-3a95fc157eed\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 03:42:43 crc kubenswrapper[4802]: I1206 03:42:43.781284 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1ef5f0e2-fd47-4f85-aa9d-3a95fc157eed-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"1ef5f0e2-fd47-4f85-aa9d-3a95fc157eed\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 03:42:43 crc kubenswrapper[4802]: I1206 03:42:43.781513 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1ef5f0e2-fd47-4f85-aa9d-3a95fc157eed-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"1ef5f0e2-fd47-4f85-aa9d-3a95fc157eed\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 03:42:43 crc kubenswrapper[4802]: I1206 03:42:43.792626 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-h7kc8"] Dec 06 03:42:43 crc kubenswrapper[4802]: I1206 03:42:43.802240 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-h7kc8"] Dec 06 03:42:43 crc kubenswrapper[4802]: I1206 03:42:43.802385 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h7kc8" Dec 06 03:42:43 crc kubenswrapper[4802]: I1206 03:42:43.805904 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 06 03:42:43 crc kubenswrapper[4802]: I1206 03:42:43.812614 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1ef5f0e2-fd47-4f85-aa9d-3a95fc157eed-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"1ef5f0e2-fd47-4f85-aa9d-3a95fc157eed\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 03:42:43 crc kubenswrapper[4802]: I1206 03:42:43.828575 4802 patch_prober.go:28] interesting pod/router-default-5444994796-rrhqx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 03:42:43 crc kubenswrapper[4802]: [-]has-synced failed: reason withheld Dec 06 03:42:43 crc kubenswrapper[4802]: [+]process-running ok Dec 06 03:42:43 crc kubenswrapper[4802]: healthz check failed Dec 06 03:42:43 crc kubenswrapper[4802]: I1206 03:42:43.828635 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-rrhqx" podUID="0fbd23d0-a6b6-4a4e-b418-07fc38e5f497" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 03:42:43 crc kubenswrapper[4802]: I1206 03:42:43.947931 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-7ckfj" Dec 06 03:42:43 crc kubenswrapper[4802]: I1206 03:42:43.948157 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-7ckfj" Dec 06 03:42:43 crc kubenswrapper[4802]: I1206 03:42:43.948056 4802 patch_prober.go:28] interesting pod/console-f9d7485db-7ckfj container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.5:8443/health\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Dec 06 03:42:43 crc kubenswrapper[4802]: I1206 03:42:43.949059 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-7ckfj" podUID="2362780e-93d1-4f97-bda0-138ed5180bb0" containerName="console" probeResult="failure" output="Get \"https://10.217.0.5:8443/health\": dial tcp 10.217.0.5:8443: connect: connection refused" Dec 06 03:42:43 crc kubenswrapper[4802]: I1206 03:42:43.970726 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 03:42:43 crc kubenswrapper[4802]: I1206 03:42:43.983662 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqtns\" (UniqueName: \"kubernetes.io/projected/edb8213d-6a0c-4051-a230-4ec109d6e578-kube-api-access-gqtns\") pod \"redhat-operators-h7kc8\" (UID: \"edb8213d-6a0c-4051-a230-4ec109d6e578\") " pod="openshift-marketplace/redhat-operators-h7kc8" Dec 06 03:42:43 crc kubenswrapper[4802]: I1206 03:42:43.983819 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/edb8213d-6a0c-4051-a230-4ec109d6e578-utilities\") pod \"redhat-operators-h7kc8\" (UID: \"edb8213d-6a0c-4051-a230-4ec109d6e578\") " pod="openshift-marketplace/redhat-operators-h7kc8" Dec 06 03:42:43 crc kubenswrapper[4802]: I1206 03:42:43.983913 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/edb8213d-6a0c-4051-a230-4ec109d6e578-catalog-content\") pod \"redhat-operators-h7kc8\" (UID: \"edb8213d-6a0c-4051-a230-4ec109d6e578\") " pod="openshift-marketplace/redhat-operators-h7kc8" Dec 06 03:42:44 crc kubenswrapper[4802]: I1206 03:42:44.085623 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/edb8213d-6a0c-4051-a230-4ec109d6e578-utilities\") pod \"redhat-operators-h7kc8\" (UID: \"edb8213d-6a0c-4051-a230-4ec109d6e578\") " pod="openshift-marketplace/redhat-operators-h7kc8" Dec 06 03:42:44 crc kubenswrapper[4802]: I1206 03:42:44.086370 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/edb8213d-6a0c-4051-a230-4ec109d6e578-utilities\") pod \"redhat-operators-h7kc8\" (UID: \"edb8213d-6a0c-4051-a230-4ec109d6e578\") " pod="openshift-marketplace/redhat-operators-h7kc8" Dec 06 03:42:44 crc kubenswrapper[4802]: I1206 03:42:44.088141 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/edb8213d-6a0c-4051-a230-4ec109d6e578-catalog-content\") pod \"redhat-operators-h7kc8\" (UID: \"edb8213d-6a0c-4051-a230-4ec109d6e578\") " pod="openshift-marketplace/redhat-operators-h7kc8" Dec 06 03:42:44 crc kubenswrapper[4802]: I1206 03:42:44.088709 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqtns\" (UniqueName: \"kubernetes.io/projected/edb8213d-6a0c-4051-a230-4ec109d6e578-kube-api-access-gqtns\") pod \"redhat-operators-h7kc8\" (UID: \"edb8213d-6a0c-4051-a230-4ec109d6e578\") " pod="openshift-marketplace/redhat-operators-h7kc8" Dec 06 03:42:44 crc kubenswrapper[4802]: I1206 03:42:44.088620 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/edb8213d-6a0c-4051-a230-4ec109d6e578-catalog-content\") pod \"redhat-operators-h7kc8\" (UID: \"edb8213d-6a0c-4051-a230-4ec109d6e578\") " pod="openshift-marketplace/redhat-operators-h7kc8" Dec 06 03:42:44 crc kubenswrapper[4802]: I1206 03:42:44.126673 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqtns\" (UniqueName: \"kubernetes.io/projected/edb8213d-6a0c-4051-a230-4ec109d6e578-kube-api-access-gqtns\") pod \"redhat-operators-h7kc8\" (UID: \"edb8213d-6a0c-4051-a230-4ec109d6e578\") " pod="openshift-marketplace/redhat-operators-h7kc8" Dec 06 03:42:44 crc kubenswrapper[4802]: I1206 03:42:44.138658 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h7kc8" Dec 06 03:42:44 crc kubenswrapper[4802]: I1206 03:42:44.165380 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-s5b86" Dec 06 03:42:44 crc kubenswrapper[4802]: I1206 03:42:44.192642 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-zm9dd"] Dec 06 03:42:44 crc kubenswrapper[4802]: I1206 03:42:44.194381 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zm9dd" Dec 06 03:42:44 crc kubenswrapper[4802]: I1206 03:42:44.203393 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zm9dd"] Dec 06 03:42:44 crc kubenswrapper[4802]: I1206 03:42:44.211681 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 06 03:42:44 crc kubenswrapper[4802]: I1206 03:42:44.326011 4802 generic.go:334] "Generic (PLEG): container finished" podID="4a20041d-53aa-4179-b92e-9e7204c800ce" containerID="7f372faed4e069189425f6e681735ed7e7a60e44247259a3cb2291b9cd69a2eb" exitCode=0 Dec 06 03:42:44 crc kubenswrapper[4802]: I1206 03:42:44.326206 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sxwkm" event={"ID":"4a20041d-53aa-4179-b92e-9e7204c800ce","Type":"ContainerDied","Data":"7f372faed4e069189425f6e681735ed7e7a60e44247259a3cb2291b9cd69a2eb"} Dec 06 03:42:44 crc kubenswrapper[4802]: I1206 03:42:44.326430 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sxwkm" event={"ID":"4a20041d-53aa-4179-b92e-9e7204c800ce","Type":"ContainerStarted","Data":"1fe6950a93a2dd7bc463ee39b919bfdfd9b0431781316a96fe69e2867eb54bfa"} Dec 06 03:42:44 crc kubenswrapper[4802]: I1206 03:42:44.331348 4802 generic.go:334] "Generic (PLEG): container finished" podID="19ade530-d0de-44d1-be83-96136753a35e" containerID="017399694bb75b5df9ba1d00a8d0ea5dd6ffc293c970cfffc5e1bdf4bc7e000a" exitCode=0 Dec 06 03:42:44 crc kubenswrapper[4802]: I1206 03:42:44.331434 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416530-88l5w" event={"ID":"19ade530-d0de-44d1-be83-96136753a35e","Type":"ContainerDied","Data":"017399694bb75b5df9ba1d00a8d0ea5dd6ffc293c970cfffc5e1bdf4bc7e000a"} Dec 06 03:42:44 crc kubenswrapper[4802]: I1206 03:42:44.336712 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"1ef5f0e2-fd47-4f85-aa9d-3a95fc157eed","Type":"ContainerStarted","Data":"dbc8aeb8cb12393643636b055a326d9c6ae8955fa97090fc5ffd2f81e065c922"} Dec 06 03:42:44 crc kubenswrapper[4802]: I1206 03:42:44.343220 4802 generic.go:334] "Generic (PLEG): container finished" podID="f3b86ac0-0a60-46f1-b82c-4bea67124a88" containerID="83129b5f5b11dadd2eb15c5a06f651ece3ab00cdf66ff0c69f51a4cde595e814" exitCode=0 Dec 06 03:42:44 crc kubenswrapper[4802]: I1206 03:42:44.345222 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5smmc" event={"ID":"f3b86ac0-0a60-46f1-b82c-4bea67124a88","Type":"ContainerDied","Data":"83129b5f5b11dadd2eb15c5a06f651ece3ab00cdf66ff0c69f51a4cde595e814"} Dec 06 03:42:44 crc kubenswrapper[4802]: I1206 03:42:44.345254 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5smmc" event={"ID":"f3b86ac0-0a60-46f1-b82c-4bea67124a88","Type":"ContainerStarted","Data":"7e224574fc248a6e6a948eabbf596ccfdea512b46261b84c98713eb75574aaee"} Dec 06 03:42:44 crc kubenswrapper[4802]: I1206 03:42:44.395346 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6eff099-6873-49c2-b0e9-e937700e1ff9-utilities\") pod \"redhat-operators-zm9dd\" (UID: \"d6eff099-6873-49c2-b0e9-e937700e1ff9\") " pod="openshift-marketplace/redhat-operators-zm9dd" Dec 06 03:42:44 crc kubenswrapper[4802]: I1206 03:42:44.395433 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7qdd\" (UniqueName: \"kubernetes.io/projected/d6eff099-6873-49c2-b0e9-e937700e1ff9-kube-api-access-b7qdd\") pod \"redhat-operators-zm9dd\" (UID: \"d6eff099-6873-49c2-b0e9-e937700e1ff9\") " pod="openshift-marketplace/redhat-operators-zm9dd" Dec 06 03:42:44 crc kubenswrapper[4802]: I1206 03:42:44.395492 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6eff099-6873-49c2-b0e9-e937700e1ff9-catalog-content\") pod \"redhat-operators-zm9dd\" (UID: \"d6eff099-6873-49c2-b0e9-e937700e1ff9\") " pod="openshift-marketplace/redhat-operators-zm9dd" Dec 06 03:42:44 crc kubenswrapper[4802]: I1206 03:42:44.428519 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-h7kc8"] Dec 06 03:42:44 crc kubenswrapper[4802]: W1206 03:42:44.433567 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podedb8213d_6a0c_4051_a230_4ec109d6e578.slice/crio-6ebe08776ad723f0640f9ddbf39e0df2605443d52062262e5138be9988831d18 WatchSource:0}: Error finding container 6ebe08776ad723f0640f9ddbf39e0df2605443d52062262e5138be9988831d18: Status 404 returned error can't find the container with id 6ebe08776ad723f0640f9ddbf39e0df2605443d52062262e5138be9988831d18 Dec 06 03:42:44 crc kubenswrapper[4802]: I1206 03:42:44.496316 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6eff099-6873-49c2-b0e9-e937700e1ff9-catalog-content\") pod \"redhat-operators-zm9dd\" (UID: \"d6eff099-6873-49c2-b0e9-e937700e1ff9\") " pod="openshift-marketplace/redhat-operators-zm9dd" Dec 06 03:42:44 crc kubenswrapper[4802]: I1206 03:42:44.496432 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6eff099-6873-49c2-b0e9-e937700e1ff9-utilities\") pod \"redhat-operators-zm9dd\" (UID: \"d6eff099-6873-49c2-b0e9-e937700e1ff9\") " pod="openshift-marketplace/redhat-operators-zm9dd" Dec 06 03:42:44 crc kubenswrapper[4802]: I1206 03:42:44.496494 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7qdd\" (UniqueName: \"kubernetes.io/projected/d6eff099-6873-49c2-b0e9-e937700e1ff9-kube-api-access-b7qdd\") pod \"redhat-operators-zm9dd\" (UID: \"d6eff099-6873-49c2-b0e9-e937700e1ff9\") " pod="openshift-marketplace/redhat-operators-zm9dd" Dec 06 03:42:44 crc kubenswrapper[4802]: I1206 03:42:44.497762 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6eff099-6873-49c2-b0e9-e937700e1ff9-utilities\") pod \"redhat-operators-zm9dd\" (UID: \"d6eff099-6873-49c2-b0e9-e937700e1ff9\") " pod="openshift-marketplace/redhat-operators-zm9dd" Dec 06 03:42:44 crc kubenswrapper[4802]: I1206 03:42:44.498345 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6eff099-6873-49c2-b0e9-e937700e1ff9-catalog-content\") pod \"redhat-operators-zm9dd\" (UID: \"d6eff099-6873-49c2-b0e9-e937700e1ff9\") " pod="openshift-marketplace/redhat-operators-zm9dd" Dec 06 03:42:44 crc kubenswrapper[4802]: I1206 03:42:44.521859 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7qdd\" (UniqueName: \"kubernetes.io/projected/d6eff099-6873-49c2-b0e9-e937700e1ff9-kube-api-access-b7qdd\") pod \"redhat-operators-zm9dd\" (UID: \"d6eff099-6873-49c2-b0e9-e937700e1ff9\") " pod="openshift-marketplace/redhat-operators-zm9dd" Dec 06 03:42:44 crc kubenswrapper[4802]: I1206 03:42:44.549987 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zm9dd" Dec 06 03:42:44 crc kubenswrapper[4802]: I1206 03:42:44.733931 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-8xz4m" Dec 06 03:42:44 crc kubenswrapper[4802]: I1206 03:42:44.734392 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-8xz4m" Dec 06 03:42:44 crc kubenswrapper[4802]: I1206 03:42:44.741607 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-8xz4m" Dec 06 03:42:44 crc kubenswrapper[4802]: I1206 03:42:44.745132 4802 patch_prober.go:28] interesting pod/downloads-7954f5f757-7cg8t container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Dec 06 03:42:44 crc kubenswrapper[4802]: I1206 03:42:44.745171 4802 patch_prober.go:28] interesting pod/downloads-7954f5f757-7cg8t container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Dec 06 03:42:44 crc kubenswrapper[4802]: I1206 03:42:44.745204 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-7cg8t" podUID="0057cff6-6085-4b41-bf4f-e1121bece99a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Dec 06 03:42:44 crc kubenswrapper[4802]: I1206 03:42:44.745216 4802 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-7cg8t" podUID="0057cff6-6085-4b41-bf4f-e1121bece99a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Dec 06 03:42:44 crc kubenswrapper[4802]: I1206 03:42:44.824362 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-rrhqx" Dec 06 03:42:44 crc kubenswrapper[4802]: I1206 03:42:44.827131 4802 patch_prober.go:28] interesting pod/router-default-5444994796-rrhqx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 03:42:44 crc kubenswrapper[4802]: [-]has-synced failed: reason withheld Dec 06 03:42:44 crc kubenswrapper[4802]: [+]process-running ok Dec 06 03:42:44 crc kubenswrapper[4802]: healthz check failed Dec 06 03:42:44 crc kubenswrapper[4802]: I1206 03:42:44.827182 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-rrhqx" podUID="0fbd23d0-a6b6-4a4e-b418-07fc38e5f497" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 03:42:44 crc kubenswrapper[4802]: I1206 03:42:44.847031 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zm9dd"] Dec 06 03:42:44 crc kubenswrapper[4802]: W1206 03:42:44.855693 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd6eff099_6873_49c2_b0e9_e937700e1ff9.slice/crio-b927bc37e72ad725e18780474a2300daa9a0e1eff333761d21a0868a8b5c8154 WatchSource:0}: Error finding container b927bc37e72ad725e18780474a2300daa9a0e1eff333761d21a0868a8b5c8154: Status 404 returned error can't find the container with id b927bc37e72ad725e18780474a2300daa9a0e1eff333761d21a0868a8b5c8154 Dec 06 03:42:45 crc kubenswrapper[4802]: I1206 03:42:45.361981 4802 generic.go:334] "Generic (PLEG): container finished" podID="1ef5f0e2-fd47-4f85-aa9d-3a95fc157eed" containerID="dca1321ca71f3c6b60e8a8590767e6f25f4ea1b8bfd101c020f688b0b43beb99" exitCode=0 Dec 06 03:42:45 crc kubenswrapper[4802]: I1206 03:42:45.362336 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"1ef5f0e2-fd47-4f85-aa9d-3a95fc157eed","Type":"ContainerDied","Data":"dca1321ca71f3c6b60e8a8590767e6f25f4ea1b8bfd101c020f688b0b43beb99"} Dec 06 03:42:45 crc kubenswrapper[4802]: I1206 03:42:45.368221 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zm9dd" event={"ID":"d6eff099-6873-49c2-b0e9-e937700e1ff9","Type":"ContainerStarted","Data":"0582177ab2c3f046a646db5bbaffb24446c1f48184610e616537757ac4ee387e"} Dec 06 03:42:45 crc kubenswrapper[4802]: I1206 03:42:45.368285 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zm9dd" event={"ID":"d6eff099-6873-49c2-b0e9-e937700e1ff9","Type":"ContainerStarted","Data":"b927bc37e72ad725e18780474a2300daa9a0e1eff333761d21a0868a8b5c8154"} Dec 06 03:42:45 crc kubenswrapper[4802]: I1206 03:42:45.372784 4802 generic.go:334] "Generic (PLEG): container finished" podID="edb8213d-6a0c-4051-a230-4ec109d6e578" containerID="26b46463a973134ae8e6eed18bdbd6ce95260bb992ec56f67f8ab6877e1bf3f0" exitCode=0 Dec 06 03:42:45 crc kubenswrapper[4802]: I1206 03:42:45.373798 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h7kc8" event={"ID":"edb8213d-6a0c-4051-a230-4ec109d6e578","Type":"ContainerDied","Data":"26b46463a973134ae8e6eed18bdbd6ce95260bb992ec56f67f8ab6877e1bf3f0"} Dec 06 03:42:45 crc kubenswrapper[4802]: I1206 03:42:45.373828 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h7kc8" event={"ID":"edb8213d-6a0c-4051-a230-4ec109d6e578","Type":"ContainerStarted","Data":"6ebe08776ad723f0640f9ddbf39e0df2605443d52062262e5138be9988831d18"} Dec 06 03:42:45 crc kubenswrapper[4802]: I1206 03:42:45.381912 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-8xz4m" Dec 06 03:42:45 crc kubenswrapper[4802]: I1206 03:42:45.710574 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416530-88l5w" Dec 06 03:42:45 crc kubenswrapper[4802]: I1206 03:42:45.767846 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 06 03:42:45 crc kubenswrapper[4802]: E1206 03:42:45.768070 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19ade530-d0de-44d1-be83-96136753a35e" containerName="collect-profiles" Dec 06 03:42:45 crc kubenswrapper[4802]: I1206 03:42:45.768080 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="19ade530-d0de-44d1-be83-96136753a35e" containerName="collect-profiles" Dec 06 03:42:45 crc kubenswrapper[4802]: I1206 03:42:45.768197 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="19ade530-d0de-44d1-be83-96136753a35e" containerName="collect-profiles" Dec 06 03:42:45 crc kubenswrapper[4802]: I1206 03:42:45.771184 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 03:42:45 crc kubenswrapper[4802]: I1206 03:42:45.775574 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 06 03:42:45 crc kubenswrapper[4802]: I1206 03:42:45.775653 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 06 03:42:45 crc kubenswrapper[4802]: I1206 03:42:45.779705 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 06 03:42:45 crc kubenswrapper[4802]: I1206 03:42:45.825239 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pbc8f\" (UniqueName: \"kubernetes.io/projected/19ade530-d0de-44d1-be83-96136753a35e-kube-api-access-pbc8f\") pod \"19ade530-d0de-44d1-be83-96136753a35e\" (UID: \"19ade530-d0de-44d1-be83-96136753a35e\") " Dec 06 03:42:45 crc kubenswrapper[4802]: I1206 03:42:45.825601 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/19ade530-d0de-44d1-be83-96136753a35e-config-volume\") pod \"19ade530-d0de-44d1-be83-96136753a35e\" (UID: \"19ade530-d0de-44d1-be83-96136753a35e\") " Dec 06 03:42:45 crc kubenswrapper[4802]: I1206 03:42:45.825645 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/19ade530-d0de-44d1-be83-96136753a35e-secret-volume\") pod \"19ade530-d0de-44d1-be83-96136753a35e\" (UID: \"19ade530-d0de-44d1-be83-96136753a35e\") " Dec 06 03:42:45 crc kubenswrapper[4802]: I1206 03:42:45.825976 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d0dacc47-2c44-4fb5-8f08-5015ab789df0-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"d0dacc47-2c44-4fb5-8f08-5015ab789df0\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 03:42:45 crc kubenswrapper[4802]: I1206 03:42:45.826184 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d0dacc47-2c44-4fb5-8f08-5015ab789df0-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"d0dacc47-2c44-4fb5-8f08-5015ab789df0\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 03:42:45 crc kubenswrapper[4802]: I1206 03:42:45.826170 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19ade530-d0de-44d1-be83-96136753a35e-config-volume" (OuterVolumeSpecName: "config-volume") pod "19ade530-d0de-44d1-be83-96136753a35e" (UID: "19ade530-d0de-44d1-be83-96136753a35e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:42:45 crc kubenswrapper[4802]: I1206 03:42:45.829110 4802 patch_prober.go:28] interesting pod/router-default-5444994796-rrhqx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 03:42:45 crc kubenswrapper[4802]: [-]has-synced failed: reason withheld Dec 06 03:42:45 crc kubenswrapper[4802]: [+]process-running ok Dec 06 03:42:45 crc kubenswrapper[4802]: healthz check failed Dec 06 03:42:45 crc kubenswrapper[4802]: I1206 03:42:45.829149 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-rrhqx" podUID="0fbd23d0-a6b6-4a4e-b418-07fc38e5f497" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 03:42:45 crc kubenswrapper[4802]: I1206 03:42:45.831723 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19ade530-d0de-44d1-be83-96136753a35e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "19ade530-d0de-44d1-be83-96136753a35e" (UID: "19ade530-d0de-44d1-be83-96136753a35e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:42:45 crc kubenswrapper[4802]: I1206 03:42:45.865296 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19ade530-d0de-44d1-be83-96136753a35e-kube-api-access-pbc8f" (OuterVolumeSpecName: "kube-api-access-pbc8f") pod "19ade530-d0de-44d1-be83-96136753a35e" (UID: "19ade530-d0de-44d1-be83-96136753a35e"). InnerVolumeSpecName "kube-api-access-pbc8f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:42:45 crc kubenswrapper[4802]: I1206 03:42:45.927134 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d0dacc47-2c44-4fb5-8f08-5015ab789df0-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"d0dacc47-2c44-4fb5-8f08-5015ab789df0\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 03:42:45 crc kubenswrapper[4802]: I1206 03:42:45.927248 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d0dacc47-2c44-4fb5-8f08-5015ab789df0-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"d0dacc47-2c44-4fb5-8f08-5015ab789df0\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 03:42:45 crc kubenswrapper[4802]: I1206 03:42:45.927372 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pbc8f\" (UniqueName: \"kubernetes.io/projected/19ade530-d0de-44d1-be83-96136753a35e-kube-api-access-pbc8f\") on node \"crc\" DevicePath \"\"" Dec 06 03:42:45 crc kubenswrapper[4802]: I1206 03:42:45.927399 4802 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/19ade530-d0de-44d1-be83-96136753a35e-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 03:42:45 crc kubenswrapper[4802]: I1206 03:42:45.927448 4802 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/19ade530-d0de-44d1-be83-96136753a35e-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 03:42:45 crc kubenswrapper[4802]: I1206 03:42:45.927537 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d0dacc47-2c44-4fb5-8f08-5015ab789df0-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"d0dacc47-2c44-4fb5-8f08-5015ab789df0\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 03:42:45 crc kubenswrapper[4802]: I1206 03:42:45.944227 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d0dacc47-2c44-4fb5-8f08-5015ab789df0-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"d0dacc47-2c44-4fb5-8f08-5015ab789df0\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 03:42:46 crc kubenswrapper[4802]: I1206 03:42:46.096030 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 03:42:46 crc kubenswrapper[4802]: I1206 03:42:46.413723 4802 generic.go:334] "Generic (PLEG): container finished" podID="d6eff099-6873-49c2-b0e9-e937700e1ff9" containerID="0582177ab2c3f046a646db5bbaffb24446c1f48184610e616537757ac4ee387e" exitCode=0 Dec 06 03:42:46 crc kubenswrapper[4802]: I1206 03:42:46.413825 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zm9dd" event={"ID":"d6eff099-6873-49c2-b0e9-e937700e1ff9","Type":"ContainerDied","Data":"0582177ab2c3f046a646db5bbaffb24446c1f48184610e616537757ac4ee387e"} Dec 06 03:42:46 crc kubenswrapper[4802]: I1206 03:42:46.420031 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416530-88l5w" event={"ID":"19ade530-d0de-44d1-be83-96136753a35e","Type":"ContainerDied","Data":"6f0a1e014d7145ba9b9f1a0e3e33e00d6baa3616ed4444a2de7bf58db8cda08e"} Dec 06 03:42:46 crc kubenswrapper[4802]: I1206 03:42:46.420074 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6f0a1e014d7145ba9b9f1a0e3e33e00d6baa3616ed4444a2de7bf58db8cda08e" Dec 06 03:42:46 crc kubenswrapper[4802]: I1206 03:42:46.420096 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 06 03:42:46 crc kubenswrapper[4802]: I1206 03:42:46.420152 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416530-88l5w" Dec 06 03:42:46 crc kubenswrapper[4802]: W1206 03:42:46.451084 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podd0dacc47_2c44_4fb5_8f08_5015ab789df0.slice/crio-dd3f60a800f995901b2ed66a25ae438d8fe7015eee21a84102d260ddc7168b33 WatchSource:0}: Error finding container dd3f60a800f995901b2ed66a25ae438d8fe7015eee21a84102d260ddc7168b33: Status 404 returned error can't find the container with id dd3f60a800f995901b2ed66a25ae438d8fe7015eee21a84102d260ddc7168b33 Dec 06 03:42:46 crc kubenswrapper[4802]: I1206 03:42:46.717667 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 03:42:46 crc kubenswrapper[4802]: I1206 03:42:46.753984 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1ef5f0e2-fd47-4f85-aa9d-3a95fc157eed-kube-api-access\") pod \"1ef5f0e2-fd47-4f85-aa9d-3a95fc157eed\" (UID: \"1ef5f0e2-fd47-4f85-aa9d-3a95fc157eed\") " Dec 06 03:42:46 crc kubenswrapper[4802]: I1206 03:42:46.754034 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1ef5f0e2-fd47-4f85-aa9d-3a95fc157eed-kubelet-dir\") pod \"1ef5f0e2-fd47-4f85-aa9d-3a95fc157eed\" (UID: \"1ef5f0e2-fd47-4f85-aa9d-3a95fc157eed\") " Dec 06 03:42:46 crc kubenswrapper[4802]: I1206 03:42:46.754266 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1ef5f0e2-fd47-4f85-aa9d-3a95fc157eed-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "1ef5f0e2-fd47-4f85-aa9d-3a95fc157eed" (UID: "1ef5f0e2-fd47-4f85-aa9d-3a95fc157eed"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:42:46 crc kubenswrapper[4802]: I1206 03:42:46.754534 4802 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1ef5f0e2-fd47-4f85-aa9d-3a95fc157eed-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 06 03:42:46 crc kubenswrapper[4802]: I1206 03:42:46.777302 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ef5f0e2-fd47-4f85-aa9d-3a95fc157eed-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1ef5f0e2-fd47-4f85-aa9d-3a95fc157eed" (UID: "1ef5f0e2-fd47-4f85-aa9d-3a95fc157eed"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:42:46 crc kubenswrapper[4802]: I1206 03:42:46.830195 4802 patch_prober.go:28] interesting pod/router-default-5444994796-rrhqx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 03:42:46 crc kubenswrapper[4802]: [-]has-synced failed: reason withheld Dec 06 03:42:46 crc kubenswrapper[4802]: [+]process-running ok Dec 06 03:42:46 crc kubenswrapper[4802]: healthz check failed Dec 06 03:42:46 crc kubenswrapper[4802]: I1206 03:42:46.830260 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-rrhqx" podUID="0fbd23d0-a6b6-4a4e-b418-07fc38e5f497" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 03:42:46 crc kubenswrapper[4802]: I1206 03:42:46.856204 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1ef5f0e2-fd47-4f85-aa9d-3a95fc157eed-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 03:42:47 crc kubenswrapper[4802]: I1206 03:42:47.626375 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"d0dacc47-2c44-4fb5-8f08-5015ab789df0","Type":"ContainerStarted","Data":"dd3f60a800f995901b2ed66a25ae438d8fe7015eee21a84102d260ddc7168b33"} Dec 06 03:42:47 crc kubenswrapper[4802]: I1206 03:42:47.638973 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"1ef5f0e2-fd47-4f85-aa9d-3a95fc157eed","Type":"ContainerDied","Data":"dbc8aeb8cb12393643636b055a326d9c6ae8955fa97090fc5ffd2f81e065c922"} Dec 06 03:42:47 crc kubenswrapper[4802]: I1206 03:42:47.639023 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dbc8aeb8cb12393643636b055a326d9c6ae8955fa97090fc5ffd2f81e065c922" Dec 06 03:42:47 crc kubenswrapper[4802]: I1206 03:42:47.639562 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 03:42:47 crc kubenswrapper[4802]: I1206 03:42:47.826261 4802 patch_prober.go:28] interesting pod/router-default-5444994796-rrhqx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 03:42:47 crc kubenswrapper[4802]: [-]has-synced failed: reason withheld Dec 06 03:42:47 crc kubenswrapper[4802]: [+]process-running ok Dec 06 03:42:47 crc kubenswrapper[4802]: healthz check failed Dec 06 03:42:47 crc kubenswrapper[4802]: I1206 03:42:47.826316 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-rrhqx" podUID="0fbd23d0-a6b6-4a4e-b418-07fc38e5f497" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 03:42:48 crc kubenswrapper[4802]: I1206 03:42:48.715961 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"d0dacc47-2c44-4fb5-8f08-5015ab789df0","Type":"ContainerStarted","Data":"30c1420c37bc20523de42042efcf93a5f716e3861ede47c9ab86c9b12b2c0471"} Dec 06 03:42:48 crc kubenswrapper[4802]: I1206 03:42:48.743711 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=3.743695482 podStartE2EDuration="3.743695482s" podCreationTimestamp="2025-12-06 03:42:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:42:48.742069138 +0000 UTC m=+161.613978300" watchObservedRunningTime="2025-12-06 03:42:48.743695482 +0000 UTC m=+161.615604634" Dec 06 03:42:48 crc kubenswrapper[4802]: I1206 03:42:48.828163 4802 patch_prober.go:28] interesting pod/router-default-5444994796-rrhqx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 03:42:48 crc kubenswrapper[4802]: [-]has-synced failed: reason withheld Dec 06 03:42:48 crc kubenswrapper[4802]: [+]process-running ok Dec 06 03:42:48 crc kubenswrapper[4802]: healthz check failed Dec 06 03:42:48 crc kubenswrapper[4802]: I1206 03:42:48.828221 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-rrhqx" podUID="0fbd23d0-a6b6-4a4e-b418-07fc38e5f497" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 03:42:48 crc kubenswrapper[4802]: I1206 03:42:48.999439 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8881599e-71e5-4578-b3f9-21f4fa2f57aa-metrics-certs\") pod \"network-metrics-daemon-5ltx2\" (UID: \"8881599e-71e5-4578-b3f9-21f4fa2f57aa\") " pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:42:49 crc kubenswrapper[4802]: I1206 03:42:49.010935 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8881599e-71e5-4578-b3f9-21f4fa2f57aa-metrics-certs\") pod \"network-metrics-daemon-5ltx2\" (UID: \"8881599e-71e5-4578-b3f9-21f4fa2f57aa\") " pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:42:49 crc kubenswrapper[4802]: I1206 03:42:49.085225 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5ltx2" Dec 06 03:42:49 crc kubenswrapper[4802]: I1206 03:42:49.729370 4802 generic.go:334] "Generic (PLEG): container finished" podID="d0dacc47-2c44-4fb5-8f08-5015ab789df0" containerID="30c1420c37bc20523de42042efcf93a5f716e3861ede47c9ab86c9b12b2c0471" exitCode=0 Dec 06 03:42:49 crc kubenswrapper[4802]: I1206 03:42:49.729419 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"d0dacc47-2c44-4fb5-8f08-5015ab789df0","Type":"ContainerDied","Data":"30c1420c37bc20523de42042efcf93a5f716e3861ede47c9ab86c9b12b2c0471"} Dec 06 03:42:49 crc kubenswrapper[4802]: I1206 03:42:49.832822 4802 patch_prober.go:28] interesting pod/router-default-5444994796-rrhqx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 03:42:49 crc kubenswrapper[4802]: [-]has-synced failed: reason withheld Dec 06 03:42:49 crc kubenswrapper[4802]: [+]process-running ok Dec 06 03:42:49 crc kubenswrapper[4802]: healthz check failed Dec 06 03:42:49 crc kubenswrapper[4802]: I1206 03:42:49.832926 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-rrhqx" podUID="0fbd23d0-a6b6-4a4e-b418-07fc38e5f497" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 03:42:50 crc kubenswrapper[4802]: I1206 03:42:50.645144 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-whhs8" Dec 06 03:42:50 crc kubenswrapper[4802]: I1206 03:42:50.827024 4802 patch_prober.go:28] interesting pod/router-default-5444994796-rrhqx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 03:42:50 crc kubenswrapper[4802]: [-]has-synced failed: reason withheld Dec 06 03:42:50 crc kubenswrapper[4802]: [+]process-running ok Dec 06 03:42:50 crc kubenswrapper[4802]: healthz check failed Dec 06 03:42:50 crc kubenswrapper[4802]: I1206 03:42:50.827078 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-rrhqx" podUID="0fbd23d0-a6b6-4a4e-b418-07fc38e5f497" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 03:42:51 crc kubenswrapper[4802]: I1206 03:42:51.826343 4802 patch_prober.go:28] interesting pod/router-default-5444994796-rrhqx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 03:42:51 crc kubenswrapper[4802]: [-]has-synced failed: reason withheld Dec 06 03:42:51 crc kubenswrapper[4802]: [+]process-running ok Dec 06 03:42:51 crc kubenswrapper[4802]: healthz check failed Dec 06 03:42:51 crc kubenswrapper[4802]: I1206 03:42:51.826486 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-rrhqx" podUID="0fbd23d0-a6b6-4a4e-b418-07fc38e5f497" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 03:42:52 crc kubenswrapper[4802]: I1206 03:42:52.825856 4802 patch_prober.go:28] interesting pod/router-default-5444994796-rrhqx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 03:42:52 crc kubenswrapper[4802]: [-]has-synced failed: reason withheld Dec 06 03:42:52 crc kubenswrapper[4802]: [+]process-running ok Dec 06 03:42:52 crc kubenswrapper[4802]: healthz check failed Dec 06 03:42:52 crc kubenswrapper[4802]: I1206 03:42:52.825973 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-rrhqx" podUID="0fbd23d0-a6b6-4a4e-b418-07fc38e5f497" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 03:42:53 crc kubenswrapper[4802]: I1206 03:42:53.947092 4802 patch_prober.go:28] interesting pod/console-f9d7485db-7ckfj container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.5:8443/health\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Dec 06 03:42:53 crc kubenswrapper[4802]: I1206 03:42:53.947215 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-7ckfj" podUID="2362780e-93d1-4f97-bda0-138ed5180bb0" containerName="console" probeResult="failure" output="Get \"https://10.217.0.5:8443/health\": dial tcp 10.217.0.5:8443: connect: connection refused" Dec 06 03:42:54 crc kubenswrapper[4802]: I1206 03:42:54.010660 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-rrhqx" Dec 06 03:42:54 crc kubenswrapper[4802]: I1206 03:42:54.014283 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-rrhqx" Dec 06 03:42:54 crc kubenswrapper[4802]: I1206 03:42:54.770960 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-7cg8t" Dec 06 03:42:55 crc kubenswrapper[4802]: I1206 03:42:55.297206 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 03:42:55 crc kubenswrapper[4802]: I1206 03:42:55.406817 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d0dacc47-2c44-4fb5-8f08-5015ab789df0-kube-api-access\") pod \"d0dacc47-2c44-4fb5-8f08-5015ab789df0\" (UID: \"d0dacc47-2c44-4fb5-8f08-5015ab789df0\") " Dec 06 03:42:55 crc kubenswrapper[4802]: I1206 03:42:55.406917 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d0dacc47-2c44-4fb5-8f08-5015ab789df0-kubelet-dir\") pod \"d0dacc47-2c44-4fb5-8f08-5015ab789df0\" (UID: \"d0dacc47-2c44-4fb5-8f08-5015ab789df0\") " Dec 06 03:42:55 crc kubenswrapper[4802]: I1206 03:42:55.407270 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d0dacc47-2c44-4fb5-8f08-5015ab789df0-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "d0dacc47-2c44-4fb5-8f08-5015ab789df0" (UID: "d0dacc47-2c44-4fb5-8f08-5015ab789df0"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:42:55 crc kubenswrapper[4802]: I1206 03:42:55.413924 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0dacc47-2c44-4fb5-8f08-5015ab789df0-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "d0dacc47-2c44-4fb5-8f08-5015ab789df0" (UID: "d0dacc47-2c44-4fb5-8f08-5015ab789df0"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:42:55 crc kubenswrapper[4802]: I1206 03:42:55.509044 4802 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d0dacc47-2c44-4fb5-8f08-5015ab789df0-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 06 03:42:55 crc kubenswrapper[4802]: I1206 03:42:55.509075 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d0dacc47-2c44-4fb5-8f08-5015ab789df0-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 03:42:55 crc kubenswrapper[4802]: I1206 03:42:55.785907 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"d0dacc47-2c44-4fb5-8f08-5015ab789df0","Type":"ContainerDied","Data":"dd3f60a800f995901b2ed66a25ae438d8fe7015eee21a84102d260ddc7168b33"} Dec 06 03:42:55 crc kubenswrapper[4802]: I1206 03:42:55.785948 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd3f60a800f995901b2ed66a25ae438d8fe7015eee21a84102d260ddc7168b33" Dec 06 03:42:55 crc kubenswrapper[4802]: I1206 03:42:55.785996 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 03:43:02 crc kubenswrapper[4802]: I1206 03:43:02.512669 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:43:03 crc kubenswrapper[4802]: I1206 03:43:03.951344 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-7ckfj" Dec 06 03:43:03 crc kubenswrapper[4802]: I1206 03:43:03.960715 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-7ckfj" Dec 06 03:43:13 crc kubenswrapper[4802]: I1206 03:43:13.283517 4802 patch_prober.go:28] interesting pod/machine-config-daemon-zpxxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 03:43:13 crc kubenswrapper[4802]: I1206 03:43:13.285017 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 03:43:13 crc kubenswrapper[4802]: E1206 03:43:13.453837 4802 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 06 03:43:13 crc kubenswrapper[4802]: E1206 03:43:13.454051 4802 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gqtns,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-h7kc8_openshift-marketplace(edb8213d-6a0c-4051-a230-4ec109d6e578): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 03:43:13 crc kubenswrapper[4802]: E1206 03:43:13.455360 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-h7kc8" podUID="edb8213d-6a0c-4051-a230-4ec109d6e578" Dec 06 03:43:14 crc kubenswrapper[4802]: E1206 03:43:14.410666 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-h7kc8" podUID="edb8213d-6a0c-4051-a230-4ec109d6e578" Dec 06 03:43:14 crc kubenswrapper[4802]: I1206 03:43:14.940082 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:43:15 crc kubenswrapper[4802]: I1206 03:43:15.521381 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wbjzw" Dec 06 03:43:19 crc kubenswrapper[4802]: E1206 03:43:19.490601 4802 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 06 03:43:19 crc kubenswrapper[4802]: E1206 03:43:19.492135 4802 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-b7qdd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-zm9dd_openshift-marketplace(d6eff099-6873-49c2-b0e9-e937700e1ff9): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 03:43:19 crc kubenswrapper[4802]: E1206 03:43:19.493778 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-zm9dd" podUID="d6eff099-6873-49c2-b0e9-e937700e1ff9" Dec 06 03:43:21 crc kubenswrapper[4802]: E1206 03:43:21.470534 4802 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 06 03:43:21 crc kubenswrapper[4802]: E1206 03:43:21.470682 4802 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-78lrv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-gq4cr_openshift-marketplace(4e6be98f-f5d2-4158-b4b8-eab47ad91564): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 03:43:21 crc kubenswrapper[4802]: E1206 03:43:21.471960 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-gq4cr" podUID="4e6be98f-f5d2-4158-b4b8-eab47ad91564" Dec 06 03:43:22 crc kubenswrapper[4802]: I1206 03:43:22.171351 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 06 03:43:22 crc kubenswrapper[4802]: E1206 03:43:22.171692 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ef5f0e2-fd47-4f85-aa9d-3a95fc157eed" containerName="pruner" Dec 06 03:43:22 crc kubenswrapper[4802]: I1206 03:43:22.171712 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ef5f0e2-fd47-4f85-aa9d-3a95fc157eed" containerName="pruner" Dec 06 03:43:22 crc kubenswrapper[4802]: E1206 03:43:22.171733 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0dacc47-2c44-4fb5-8f08-5015ab789df0" containerName="pruner" Dec 06 03:43:22 crc kubenswrapper[4802]: I1206 03:43:22.171745 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0dacc47-2c44-4fb5-8f08-5015ab789df0" containerName="pruner" Dec 06 03:43:22 crc kubenswrapper[4802]: I1206 03:43:22.171965 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0dacc47-2c44-4fb5-8f08-5015ab789df0" containerName="pruner" Dec 06 03:43:22 crc kubenswrapper[4802]: I1206 03:43:22.171993 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ef5f0e2-fd47-4f85-aa9d-3a95fc157eed" containerName="pruner" Dec 06 03:43:22 crc kubenswrapper[4802]: I1206 03:43:22.172628 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 03:43:22 crc kubenswrapper[4802]: I1206 03:43:22.175941 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 06 03:43:22 crc kubenswrapper[4802]: I1206 03:43:22.176293 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 06 03:43:22 crc kubenswrapper[4802]: I1206 03:43:22.182367 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 06 03:43:22 crc kubenswrapper[4802]: I1206 03:43:22.359928 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ecfd6acc-8b75-4988-b832-4e63778ca053-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"ecfd6acc-8b75-4988-b832-4e63778ca053\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 03:43:22 crc kubenswrapper[4802]: I1206 03:43:22.360088 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ecfd6acc-8b75-4988-b832-4e63778ca053-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"ecfd6acc-8b75-4988-b832-4e63778ca053\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 03:43:22 crc kubenswrapper[4802]: I1206 03:43:22.461639 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ecfd6acc-8b75-4988-b832-4e63778ca053-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"ecfd6acc-8b75-4988-b832-4e63778ca053\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 03:43:22 crc kubenswrapper[4802]: I1206 03:43:22.461770 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ecfd6acc-8b75-4988-b832-4e63778ca053-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"ecfd6acc-8b75-4988-b832-4e63778ca053\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 03:43:22 crc kubenswrapper[4802]: I1206 03:43:22.462219 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ecfd6acc-8b75-4988-b832-4e63778ca053-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"ecfd6acc-8b75-4988-b832-4e63778ca053\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 03:43:22 crc kubenswrapper[4802]: I1206 03:43:22.488561 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ecfd6acc-8b75-4988-b832-4e63778ca053-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"ecfd6acc-8b75-4988-b832-4e63778ca053\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 03:43:22 crc kubenswrapper[4802]: I1206 03:43:22.511274 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 03:43:22 crc kubenswrapper[4802]: E1206 03:43:22.559104 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-gq4cr" podUID="4e6be98f-f5d2-4158-b4b8-eab47ad91564" Dec 06 03:43:22 crc kubenswrapper[4802]: E1206 03:43:22.559113 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-zm9dd" podUID="d6eff099-6873-49c2-b0e9-e937700e1ff9" Dec 06 03:43:24 crc kubenswrapper[4802]: E1206 03:43:24.394771 4802 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 06 03:43:24 crc kubenswrapper[4802]: E1206 03:43:24.394933 4802 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wr5tq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-5smmc_openshift-marketplace(f3b86ac0-0a60-46f1-b82c-4bea67124a88): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 03:43:24 crc kubenswrapper[4802]: E1206 03:43:24.396147 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-5smmc" podUID="f3b86ac0-0a60-46f1-b82c-4bea67124a88" Dec 06 03:43:25 crc kubenswrapper[4802]: E1206 03:43:25.702519 4802 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 06 03:43:25 crc kubenswrapper[4802]: E1206 03:43:25.702689 4802 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-t5n77,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-z29bw_openshift-marketplace(9b68d090-3526-4778-ad64-51da5e2191a4): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 03:43:25 crc kubenswrapper[4802]: E1206 03:43:25.704006 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-z29bw" podUID="9b68d090-3526-4778-ad64-51da5e2191a4" Dec 06 03:43:26 crc kubenswrapper[4802]: E1206 03:43:26.131533 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-z29bw" podUID="9b68d090-3526-4778-ad64-51da5e2191a4" Dec 06 03:43:26 crc kubenswrapper[4802]: E1206 03:43:26.131606 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-5smmc" podUID="f3b86ac0-0a60-46f1-b82c-4bea67124a88" Dec 06 03:43:26 crc kubenswrapper[4802]: E1206 03:43:26.167398 4802 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 06 03:43:26 crc kubenswrapper[4802]: E1206 03:43:26.167879 4802 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-222z2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-sxwkm_openshift-marketplace(4a20041d-53aa-4179-b92e-9e7204c800ce): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 03:43:26 crc kubenswrapper[4802]: E1206 03:43:26.169730 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-sxwkm" podUID="4a20041d-53aa-4179-b92e-9e7204c800ce" Dec 06 03:43:26 crc kubenswrapper[4802]: I1206 03:43:26.553737 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 06 03:43:26 crc kubenswrapper[4802]: W1206 03:43:26.569249 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podecfd6acc_8b75_4988_b832_4e63778ca053.slice/crio-f02fea4d9b27a606b8c1e954a8670e83b3b7683ddccbbcdeab200f71598abf51 WatchSource:0}: Error finding container f02fea4d9b27a606b8c1e954a8670e83b3b7683ddccbbcdeab200f71598abf51: Status 404 returned error can't find the container with id f02fea4d9b27a606b8c1e954a8670e83b3b7683ddccbbcdeab200f71598abf51 Dec 06 03:43:26 crc kubenswrapper[4802]: I1206 03:43:26.631561 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-5ltx2"] Dec 06 03:43:26 crc kubenswrapper[4802]: W1206 03:43:26.648941 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8881599e_71e5_4578_b3f9_21f4fa2f57aa.slice/crio-e8b6059e6876558b3d5fe71c667ba0e900d95cea3a48de6eef67ffff5ef6366c WatchSource:0}: Error finding container e8b6059e6876558b3d5fe71c667ba0e900d95cea3a48de6eef67ffff5ef6366c: Status 404 returned error can't find the container with id e8b6059e6876558b3d5fe71c667ba0e900d95cea3a48de6eef67ffff5ef6366c Dec 06 03:43:26 crc kubenswrapper[4802]: I1206 03:43:26.764000 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 06 03:43:26 crc kubenswrapper[4802]: I1206 03:43:26.765008 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 06 03:43:26 crc kubenswrapper[4802]: I1206 03:43:26.786819 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 06 03:43:26 crc kubenswrapper[4802]: I1206 03:43:26.844968 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b58e10a8-7605-4863-bc97-91714add5409-kubelet-dir\") pod \"installer-9-crc\" (UID: \"b58e10a8-7605-4863-bc97-91714add5409\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 03:43:26 crc kubenswrapper[4802]: I1206 03:43:26.845311 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b58e10a8-7605-4863-bc97-91714add5409-var-lock\") pod \"installer-9-crc\" (UID: \"b58e10a8-7605-4863-bc97-91714add5409\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 03:43:26 crc kubenswrapper[4802]: I1206 03:43:26.845511 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b58e10a8-7605-4863-bc97-91714add5409-kube-api-access\") pod \"installer-9-crc\" (UID: \"b58e10a8-7605-4863-bc97-91714add5409\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 03:43:26 crc kubenswrapper[4802]: I1206 03:43:26.946735 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b58e10a8-7605-4863-bc97-91714add5409-kubelet-dir\") pod \"installer-9-crc\" (UID: \"b58e10a8-7605-4863-bc97-91714add5409\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 03:43:26 crc kubenswrapper[4802]: I1206 03:43:26.946855 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b58e10a8-7605-4863-bc97-91714add5409-var-lock\") pod \"installer-9-crc\" (UID: \"b58e10a8-7605-4863-bc97-91714add5409\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 03:43:26 crc kubenswrapper[4802]: I1206 03:43:26.946866 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b58e10a8-7605-4863-bc97-91714add5409-kubelet-dir\") pod \"installer-9-crc\" (UID: \"b58e10a8-7605-4863-bc97-91714add5409\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 03:43:26 crc kubenswrapper[4802]: I1206 03:43:26.946984 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b58e10a8-7605-4863-bc97-91714add5409-var-lock\") pod \"installer-9-crc\" (UID: \"b58e10a8-7605-4863-bc97-91714add5409\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 03:43:26 crc kubenswrapper[4802]: I1206 03:43:26.947094 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b58e10a8-7605-4863-bc97-91714add5409-kube-api-access\") pod \"installer-9-crc\" (UID: \"b58e10a8-7605-4863-bc97-91714add5409\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 03:43:26 crc kubenswrapper[4802]: I1206 03:43:26.972944 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b58e10a8-7605-4863-bc97-91714add5409-kube-api-access\") pod \"installer-9-crc\" (UID: \"b58e10a8-7605-4863-bc97-91714add5409\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 03:43:26 crc kubenswrapper[4802]: I1206 03:43:26.978841 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"ecfd6acc-8b75-4988-b832-4e63778ca053","Type":"ContainerStarted","Data":"f02fea4d9b27a606b8c1e954a8670e83b3b7683ddccbbcdeab200f71598abf51"} Dec 06 03:43:26 crc kubenswrapper[4802]: I1206 03:43:26.980287 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-5ltx2" event={"ID":"8881599e-71e5-4578-b3f9-21f4fa2f57aa","Type":"ContainerStarted","Data":"e8b6059e6876558b3d5fe71c667ba0e900d95cea3a48de6eef67ffff5ef6366c"} Dec 06 03:43:27 crc kubenswrapper[4802]: E1206 03:43:27.047959 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-sxwkm" podUID="4a20041d-53aa-4179-b92e-9e7204c800ce" Dec 06 03:43:27 crc kubenswrapper[4802]: I1206 03:43:27.097971 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 06 03:43:27 crc kubenswrapper[4802]: I1206 03:43:27.565042 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 06 03:43:27 crc kubenswrapper[4802]: W1206 03:43:27.591801 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podb58e10a8_7605_4863_bc97_91714add5409.slice/crio-96a2340d0cb485091cf317bb1624e1303513182f8ee577fdf1ad9284815390f5 WatchSource:0}: Error finding container 96a2340d0cb485091cf317bb1624e1303513182f8ee577fdf1ad9284815390f5: Status 404 returned error can't find the container with id 96a2340d0cb485091cf317bb1624e1303513182f8ee577fdf1ad9284815390f5 Dec 06 03:43:27 crc kubenswrapper[4802]: I1206 03:43:27.988360 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"b58e10a8-7605-4863-bc97-91714add5409","Type":"ContainerStarted","Data":"96a2340d0cb485091cf317bb1624e1303513182f8ee577fdf1ad9284815390f5"} Dec 06 03:43:28 crc kubenswrapper[4802]: E1206 03:43:28.812824 4802 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 06 03:43:28 crc kubenswrapper[4802]: E1206 03:43:28.813310 4802 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sstvn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-hqh24_openshift-marketplace(9ef50a75-3fb0-4a65-ae6a-45f981d5c979): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 03:43:28 crc kubenswrapper[4802]: E1206 03:43:28.814625 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-hqh24" podUID="9ef50a75-3fb0-4a65-ae6a-45f981d5c979" Dec 06 03:43:28 crc kubenswrapper[4802]: E1206 03:43:28.995548 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-hqh24" podUID="9ef50a75-3fb0-4a65-ae6a-45f981d5c979" Dec 06 03:43:30 crc kubenswrapper[4802]: I1206 03:43:30.006116 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-5ltx2" event={"ID":"8881599e-71e5-4578-b3f9-21f4fa2f57aa","Type":"ContainerStarted","Data":"65417a7b8bca84c54a11708c9a313548f221bedb97fd7c4ff55f34efa36e97ca"} Dec 06 03:43:30 crc kubenswrapper[4802]: I1206 03:43:30.009161 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"ecfd6acc-8b75-4988-b832-4e63778ca053","Type":"ContainerStarted","Data":"43e13c2b8597f2df091ea9ae01564d13671308276c8414d8976283f6b1479f31"} Dec 06 03:43:30 crc kubenswrapper[4802]: E1206 03:43:30.072892 4802 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 06 03:43:30 crc kubenswrapper[4802]: E1206 03:43:30.073066 4802 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wv7n5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-zzqgx_openshift-marketplace(03be8283-9ddf-4ecb-a1e7-579c322ca5d6): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 03:43:30 crc kubenswrapper[4802]: E1206 03:43:30.074283 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-zzqgx" podUID="03be8283-9ddf-4ecb-a1e7-579c322ca5d6" Dec 06 03:43:31 crc kubenswrapper[4802]: I1206 03:43:31.021303 4802 generic.go:334] "Generic (PLEG): container finished" podID="ecfd6acc-8b75-4988-b832-4e63778ca053" containerID="43e13c2b8597f2df091ea9ae01564d13671308276c8414d8976283f6b1479f31" exitCode=0 Dec 06 03:43:31 crc kubenswrapper[4802]: I1206 03:43:31.021361 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"ecfd6acc-8b75-4988-b832-4e63778ca053","Type":"ContainerDied","Data":"43e13c2b8597f2df091ea9ae01564d13671308276c8414d8976283f6b1479f31"} Dec 06 03:43:31 crc kubenswrapper[4802]: I1206 03:43:31.024052 4802 generic.go:334] "Generic (PLEG): container finished" podID="edb8213d-6a0c-4051-a230-4ec109d6e578" containerID="4493b583a803e12b2a5080c894164167f920ccc635ab9ad94afb11aecd4e0417" exitCode=0 Dec 06 03:43:31 crc kubenswrapper[4802]: I1206 03:43:31.024097 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h7kc8" event={"ID":"edb8213d-6a0c-4051-a230-4ec109d6e578","Type":"ContainerDied","Data":"4493b583a803e12b2a5080c894164167f920ccc635ab9ad94afb11aecd4e0417"} Dec 06 03:43:31 crc kubenswrapper[4802]: I1206 03:43:31.027970 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"b58e10a8-7605-4863-bc97-91714add5409","Type":"ContainerStarted","Data":"af3ed61c6bef9dc4e55515ee327d8928fa08e228412d55611fd95b430c653929"} Dec 06 03:43:31 crc kubenswrapper[4802]: I1206 03:43:31.029630 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-5ltx2" event={"ID":"8881599e-71e5-4578-b3f9-21f4fa2f57aa","Type":"ContainerStarted","Data":"f5bd8ffe79e78e1c8abfdbf7be496d80881338ea054c8023e28e7532aea9f5a0"} Dec 06 03:43:31 crc kubenswrapper[4802]: E1206 03:43:31.030844 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-zzqgx" podUID="03be8283-9ddf-4ecb-a1e7-579c322ca5d6" Dec 06 03:43:31 crc kubenswrapper[4802]: I1206 03:43:31.069328 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=5.069309242 podStartE2EDuration="5.069309242s" podCreationTimestamp="2025-12-06 03:43:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:43:31.062903807 +0000 UTC m=+203.934812969" watchObservedRunningTime="2025-12-06 03:43:31.069309242 +0000 UTC m=+203.941218394" Dec 06 03:43:31 crc kubenswrapper[4802]: I1206 03:43:31.081368 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-5ltx2" podStartSLOduration=185.081345942 podStartE2EDuration="3m5.081345942s" podCreationTimestamp="2025-12-06 03:40:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:43:31.079661415 +0000 UTC m=+203.951570577" watchObservedRunningTime="2025-12-06 03:43:31.081345942 +0000 UTC m=+203.953255104" Dec 06 03:43:32 crc kubenswrapper[4802]: I1206 03:43:32.037465 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h7kc8" event={"ID":"edb8213d-6a0c-4051-a230-4ec109d6e578","Type":"ContainerStarted","Data":"90422ef940487979c738c9408947ea9a9c23db7a8748cbfb02434a53708b6a53"} Dec 06 03:43:32 crc kubenswrapper[4802]: I1206 03:43:32.060948 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-h7kc8" podStartSLOduration=2.827244465 podStartE2EDuration="49.060930212s" podCreationTimestamp="2025-12-06 03:42:43 +0000 UTC" firstStartedPulling="2025-12-06 03:42:45.376722198 +0000 UTC m=+158.248631350" lastFinishedPulling="2025-12-06 03:43:31.610407915 +0000 UTC m=+204.482317097" observedRunningTime="2025-12-06 03:43:32.057517499 +0000 UTC m=+204.929426651" watchObservedRunningTime="2025-12-06 03:43:32.060930212 +0000 UTC m=+204.932839364" Dec 06 03:43:32 crc kubenswrapper[4802]: I1206 03:43:32.360814 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 03:43:32 crc kubenswrapper[4802]: I1206 03:43:32.558009 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ecfd6acc-8b75-4988-b832-4e63778ca053-kube-api-access\") pod \"ecfd6acc-8b75-4988-b832-4e63778ca053\" (UID: \"ecfd6acc-8b75-4988-b832-4e63778ca053\") " Dec 06 03:43:32 crc kubenswrapper[4802]: I1206 03:43:32.558071 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ecfd6acc-8b75-4988-b832-4e63778ca053-kubelet-dir\") pod \"ecfd6acc-8b75-4988-b832-4e63778ca053\" (UID: \"ecfd6acc-8b75-4988-b832-4e63778ca053\") " Dec 06 03:43:32 crc kubenswrapper[4802]: I1206 03:43:32.558321 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ecfd6acc-8b75-4988-b832-4e63778ca053-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "ecfd6acc-8b75-4988-b832-4e63778ca053" (UID: "ecfd6acc-8b75-4988-b832-4e63778ca053"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:43:32 crc kubenswrapper[4802]: I1206 03:43:32.562892 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecfd6acc-8b75-4988-b832-4e63778ca053-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "ecfd6acc-8b75-4988-b832-4e63778ca053" (UID: "ecfd6acc-8b75-4988-b832-4e63778ca053"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:43:32 crc kubenswrapper[4802]: I1206 03:43:32.659603 4802 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ecfd6acc-8b75-4988-b832-4e63778ca053-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 06 03:43:32 crc kubenswrapper[4802]: I1206 03:43:32.659635 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ecfd6acc-8b75-4988-b832-4e63778ca053-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 03:43:33 crc kubenswrapper[4802]: I1206 03:43:33.047412 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"ecfd6acc-8b75-4988-b832-4e63778ca053","Type":"ContainerDied","Data":"f02fea4d9b27a606b8c1e954a8670e83b3b7683ddccbbcdeab200f71598abf51"} Dec 06 03:43:33 crc kubenswrapper[4802]: I1206 03:43:33.047474 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f02fea4d9b27a606b8c1e954a8670e83b3b7683ddccbbcdeab200f71598abf51" Dec 06 03:43:33 crc kubenswrapper[4802]: I1206 03:43:33.047580 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 03:43:34 crc kubenswrapper[4802]: I1206 03:43:34.140464 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-h7kc8" Dec 06 03:43:34 crc kubenswrapper[4802]: I1206 03:43:34.140889 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-h7kc8" Dec 06 03:43:35 crc kubenswrapper[4802]: I1206 03:43:35.242115 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-h7kc8" podUID="edb8213d-6a0c-4051-a230-4ec109d6e578" containerName="registry-server" probeResult="failure" output=< Dec 06 03:43:35 crc kubenswrapper[4802]: timeout: failed to connect service ":50051" within 1s Dec 06 03:43:35 crc kubenswrapper[4802]: > Dec 06 03:43:36 crc kubenswrapper[4802]: I1206 03:43:36.071323 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zm9dd" event={"ID":"d6eff099-6873-49c2-b0e9-e937700e1ff9","Type":"ContainerStarted","Data":"cc62a36bdacf3ba8a99e11eaf8ebb097bac4d29f41a5d54766153cad81379503"} Dec 06 03:43:36 crc kubenswrapper[4802]: I1206 03:43:36.076204 4802 generic.go:334] "Generic (PLEG): container finished" podID="4e6be98f-f5d2-4158-b4b8-eab47ad91564" containerID="a529b48adcf34e1d5cade9edc43807acaf894be32f2848981ea7c16525ca0d30" exitCode=0 Dec 06 03:43:36 crc kubenswrapper[4802]: I1206 03:43:36.076261 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gq4cr" event={"ID":"4e6be98f-f5d2-4158-b4b8-eab47ad91564","Type":"ContainerDied","Data":"a529b48adcf34e1d5cade9edc43807acaf894be32f2848981ea7c16525ca0d30"} Dec 06 03:43:37 crc kubenswrapper[4802]: I1206 03:43:37.082571 4802 generic.go:334] "Generic (PLEG): container finished" podID="d6eff099-6873-49c2-b0e9-e937700e1ff9" containerID="cc62a36bdacf3ba8a99e11eaf8ebb097bac4d29f41a5d54766153cad81379503" exitCode=0 Dec 06 03:43:37 crc kubenswrapper[4802]: I1206 03:43:37.082631 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zm9dd" event={"ID":"d6eff099-6873-49c2-b0e9-e937700e1ff9","Type":"ContainerDied","Data":"cc62a36bdacf3ba8a99e11eaf8ebb097bac4d29f41a5d54766153cad81379503"} Dec 06 03:43:37 crc kubenswrapper[4802]: I1206 03:43:37.086484 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gq4cr" event={"ID":"4e6be98f-f5d2-4158-b4b8-eab47ad91564","Type":"ContainerStarted","Data":"9873ca269bb9418c9f670d016e76aafb2cfb6ca2bb7874c3e0b5a44bdf212b52"} Dec 06 03:43:37 crc kubenswrapper[4802]: I1206 03:43:37.114697 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-gq4cr" podStartSLOduration=2.659281584 podStartE2EDuration="56.114682367s" podCreationTimestamp="2025-12-06 03:42:41 +0000 UTC" firstStartedPulling="2025-12-06 03:42:43.2974946 +0000 UTC m=+156.169403752" lastFinishedPulling="2025-12-06 03:43:36.752895373 +0000 UTC m=+209.624804535" observedRunningTime="2025-12-06 03:43:37.112518578 +0000 UTC m=+209.984427730" watchObservedRunningTime="2025-12-06 03:43:37.114682367 +0000 UTC m=+209.986591519" Dec 06 03:43:39 crc kubenswrapper[4802]: I1206 03:43:39.107795 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zm9dd" event={"ID":"d6eff099-6873-49c2-b0e9-e937700e1ff9","Type":"ContainerStarted","Data":"8f5b2f642eb25588559a3e8e6cdf0906915e627fbe7c5e4b91d5aa05018ede14"} Dec 06 03:43:39 crc kubenswrapper[4802]: I1206 03:43:39.132898 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-zm9dd" podStartSLOduration=3.521561057 podStartE2EDuration="55.132876383s" podCreationTimestamp="2025-12-06 03:42:44 +0000 UTC" firstStartedPulling="2025-12-06 03:42:46.422556153 +0000 UTC m=+159.294465305" lastFinishedPulling="2025-12-06 03:43:38.033871469 +0000 UTC m=+210.905780631" observedRunningTime="2025-12-06 03:43:39.131436453 +0000 UTC m=+212.003345605" watchObservedRunningTime="2025-12-06 03:43:39.132876383 +0000 UTC m=+212.004785535" Dec 06 03:43:39 crc kubenswrapper[4802]: E1206 03:43:39.412206 4802 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9b68d090_3526_4778_ad64_51da5e2191a4.slice/crio-eda8799283b13a2b7a49f02e60f9897e2b46fc5cd80385de5d493d3bafa10830.scope\": RecentStats: unable to find data in memory cache]" Dec 06 03:43:40 crc kubenswrapper[4802]: I1206 03:43:40.113987 4802 generic.go:334] "Generic (PLEG): container finished" podID="9b68d090-3526-4778-ad64-51da5e2191a4" containerID="eda8799283b13a2b7a49f02e60f9897e2b46fc5cd80385de5d493d3bafa10830" exitCode=0 Dec 06 03:43:40 crc kubenswrapper[4802]: I1206 03:43:40.114096 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z29bw" event={"ID":"9b68d090-3526-4778-ad64-51da5e2191a4","Type":"ContainerDied","Data":"eda8799283b13a2b7a49f02e60f9897e2b46fc5cd80385de5d493d3bafa10830"} Dec 06 03:43:41 crc kubenswrapper[4802]: I1206 03:43:41.516565 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-gq4cr" Dec 06 03:43:41 crc kubenswrapper[4802]: I1206 03:43:41.516647 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-gq4cr" Dec 06 03:43:41 crc kubenswrapper[4802]: I1206 03:43:41.560003 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-gq4cr" Dec 06 03:43:42 crc kubenswrapper[4802]: I1206 03:43:42.134030 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z29bw" event={"ID":"9b68d090-3526-4778-ad64-51da5e2191a4","Type":"ContainerStarted","Data":"aa53a5441403a493a64d80cb023e5f5513aef3f032db63372ccab0cf61f3c480"} Dec 06 03:43:42 crc kubenswrapper[4802]: I1206 03:43:42.172999 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-z29bw" podStartSLOduration=4.284409809 podStartE2EDuration="1m2.172984736s" podCreationTimestamp="2025-12-06 03:42:40 +0000 UTC" firstStartedPulling="2025-12-06 03:42:43.303217724 +0000 UTC m=+156.175126876" lastFinishedPulling="2025-12-06 03:43:41.191792611 +0000 UTC m=+214.063701803" observedRunningTime="2025-12-06 03:43:42.172139703 +0000 UTC m=+215.044048855" watchObservedRunningTime="2025-12-06 03:43:42.172984736 +0000 UTC m=+215.044893888" Dec 06 03:43:42 crc kubenswrapper[4802]: I1206 03:43:42.194405 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-gq4cr" Dec 06 03:43:43 crc kubenswrapper[4802]: I1206 03:43:43.283588 4802 patch_prober.go:28] interesting pod/machine-config-daemon-zpxxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 03:43:43 crc kubenswrapper[4802]: I1206 03:43:43.283654 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 03:43:43 crc kubenswrapper[4802]: I1206 03:43:43.283730 4802 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" Dec 06 03:43:43 crc kubenswrapper[4802]: I1206 03:43:43.284331 4802 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3bf5e41701c7a2d45adc4740ceb4ac1785db17bc7cb15daed439f7556812d1e9"} pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 03:43:43 crc kubenswrapper[4802]: I1206 03:43:43.284429 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" containerID="cri-o://3bf5e41701c7a2d45adc4740ceb4ac1785db17bc7cb15daed439f7556812d1e9" gracePeriod=600 Dec 06 03:43:44 crc kubenswrapper[4802]: I1206 03:43:44.149397 4802 generic.go:334] "Generic (PLEG): container finished" podID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerID="3bf5e41701c7a2d45adc4740ceb4ac1785db17bc7cb15daed439f7556812d1e9" exitCode=0 Dec 06 03:43:44 crc kubenswrapper[4802]: I1206 03:43:44.149773 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" event={"ID":"6df38316-e0d3-4018-8d27-3620eba3a68d","Type":"ContainerDied","Data":"3bf5e41701c7a2d45adc4740ceb4ac1785db17bc7cb15daed439f7556812d1e9"} Dec 06 03:43:44 crc kubenswrapper[4802]: I1206 03:43:44.187944 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-h7kc8" Dec 06 03:43:44 crc kubenswrapper[4802]: I1206 03:43:44.240595 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-h7kc8" Dec 06 03:43:44 crc kubenswrapper[4802]: I1206 03:43:44.550968 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-zm9dd" Dec 06 03:43:44 crc kubenswrapper[4802]: I1206 03:43:44.551021 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-zm9dd" Dec 06 03:43:44 crc kubenswrapper[4802]: I1206 03:43:44.591000 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-zm9dd" Dec 06 03:43:45 crc kubenswrapper[4802]: I1206 03:43:45.196506 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-zm9dd" Dec 06 03:43:45 crc kubenswrapper[4802]: I1206 03:43:45.295960 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gq4cr"] Dec 06 03:43:45 crc kubenswrapper[4802]: I1206 03:43:45.296183 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-gq4cr" podUID="4e6be98f-f5d2-4158-b4b8-eab47ad91564" containerName="registry-server" containerID="cri-o://9873ca269bb9418c9f670d016e76aafb2cfb6ca2bb7874c3e0b5a44bdf212b52" gracePeriod=2 Dec 06 03:43:47 crc kubenswrapper[4802]: I1206 03:43:47.168046 4802 generic.go:334] "Generic (PLEG): container finished" podID="4e6be98f-f5d2-4158-b4b8-eab47ad91564" containerID="9873ca269bb9418c9f670d016e76aafb2cfb6ca2bb7874c3e0b5a44bdf212b52" exitCode=0 Dec 06 03:43:47 crc kubenswrapper[4802]: I1206 03:43:47.168093 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gq4cr" event={"ID":"4e6be98f-f5d2-4158-b4b8-eab47ad91564","Type":"ContainerDied","Data":"9873ca269bb9418c9f670d016e76aafb2cfb6ca2bb7874c3e0b5a44bdf212b52"} Dec 06 03:43:48 crc kubenswrapper[4802]: I1206 03:43:48.294070 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zm9dd"] Dec 06 03:43:48 crc kubenswrapper[4802]: I1206 03:43:48.294357 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-zm9dd" podUID="d6eff099-6873-49c2-b0e9-e937700e1ff9" containerName="registry-server" containerID="cri-o://8f5b2f642eb25588559a3e8e6cdf0906915e627fbe7c5e4b91d5aa05018ede14" gracePeriod=2 Dec 06 03:43:50 crc kubenswrapper[4802]: I1206 03:43:50.186829 4802 generic.go:334] "Generic (PLEG): container finished" podID="d6eff099-6873-49c2-b0e9-e937700e1ff9" containerID="8f5b2f642eb25588559a3e8e6cdf0906915e627fbe7c5e4b91d5aa05018ede14" exitCode=0 Dec 06 03:43:50 crc kubenswrapper[4802]: I1206 03:43:50.186909 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zm9dd" event={"ID":"d6eff099-6873-49c2-b0e9-e937700e1ff9","Type":"ContainerDied","Data":"8f5b2f642eb25588559a3e8e6cdf0906915e627fbe7c5e4b91d5aa05018ede14"} Dec 06 03:43:51 crc kubenswrapper[4802]: I1206 03:43:51.101142 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-z29bw" Dec 06 03:43:51 crc kubenswrapper[4802]: I1206 03:43:51.101184 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-z29bw" Dec 06 03:43:51 crc kubenswrapper[4802]: I1206 03:43:51.146986 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-z29bw" Dec 06 03:43:51 crc kubenswrapper[4802]: I1206 03:43:51.233491 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-z29bw" Dec 06 03:43:51 crc kubenswrapper[4802]: I1206 03:43:51.418300 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gq4cr" Dec 06 03:43:51 crc kubenswrapper[4802]: I1206 03:43:51.604110 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-78lrv\" (UniqueName: \"kubernetes.io/projected/4e6be98f-f5d2-4158-b4b8-eab47ad91564-kube-api-access-78lrv\") pod \"4e6be98f-f5d2-4158-b4b8-eab47ad91564\" (UID: \"4e6be98f-f5d2-4158-b4b8-eab47ad91564\") " Dec 06 03:43:51 crc kubenswrapper[4802]: I1206 03:43:51.604511 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e6be98f-f5d2-4158-b4b8-eab47ad91564-catalog-content\") pod \"4e6be98f-f5d2-4158-b4b8-eab47ad91564\" (UID: \"4e6be98f-f5d2-4158-b4b8-eab47ad91564\") " Dec 06 03:43:51 crc kubenswrapper[4802]: I1206 03:43:51.604562 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e6be98f-f5d2-4158-b4b8-eab47ad91564-utilities\") pod \"4e6be98f-f5d2-4158-b4b8-eab47ad91564\" (UID: \"4e6be98f-f5d2-4158-b4b8-eab47ad91564\") " Dec 06 03:43:51 crc kubenswrapper[4802]: I1206 03:43:51.605373 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e6be98f-f5d2-4158-b4b8-eab47ad91564-utilities" (OuterVolumeSpecName: "utilities") pod "4e6be98f-f5d2-4158-b4b8-eab47ad91564" (UID: "4e6be98f-f5d2-4158-b4b8-eab47ad91564"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:43:51 crc kubenswrapper[4802]: I1206 03:43:51.614924 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e6be98f-f5d2-4158-b4b8-eab47ad91564-kube-api-access-78lrv" (OuterVolumeSpecName: "kube-api-access-78lrv") pod "4e6be98f-f5d2-4158-b4b8-eab47ad91564" (UID: "4e6be98f-f5d2-4158-b4b8-eab47ad91564"). InnerVolumeSpecName "kube-api-access-78lrv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:43:51 crc kubenswrapper[4802]: I1206 03:43:51.635187 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zm9dd" Dec 06 03:43:51 crc kubenswrapper[4802]: I1206 03:43:51.659190 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e6be98f-f5d2-4158-b4b8-eab47ad91564-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4e6be98f-f5d2-4158-b4b8-eab47ad91564" (UID: "4e6be98f-f5d2-4158-b4b8-eab47ad91564"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:43:51 crc kubenswrapper[4802]: I1206 03:43:51.705168 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-78lrv\" (UniqueName: \"kubernetes.io/projected/4e6be98f-f5d2-4158-b4b8-eab47ad91564-kube-api-access-78lrv\") on node \"crc\" DevicePath \"\"" Dec 06 03:43:51 crc kubenswrapper[4802]: I1206 03:43:51.705230 4802 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e6be98f-f5d2-4158-b4b8-eab47ad91564-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 03:43:51 crc kubenswrapper[4802]: I1206 03:43:51.705248 4802 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e6be98f-f5d2-4158-b4b8-eab47ad91564-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 03:43:51 crc kubenswrapper[4802]: I1206 03:43:51.805645 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6eff099-6873-49c2-b0e9-e937700e1ff9-catalog-content\") pod \"d6eff099-6873-49c2-b0e9-e937700e1ff9\" (UID: \"d6eff099-6873-49c2-b0e9-e937700e1ff9\") " Dec 06 03:43:51 crc kubenswrapper[4802]: I1206 03:43:51.807206 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b7qdd\" (UniqueName: \"kubernetes.io/projected/d6eff099-6873-49c2-b0e9-e937700e1ff9-kube-api-access-b7qdd\") pod \"d6eff099-6873-49c2-b0e9-e937700e1ff9\" (UID: \"d6eff099-6873-49c2-b0e9-e937700e1ff9\") " Dec 06 03:43:51 crc kubenswrapper[4802]: I1206 03:43:51.807346 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6eff099-6873-49c2-b0e9-e937700e1ff9-utilities\") pod \"d6eff099-6873-49c2-b0e9-e937700e1ff9\" (UID: \"d6eff099-6873-49c2-b0e9-e937700e1ff9\") " Dec 06 03:43:51 crc kubenswrapper[4802]: I1206 03:43:51.808546 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6eff099-6873-49c2-b0e9-e937700e1ff9-utilities" (OuterVolumeSpecName: "utilities") pod "d6eff099-6873-49c2-b0e9-e937700e1ff9" (UID: "d6eff099-6873-49c2-b0e9-e937700e1ff9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:43:51 crc kubenswrapper[4802]: I1206 03:43:51.810630 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6eff099-6873-49c2-b0e9-e937700e1ff9-kube-api-access-b7qdd" (OuterVolumeSpecName: "kube-api-access-b7qdd") pod "d6eff099-6873-49c2-b0e9-e937700e1ff9" (UID: "d6eff099-6873-49c2-b0e9-e937700e1ff9"). InnerVolumeSpecName "kube-api-access-b7qdd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:43:51 crc kubenswrapper[4802]: I1206 03:43:51.908703 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b7qdd\" (UniqueName: \"kubernetes.io/projected/d6eff099-6873-49c2-b0e9-e937700e1ff9-kube-api-access-b7qdd\") on node \"crc\" DevicePath \"\"" Dec 06 03:43:51 crc kubenswrapper[4802]: I1206 03:43:51.908736 4802 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6eff099-6873-49c2-b0e9-e937700e1ff9-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 03:43:51 crc kubenswrapper[4802]: I1206 03:43:51.928831 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6eff099-6873-49c2-b0e9-e937700e1ff9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d6eff099-6873-49c2-b0e9-e937700e1ff9" (UID: "d6eff099-6873-49c2-b0e9-e937700e1ff9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:43:52 crc kubenswrapper[4802]: I1206 03:43:52.010743 4802 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6eff099-6873-49c2-b0e9-e937700e1ff9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 03:43:52 crc kubenswrapper[4802]: I1206 03:43:52.198007 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zm9dd" event={"ID":"d6eff099-6873-49c2-b0e9-e937700e1ff9","Type":"ContainerDied","Data":"b927bc37e72ad725e18780474a2300daa9a0e1eff333761d21a0868a8b5c8154"} Dec 06 03:43:52 crc kubenswrapper[4802]: I1206 03:43:52.198029 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zm9dd" Dec 06 03:43:52 crc kubenswrapper[4802]: I1206 03:43:52.198336 4802 scope.go:117] "RemoveContainer" containerID="8f5b2f642eb25588559a3e8e6cdf0906915e627fbe7c5e4b91d5aa05018ede14" Dec 06 03:43:52 crc kubenswrapper[4802]: I1206 03:43:52.200322 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" event={"ID":"6df38316-e0d3-4018-8d27-3620eba3a68d","Type":"ContainerStarted","Data":"2200194f7108affdd47802bf08946e25d3e937d094def5243f711369c5ae1305"} Dec 06 03:43:52 crc kubenswrapper[4802]: I1206 03:43:52.203296 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gq4cr" event={"ID":"4e6be98f-f5d2-4158-b4b8-eab47ad91564","Type":"ContainerDied","Data":"afa20aa2013c5eda74a9c65ef47f3598d8e6a2395f06742ef522a880728662ba"} Dec 06 03:43:52 crc kubenswrapper[4802]: I1206 03:43:52.203338 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gq4cr" Dec 06 03:43:52 crc kubenswrapper[4802]: I1206 03:43:52.211260 4802 scope.go:117] "RemoveContainer" containerID="cc62a36bdacf3ba8a99e11eaf8ebb097bac4d29f41a5d54766153cad81379503" Dec 06 03:43:52 crc kubenswrapper[4802]: I1206 03:43:52.232507 4802 scope.go:117] "RemoveContainer" containerID="0582177ab2c3f046a646db5bbaffb24446c1f48184610e616537757ac4ee387e" Dec 06 03:43:52 crc kubenswrapper[4802]: I1206 03:43:52.242399 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gq4cr"] Dec 06 03:43:52 crc kubenswrapper[4802]: I1206 03:43:52.246581 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-gq4cr"] Dec 06 03:43:52 crc kubenswrapper[4802]: I1206 03:43:52.256399 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zm9dd"] Dec 06 03:43:52 crc kubenswrapper[4802]: I1206 03:43:52.261016 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-zm9dd"] Dec 06 03:43:52 crc kubenswrapper[4802]: I1206 03:43:52.263546 4802 scope.go:117] "RemoveContainer" containerID="9873ca269bb9418c9f670d016e76aafb2cfb6ca2bb7874c3e0b5a44bdf212b52" Dec 06 03:43:52 crc kubenswrapper[4802]: I1206 03:43:52.274684 4802 scope.go:117] "RemoveContainer" containerID="a529b48adcf34e1d5cade9edc43807acaf894be32f2848981ea7c16525ca0d30" Dec 06 03:43:52 crc kubenswrapper[4802]: I1206 03:43:52.286633 4802 scope.go:117] "RemoveContainer" containerID="030dcc3082c01a924aff1c1d291341852b7b09c6ccb9682fff26368e3879f542" Dec 06 03:43:53 crc kubenswrapper[4802]: I1206 03:43:53.210974 4802 generic.go:334] "Generic (PLEG): container finished" podID="9ef50a75-3fb0-4a65-ae6a-45f981d5c979" containerID="ffb979adb26fcbc6e9bd32c1fb2aca0c0a69106462aaedd90702d4fdbaf95047" exitCode=0 Dec 06 03:43:53 crc kubenswrapper[4802]: I1206 03:43:53.211049 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hqh24" event={"ID":"9ef50a75-3fb0-4a65-ae6a-45f981d5c979","Type":"ContainerDied","Data":"ffb979adb26fcbc6e9bd32c1fb2aca0c0a69106462aaedd90702d4fdbaf95047"} Dec 06 03:43:53 crc kubenswrapper[4802]: I1206 03:43:53.217412 4802 generic.go:334] "Generic (PLEG): container finished" podID="03be8283-9ddf-4ecb-a1e7-579c322ca5d6" containerID="b5466e90518b9c52f5efd636a232d9dc07215fcbacbe49f3067e85ab55e8a114" exitCode=0 Dec 06 03:43:53 crc kubenswrapper[4802]: I1206 03:43:53.217457 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zzqgx" event={"ID":"03be8283-9ddf-4ecb-a1e7-579c322ca5d6","Type":"ContainerDied","Data":"b5466e90518b9c52f5efd636a232d9dc07215fcbacbe49f3067e85ab55e8a114"} Dec 06 03:43:53 crc kubenswrapper[4802]: I1206 03:43:53.222422 4802 generic.go:334] "Generic (PLEG): container finished" podID="f3b86ac0-0a60-46f1-b82c-4bea67124a88" containerID="7f766f943a6941f0810f824d83449e726b163fa01b79b4d68e64a07213508279" exitCode=0 Dec 06 03:43:53 crc kubenswrapper[4802]: I1206 03:43:53.222473 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5smmc" event={"ID":"f3b86ac0-0a60-46f1-b82c-4bea67124a88","Type":"ContainerDied","Data":"7f766f943a6941f0810f824d83449e726b163fa01b79b4d68e64a07213508279"} Dec 06 03:43:53 crc kubenswrapper[4802]: I1206 03:43:53.226779 4802 generic.go:334] "Generic (PLEG): container finished" podID="4a20041d-53aa-4179-b92e-9e7204c800ce" containerID="ee4f28d8dcd84e0f39cad21e47bf46a87f02d12b3516a7ffa1b02c0c1882a796" exitCode=0 Dec 06 03:43:53 crc kubenswrapper[4802]: I1206 03:43:53.226914 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sxwkm" event={"ID":"4a20041d-53aa-4179-b92e-9e7204c800ce","Type":"ContainerDied","Data":"ee4f28d8dcd84e0f39cad21e47bf46a87f02d12b3516a7ffa1b02c0c1882a796"} Dec 06 03:43:53 crc kubenswrapper[4802]: I1206 03:43:53.463081 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e6be98f-f5d2-4158-b4b8-eab47ad91564" path="/var/lib/kubelet/pods/4e6be98f-f5d2-4158-b4b8-eab47ad91564/volumes" Dec 06 03:43:53 crc kubenswrapper[4802]: I1206 03:43:53.464087 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6eff099-6873-49c2-b0e9-e937700e1ff9" path="/var/lib/kubelet/pods/d6eff099-6873-49c2-b0e9-e937700e1ff9/volumes" Dec 06 03:43:54 crc kubenswrapper[4802]: I1206 03:43:54.235626 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hqh24" event={"ID":"9ef50a75-3fb0-4a65-ae6a-45f981d5c979","Type":"ContainerStarted","Data":"526d21aa930fe3d9a935b02bf3298f63b99e6fd460f906b2cb8ff06a5cc1ed1a"} Dec 06 03:43:54 crc kubenswrapper[4802]: I1206 03:43:54.238408 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zzqgx" event={"ID":"03be8283-9ddf-4ecb-a1e7-579c322ca5d6","Type":"ContainerStarted","Data":"27f211beea61c680cd529158e01283a1070e9c6a850425273880787fd7750485"} Dec 06 03:43:54 crc kubenswrapper[4802]: I1206 03:43:54.241543 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5smmc" event={"ID":"f3b86ac0-0a60-46f1-b82c-4bea67124a88","Type":"ContainerStarted","Data":"7d3b07a31ed9406630f598bfd3d5db7a2d6d651c5b05ccadb53cf46f1c71b97b"} Dec 06 03:43:54 crc kubenswrapper[4802]: I1206 03:43:54.243797 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sxwkm" event={"ID":"4a20041d-53aa-4179-b92e-9e7204c800ce","Type":"ContainerStarted","Data":"01f107ede9c9ccf901332436b6fbe30835290ff8a074f09d367bacbcbd39ee7a"} Dec 06 03:43:54 crc kubenswrapper[4802]: I1206 03:43:54.252866 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hqh24" podStartSLOduration=3.8331304680000002 podStartE2EDuration="1m14.252850618s" podCreationTimestamp="2025-12-06 03:42:40 +0000 UTC" firstStartedPulling="2025-12-06 03:42:43.305081853 +0000 UTC m=+156.176991005" lastFinishedPulling="2025-12-06 03:43:53.724802003 +0000 UTC m=+226.596711155" observedRunningTime="2025-12-06 03:43:54.250545925 +0000 UTC m=+227.122455087" watchObservedRunningTime="2025-12-06 03:43:54.252850618 +0000 UTC m=+227.124759770" Dec 06 03:43:54 crc kubenswrapper[4802]: I1206 03:43:54.287022 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zzqgx" podStartSLOduration=3.976472589 podStartE2EDuration="1m14.287002231s" podCreationTimestamp="2025-12-06 03:42:40 +0000 UTC" firstStartedPulling="2025-12-06 03:42:43.296438381 +0000 UTC m=+156.168347533" lastFinishedPulling="2025-12-06 03:43:53.606968013 +0000 UTC m=+226.478877175" observedRunningTime="2025-12-06 03:43:54.265316619 +0000 UTC m=+227.137225771" watchObservedRunningTime="2025-12-06 03:43:54.287002231 +0000 UTC m=+227.158911383" Dec 06 03:43:54 crc kubenswrapper[4802]: I1206 03:43:54.290081 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5smmc" podStartSLOduration=2.997952639 podStartE2EDuration="1m12.290064915s" podCreationTimestamp="2025-12-06 03:42:42 +0000 UTC" firstStartedPulling="2025-12-06 03:42:44.346673116 +0000 UTC m=+157.218582268" lastFinishedPulling="2025-12-06 03:43:53.638785392 +0000 UTC m=+226.510694544" observedRunningTime="2025-12-06 03:43:54.285939502 +0000 UTC m=+227.157848654" watchObservedRunningTime="2025-12-06 03:43:54.290064915 +0000 UTC m=+227.161974067" Dec 06 03:43:54 crc kubenswrapper[4802]: I1206 03:43:54.316064 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-sxwkm" podStartSLOduration=1.9849722220000001 podStartE2EDuration="1m11.316028584s" podCreationTimestamp="2025-12-06 03:42:43 +0000 UTC" firstStartedPulling="2025-12-06 03:42:44.327655215 +0000 UTC m=+157.199564367" lastFinishedPulling="2025-12-06 03:43:53.658711577 +0000 UTC m=+226.530620729" observedRunningTime="2025-12-06 03:43:54.313882275 +0000 UTC m=+227.185791437" watchObservedRunningTime="2025-12-06 03:43:54.316028584 +0000 UTC m=+227.187937736" Dec 06 03:44:00 crc kubenswrapper[4802]: I1206 03:44:00.813370 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-z29bw"] Dec 06 03:44:00 crc kubenswrapper[4802]: I1206 03:44:00.814082 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-z29bw" podUID="9b68d090-3526-4778-ad64-51da5e2191a4" containerName="registry-server" containerID="cri-o://aa53a5441403a493a64d80cb023e5f5513aef3f032db63372ccab0cf61f3c480" gracePeriod=30 Dec 06 03:44:00 crc kubenswrapper[4802]: I1206 03:44:00.823629 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hqh24"] Dec 06 03:44:00 crc kubenswrapper[4802]: I1206 03:44:00.824080 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hqh24" podUID="9ef50a75-3fb0-4a65-ae6a-45f981d5c979" containerName="registry-server" containerID="cri-o://526d21aa930fe3d9a935b02bf3298f63b99e6fd460f906b2cb8ff06a5cc1ed1a" gracePeriod=30 Dec 06 03:44:00 crc kubenswrapper[4802]: I1206 03:44:00.834448 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zzqgx"] Dec 06 03:44:00 crc kubenswrapper[4802]: I1206 03:44:00.834807 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zzqgx" podUID="03be8283-9ddf-4ecb-a1e7-579c322ca5d6" containerName="registry-server" containerID="cri-o://27f211beea61c680cd529158e01283a1070e9c6a850425273880787fd7750485" gracePeriod=30 Dec 06 03:44:00 crc kubenswrapper[4802]: I1206 03:44:00.839112 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ddpmx"] Dec 06 03:44:00 crc kubenswrapper[4802]: I1206 03:44:00.839313 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-ddpmx" podUID="ec27be85-6d9d-4d96-b3b8-b78a7a941acb" containerName="marketplace-operator" containerID="cri-o://bf69b7afe30f3da1ad1bf2efa2ab232935a02b768403437dfc405d6d91f019da" gracePeriod=30 Dec 06 03:44:00 crc kubenswrapper[4802]: I1206 03:44:00.847764 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5smmc"] Dec 06 03:44:00 crc kubenswrapper[4802]: I1206 03:44:00.848010 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5smmc" podUID="f3b86ac0-0a60-46f1-b82c-4bea67124a88" containerName="registry-server" containerID="cri-o://7d3b07a31ed9406630f598bfd3d5db7a2d6d651c5b05ccadb53cf46f1c71b97b" gracePeriod=30 Dec 06 03:44:00 crc kubenswrapper[4802]: I1206 03:44:00.863738 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wxqsx"] Dec 06 03:44:00 crc kubenswrapper[4802]: E1206 03:44:00.864011 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6eff099-6873-49c2-b0e9-e937700e1ff9" containerName="extract-utilities" Dec 06 03:44:00 crc kubenswrapper[4802]: I1206 03:44:00.864022 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6eff099-6873-49c2-b0e9-e937700e1ff9" containerName="extract-utilities" Dec 06 03:44:00 crc kubenswrapper[4802]: E1206 03:44:00.864036 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e6be98f-f5d2-4158-b4b8-eab47ad91564" containerName="extract-content" Dec 06 03:44:00 crc kubenswrapper[4802]: I1206 03:44:00.864042 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e6be98f-f5d2-4158-b4b8-eab47ad91564" containerName="extract-content" Dec 06 03:44:00 crc kubenswrapper[4802]: E1206 03:44:00.864051 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e6be98f-f5d2-4158-b4b8-eab47ad91564" containerName="registry-server" Dec 06 03:44:00 crc kubenswrapper[4802]: I1206 03:44:00.864059 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e6be98f-f5d2-4158-b4b8-eab47ad91564" containerName="registry-server" Dec 06 03:44:00 crc kubenswrapper[4802]: E1206 03:44:00.864068 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e6be98f-f5d2-4158-b4b8-eab47ad91564" containerName="extract-utilities" Dec 06 03:44:00 crc kubenswrapper[4802]: I1206 03:44:00.864073 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e6be98f-f5d2-4158-b4b8-eab47ad91564" containerName="extract-utilities" Dec 06 03:44:00 crc kubenswrapper[4802]: E1206 03:44:00.864086 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6eff099-6873-49c2-b0e9-e937700e1ff9" containerName="extract-content" Dec 06 03:44:00 crc kubenswrapper[4802]: I1206 03:44:00.864092 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6eff099-6873-49c2-b0e9-e937700e1ff9" containerName="extract-content" Dec 06 03:44:00 crc kubenswrapper[4802]: E1206 03:44:00.864101 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6eff099-6873-49c2-b0e9-e937700e1ff9" containerName="registry-server" Dec 06 03:44:00 crc kubenswrapper[4802]: I1206 03:44:00.864107 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6eff099-6873-49c2-b0e9-e937700e1ff9" containerName="registry-server" Dec 06 03:44:00 crc kubenswrapper[4802]: E1206 03:44:00.864114 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecfd6acc-8b75-4988-b832-4e63778ca053" containerName="pruner" Dec 06 03:44:00 crc kubenswrapper[4802]: I1206 03:44:00.864121 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecfd6acc-8b75-4988-b832-4e63778ca053" containerName="pruner" Dec 06 03:44:00 crc kubenswrapper[4802]: I1206 03:44:00.864212 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecfd6acc-8b75-4988-b832-4e63778ca053" containerName="pruner" Dec 06 03:44:00 crc kubenswrapper[4802]: I1206 03:44:00.864223 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6eff099-6873-49c2-b0e9-e937700e1ff9" containerName="registry-server" Dec 06 03:44:00 crc kubenswrapper[4802]: I1206 03:44:00.864235 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e6be98f-f5d2-4158-b4b8-eab47ad91564" containerName="registry-server" Dec 06 03:44:00 crc kubenswrapper[4802]: I1206 03:44:00.864591 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-wxqsx" Dec 06 03:44:00 crc kubenswrapper[4802]: I1206 03:44:00.868795 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-sxwkm"] Dec 06 03:44:00 crc kubenswrapper[4802]: I1206 03:44:00.869376 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-sxwkm" podUID="4a20041d-53aa-4179-b92e-9e7204c800ce" containerName="registry-server" containerID="cri-o://01f107ede9c9ccf901332436b6fbe30835290ff8a074f09d367bacbcbd39ee7a" gracePeriod=30 Dec 06 03:44:00 crc kubenswrapper[4802]: I1206 03:44:00.877792 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wxqsx"] Dec 06 03:44:00 crc kubenswrapper[4802]: I1206 03:44:00.880711 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-h7kc8"] Dec 06 03:44:00 crc kubenswrapper[4802]: I1206 03:44:00.881005 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-h7kc8" podUID="edb8213d-6a0c-4051-a230-4ec109d6e578" containerName="registry-server" containerID="cri-o://90422ef940487979c738c9408947ea9a9c23db7a8748cbfb02434a53708b6a53" gracePeriod=30 Dec 06 03:44:00 crc kubenswrapper[4802]: I1206 03:44:00.932834 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zzqgx" Dec 06 03:44:01 crc kubenswrapper[4802]: I1206 03:44:01.008650 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cffba732-050e-470a-9876-5954b2854ba5-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-wxqsx\" (UID: \"cffba732-050e-470a-9876-5954b2854ba5\") " pod="openshift-marketplace/marketplace-operator-79b997595-wxqsx" Dec 06 03:44:01 crc kubenswrapper[4802]: I1206 03:44:01.008703 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/cffba732-050e-470a-9876-5954b2854ba5-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-wxqsx\" (UID: \"cffba732-050e-470a-9876-5954b2854ba5\") " pod="openshift-marketplace/marketplace-operator-79b997595-wxqsx" Dec 06 03:44:01 crc kubenswrapper[4802]: I1206 03:44:01.008815 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47dp4\" (UniqueName: \"kubernetes.io/projected/cffba732-050e-470a-9876-5954b2854ba5-kube-api-access-47dp4\") pod \"marketplace-operator-79b997595-wxqsx\" (UID: \"cffba732-050e-470a-9876-5954b2854ba5\") " pod="openshift-marketplace/marketplace-operator-79b997595-wxqsx" Dec 06 03:44:01 crc kubenswrapper[4802]: E1206 03:44:01.102186 4802 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of aa53a5441403a493a64d80cb023e5f5513aef3f032db63372ccab0cf61f3c480 is running failed: container process not found" containerID="aa53a5441403a493a64d80cb023e5f5513aef3f032db63372ccab0cf61f3c480" cmd=["grpc_health_probe","-addr=:50051"] Dec 06 03:44:01 crc kubenswrapper[4802]: E1206 03:44:01.103086 4802 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of aa53a5441403a493a64d80cb023e5f5513aef3f032db63372ccab0cf61f3c480 is running failed: container process not found" containerID="aa53a5441403a493a64d80cb023e5f5513aef3f032db63372ccab0cf61f3c480" cmd=["grpc_health_probe","-addr=:50051"] Dec 06 03:44:01 crc kubenswrapper[4802]: E1206 03:44:01.103546 4802 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of aa53a5441403a493a64d80cb023e5f5513aef3f032db63372ccab0cf61f3c480 is running failed: container process not found" containerID="aa53a5441403a493a64d80cb023e5f5513aef3f032db63372ccab0cf61f3c480" cmd=["grpc_health_probe","-addr=:50051"] Dec 06 03:44:01 crc kubenswrapper[4802]: E1206 03:44:01.103582 4802 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of aa53a5441403a493a64d80cb023e5f5513aef3f032db63372ccab0cf61f3c480 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-z29bw" podUID="9b68d090-3526-4778-ad64-51da5e2191a4" containerName="registry-server" Dec 06 03:44:01 crc kubenswrapper[4802]: I1206 03:44:01.109687 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cffba732-050e-470a-9876-5954b2854ba5-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-wxqsx\" (UID: \"cffba732-050e-470a-9876-5954b2854ba5\") " pod="openshift-marketplace/marketplace-operator-79b997595-wxqsx" Dec 06 03:44:01 crc kubenswrapper[4802]: I1206 03:44:01.109780 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/cffba732-050e-470a-9876-5954b2854ba5-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-wxqsx\" (UID: \"cffba732-050e-470a-9876-5954b2854ba5\") " pod="openshift-marketplace/marketplace-operator-79b997595-wxqsx" Dec 06 03:44:01 crc kubenswrapper[4802]: I1206 03:44:01.109878 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47dp4\" (UniqueName: \"kubernetes.io/projected/cffba732-050e-470a-9876-5954b2854ba5-kube-api-access-47dp4\") pod \"marketplace-operator-79b997595-wxqsx\" (UID: \"cffba732-050e-470a-9876-5954b2854ba5\") " pod="openshift-marketplace/marketplace-operator-79b997595-wxqsx" Dec 06 03:44:01 crc kubenswrapper[4802]: I1206 03:44:01.111349 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cffba732-050e-470a-9876-5954b2854ba5-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-wxqsx\" (UID: \"cffba732-050e-470a-9876-5954b2854ba5\") " pod="openshift-marketplace/marketplace-operator-79b997595-wxqsx" Dec 06 03:44:01 crc kubenswrapper[4802]: I1206 03:44:01.120524 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/cffba732-050e-470a-9876-5954b2854ba5-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-wxqsx\" (UID: \"cffba732-050e-470a-9876-5954b2854ba5\") " pod="openshift-marketplace/marketplace-operator-79b997595-wxqsx" Dec 06 03:44:01 crc kubenswrapper[4802]: I1206 03:44:01.137380 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47dp4\" (UniqueName: \"kubernetes.io/projected/cffba732-050e-470a-9876-5954b2854ba5-kube-api-access-47dp4\") pod \"marketplace-operator-79b997595-wxqsx\" (UID: \"cffba732-050e-470a-9876-5954b2854ba5\") " pod="openshift-marketplace/marketplace-operator-79b997595-wxqsx" Dec 06 03:44:01 crc kubenswrapper[4802]: I1206 03:44:01.189357 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-wxqsx" Dec 06 03:44:01 crc kubenswrapper[4802]: I1206 03:44:01.252361 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-rqs9q"] Dec 06 03:44:01 crc kubenswrapper[4802]: I1206 03:44:01.296450 4802 generic.go:334] "Generic (PLEG): container finished" podID="ec27be85-6d9d-4d96-b3b8-b78a7a941acb" containerID="bf69b7afe30f3da1ad1bf2efa2ab232935a02b768403437dfc405d6d91f019da" exitCode=0 Dec 06 03:44:01 crc kubenswrapper[4802]: I1206 03:44:01.296856 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ddpmx" event={"ID":"ec27be85-6d9d-4d96-b3b8-b78a7a941acb","Type":"ContainerDied","Data":"bf69b7afe30f3da1ad1bf2efa2ab232935a02b768403437dfc405d6d91f019da"} Dec 06 03:44:01 crc kubenswrapper[4802]: I1206 03:44:01.299234 4802 generic.go:334] "Generic (PLEG): container finished" podID="4a20041d-53aa-4179-b92e-9e7204c800ce" containerID="01f107ede9c9ccf901332436b6fbe30835290ff8a074f09d367bacbcbd39ee7a" exitCode=0 Dec 06 03:44:01 crc kubenswrapper[4802]: I1206 03:44:01.299268 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sxwkm" event={"ID":"4a20041d-53aa-4179-b92e-9e7204c800ce","Type":"ContainerDied","Data":"01f107ede9c9ccf901332436b6fbe30835290ff8a074f09d367bacbcbd39ee7a"} Dec 06 03:44:01 crc kubenswrapper[4802]: I1206 03:44:01.301306 4802 generic.go:334] "Generic (PLEG): container finished" podID="9ef50a75-3fb0-4a65-ae6a-45f981d5c979" containerID="526d21aa930fe3d9a935b02bf3298f63b99e6fd460f906b2cb8ff06a5cc1ed1a" exitCode=0 Dec 06 03:44:01 crc kubenswrapper[4802]: I1206 03:44:01.301338 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hqh24" event={"ID":"9ef50a75-3fb0-4a65-ae6a-45f981d5c979","Type":"ContainerDied","Data":"526d21aa930fe3d9a935b02bf3298f63b99e6fd460f906b2cb8ff06a5cc1ed1a"} Dec 06 03:44:01 crc kubenswrapper[4802]: I1206 03:44:01.318320 4802 generic.go:334] "Generic (PLEG): container finished" podID="03be8283-9ddf-4ecb-a1e7-579c322ca5d6" containerID="27f211beea61c680cd529158e01283a1070e9c6a850425273880787fd7750485" exitCode=0 Dec 06 03:44:01 crc kubenswrapper[4802]: I1206 03:44:01.318507 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zzqgx" event={"ID":"03be8283-9ddf-4ecb-a1e7-579c322ca5d6","Type":"ContainerDied","Data":"27f211beea61c680cd529158e01283a1070e9c6a850425273880787fd7750485"} Dec 06 03:44:01 crc kubenswrapper[4802]: I1206 03:44:01.327893 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hqh24" Dec 06 03:44:01 crc kubenswrapper[4802]: I1206 03:44:01.349151 4802 generic.go:334] "Generic (PLEG): container finished" podID="edb8213d-6a0c-4051-a230-4ec109d6e578" containerID="90422ef940487979c738c9408947ea9a9c23db7a8748cbfb02434a53708b6a53" exitCode=0 Dec 06 03:44:01 crc kubenswrapper[4802]: I1206 03:44:01.349216 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h7kc8" event={"ID":"edb8213d-6a0c-4051-a230-4ec109d6e578","Type":"ContainerDied","Data":"90422ef940487979c738c9408947ea9a9c23db7a8748cbfb02434a53708b6a53"} Dec 06 03:44:01 crc kubenswrapper[4802]: I1206 03:44:01.364371 4802 generic.go:334] "Generic (PLEG): container finished" podID="f3b86ac0-0a60-46f1-b82c-4bea67124a88" containerID="7d3b07a31ed9406630f598bfd3d5db7a2d6d651c5b05ccadb53cf46f1c71b97b" exitCode=0 Dec 06 03:44:01 crc kubenswrapper[4802]: I1206 03:44:01.364461 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5smmc" event={"ID":"f3b86ac0-0a60-46f1-b82c-4bea67124a88","Type":"ContainerDied","Data":"7d3b07a31ed9406630f598bfd3d5db7a2d6d651c5b05ccadb53cf46f1c71b97b"} Dec 06 03:44:01 crc kubenswrapper[4802]: I1206 03:44:01.372887 4802 generic.go:334] "Generic (PLEG): container finished" podID="9b68d090-3526-4778-ad64-51da5e2191a4" containerID="aa53a5441403a493a64d80cb023e5f5513aef3f032db63372ccab0cf61f3c480" exitCode=0 Dec 06 03:44:01 crc kubenswrapper[4802]: I1206 03:44:01.372902 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z29bw" event={"ID":"9b68d090-3526-4778-ad64-51da5e2191a4","Type":"ContainerDied","Data":"aa53a5441403a493a64d80cb023e5f5513aef3f032db63372ccab0cf61f3c480"} Dec 06 03:44:01 crc kubenswrapper[4802]: I1206 03:44:01.389993 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sxwkm" Dec 06 03:44:01 crc kubenswrapper[4802]: I1206 03:44:01.429489 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-222z2\" (UniqueName: \"kubernetes.io/projected/4a20041d-53aa-4179-b92e-9e7204c800ce-kube-api-access-222z2\") pod \"4a20041d-53aa-4179-b92e-9e7204c800ce\" (UID: \"4a20041d-53aa-4179-b92e-9e7204c800ce\") " Dec 06 03:44:01 crc kubenswrapper[4802]: I1206 03:44:01.429536 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a20041d-53aa-4179-b92e-9e7204c800ce-catalog-content\") pod \"4a20041d-53aa-4179-b92e-9e7204c800ce\" (UID: \"4a20041d-53aa-4179-b92e-9e7204c800ce\") " Dec 06 03:44:01 crc kubenswrapper[4802]: I1206 03:44:01.429576 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a20041d-53aa-4179-b92e-9e7204c800ce-utilities\") pod \"4a20041d-53aa-4179-b92e-9e7204c800ce\" (UID: \"4a20041d-53aa-4179-b92e-9e7204c800ce\") " Dec 06 03:44:01 crc kubenswrapper[4802]: I1206 03:44:01.430389 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4a20041d-53aa-4179-b92e-9e7204c800ce-utilities" (OuterVolumeSpecName: "utilities") pod "4a20041d-53aa-4179-b92e-9e7204c800ce" (UID: "4a20041d-53aa-4179-b92e-9e7204c800ce"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:44:01 crc kubenswrapper[4802]: I1206 03:44:01.432606 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a20041d-53aa-4179-b92e-9e7204c800ce-kube-api-access-222z2" (OuterVolumeSpecName: "kube-api-access-222z2") pod "4a20041d-53aa-4179-b92e-9e7204c800ce" (UID: "4a20041d-53aa-4179-b92e-9e7204c800ce"). InnerVolumeSpecName "kube-api-access-222z2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:44:01 crc kubenswrapper[4802]: I1206 03:44:01.502912 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4a20041d-53aa-4179-b92e-9e7204c800ce-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4a20041d-53aa-4179-b92e-9e7204c800ce" (UID: "4a20041d-53aa-4179-b92e-9e7204c800ce"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:44:01 crc kubenswrapper[4802]: I1206 03:44:01.530947 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-222z2\" (UniqueName: \"kubernetes.io/projected/4a20041d-53aa-4179-b92e-9e7204c800ce-kube-api-access-222z2\") on node \"crc\" DevicePath \"\"" Dec 06 03:44:01 crc kubenswrapper[4802]: I1206 03:44:01.531031 4802 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a20041d-53aa-4179-b92e-9e7204c800ce-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 03:44:01 crc kubenswrapper[4802]: I1206 03:44:01.531043 4802 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a20041d-53aa-4179-b92e-9e7204c800ce-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 03:44:01 crc kubenswrapper[4802]: I1206 03:44:01.890921 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z29bw" Dec 06 03:44:01 crc kubenswrapper[4802]: I1206 03:44:01.897134 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5smmc" Dec 06 03:44:01 crc kubenswrapper[4802]: I1206 03:44:01.911992 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h7kc8" Dec 06 03:44:01 crc kubenswrapper[4802]: I1206 03:44:01.922449 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zzqgx" Dec 06 03:44:01 crc kubenswrapper[4802]: I1206 03:44:01.997614 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ddpmx" Dec 06 03:44:01 crc kubenswrapper[4802]: I1206 03:44:01.999871 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hqh24" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.035909 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wxqsx"] Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.039107 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wr5tq\" (UniqueName: \"kubernetes.io/projected/f3b86ac0-0a60-46f1-b82c-4bea67124a88-kube-api-access-wr5tq\") pod \"f3b86ac0-0a60-46f1-b82c-4bea67124a88\" (UID: \"f3b86ac0-0a60-46f1-b82c-4bea67124a88\") " Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.039182 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03be8283-9ddf-4ecb-a1e7-579c322ca5d6-utilities\") pod \"03be8283-9ddf-4ecb-a1e7-579c322ca5d6\" (UID: \"03be8283-9ddf-4ecb-a1e7-579c322ca5d6\") " Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.039217 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wv7n5\" (UniqueName: \"kubernetes.io/projected/03be8283-9ddf-4ecb-a1e7-579c322ca5d6-kube-api-access-wv7n5\") pod \"03be8283-9ddf-4ecb-a1e7-579c322ca5d6\" (UID: \"03be8283-9ddf-4ecb-a1e7-579c322ca5d6\") " Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.039241 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b68d090-3526-4778-ad64-51da5e2191a4-utilities\") pod \"9b68d090-3526-4778-ad64-51da5e2191a4\" (UID: \"9b68d090-3526-4778-ad64-51da5e2191a4\") " Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.039286 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b68d090-3526-4778-ad64-51da5e2191a4-catalog-content\") pod \"9b68d090-3526-4778-ad64-51da5e2191a4\" (UID: \"9b68d090-3526-4778-ad64-51da5e2191a4\") " Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.039324 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03be8283-9ddf-4ecb-a1e7-579c322ca5d6-catalog-content\") pod \"03be8283-9ddf-4ecb-a1e7-579c322ca5d6\" (UID: \"03be8283-9ddf-4ecb-a1e7-579c322ca5d6\") " Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.039367 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gqtns\" (UniqueName: \"kubernetes.io/projected/edb8213d-6a0c-4051-a230-4ec109d6e578-kube-api-access-gqtns\") pod \"edb8213d-6a0c-4051-a230-4ec109d6e578\" (UID: \"edb8213d-6a0c-4051-a230-4ec109d6e578\") " Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.039402 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3b86ac0-0a60-46f1-b82c-4bea67124a88-catalog-content\") pod \"f3b86ac0-0a60-46f1-b82c-4bea67124a88\" (UID: \"f3b86ac0-0a60-46f1-b82c-4bea67124a88\") " Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.039428 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t5n77\" (UniqueName: \"kubernetes.io/projected/9b68d090-3526-4778-ad64-51da5e2191a4-kube-api-access-t5n77\") pod \"9b68d090-3526-4778-ad64-51da5e2191a4\" (UID: \"9b68d090-3526-4778-ad64-51da5e2191a4\") " Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.039470 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3b86ac0-0a60-46f1-b82c-4bea67124a88-utilities\") pod \"f3b86ac0-0a60-46f1-b82c-4bea67124a88\" (UID: \"f3b86ac0-0a60-46f1-b82c-4bea67124a88\") " Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.039487 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/edb8213d-6a0c-4051-a230-4ec109d6e578-utilities\") pod \"edb8213d-6a0c-4051-a230-4ec109d6e578\" (UID: \"edb8213d-6a0c-4051-a230-4ec109d6e578\") " Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.039509 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/edb8213d-6a0c-4051-a230-4ec109d6e578-catalog-content\") pod \"edb8213d-6a0c-4051-a230-4ec109d6e578\" (UID: \"edb8213d-6a0c-4051-a230-4ec109d6e578\") " Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.041689 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b68d090-3526-4778-ad64-51da5e2191a4-utilities" (OuterVolumeSpecName: "utilities") pod "9b68d090-3526-4778-ad64-51da5e2191a4" (UID: "9b68d090-3526-4778-ad64-51da5e2191a4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.042105 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3b86ac0-0a60-46f1-b82c-4bea67124a88-utilities" (OuterVolumeSpecName: "utilities") pod "f3b86ac0-0a60-46f1-b82c-4bea67124a88" (UID: "f3b86ac0-0a60-46f1-b82c-4bea67124a88"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.042321 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/edb8213d-6a0c-4051-a230-4ec109d6e578-utilities" (OuterVolumeSpecName: "utilities") pod "edb8213d-6a0c-4051-a230-4ec109d6e578" (UID: "edb8213d-6a0c-4051-a230-4ec109d6e578"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.053069 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3b86ac0-0a60-46f1-b82c-4bea67124a88-kube-api-access-wr5tq" (OuterVolumeSpecName: "kube-api-access-wr5tq") pod "f3b86ac0-0a60-46f1-b82c-4bea67124a88" (UID: "f3b86ac0-0a60-46f1-b82c-4bea67124a88"). InnerVolumeSpecName "kube-api-access-wr5tq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.068562 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b68d090-3526-4778-ad64-51da5e2191a4-kube-api-access-t5n77" (OuterVolumeSpecName: "kube-api-access-t5n77") pod "9b68d090-3526-4778-ad64-51da5e2191a4" (UID: "9b68d090-3526-4778-ad64-51da5e2191a4"). InnerVolumeSpecName "kube-api-access-t5n77". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.068728 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/edb8213d-6a0c-4051-a230-4ec109d6e578-kube-api-access-gqtns" (OuterVolumeSpecName: "kube-api-access-gqtns") pod "edb8213d-6a0c-4051-a230-4ec109d6e578" (UID: "edb8213d-6a0c-4051-a230-4ec109d6e578"). InnerVolumeSpecName "kube-api-access-gqtns". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.072273 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03be8283-9ddf-4ecb-a1e7-579c322ca5d6-utilities" (OuterVolumeSpecName: "utilities") pod "03be8283-9ddf-4ecb-a1e7-579c322ca5d6" (UID: "03be8283-9ddf-4ecb-a1e7-579c322ca5d6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.076245 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03be8283-9ddf-4ecb-a1e7-579c322ca5d6-kube-api-access-wv7n5" (OuterVolumeSpecName: "kube-api-access-wv7n5") pod "03be8283-9ddf-4ecb-a1e7-579c322ca5d6" (UID: "03be8283-9ddf-4ecb-a1e7-579c322ca5d6"). InnerVolumeSpecName "kube-api-access-wv7n5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.077587 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3b86ac0-0a60-46f1-b82c-4bea67124a88-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f3b86ac0-0a60-46f1-b82c-4bea67124a88" (UID: "f3b86ac0-0a60-46f1-b82c-4bea67124a88"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.113137 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b68d090-3526-4778-ad64-51da5e2191a4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9b68d090-3526-4778-ad64-51da5e2191a4" (UID: "9b68d090-3526-4778-ad64-51da5e2191a4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.141098 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ec27be85-6d9d-4d96-b3b8-b78a7a941acb-marketplace-trusted-ca\") pod \"ec27be85-6d9d-4d96-b3b8-b78a7a941acb\" (UID: \"ec27be85-6d9d-4d96-b3b8-b78a7a941acb\") " Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.141180 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ef50a75-3fb0-4a65-ae6a-45f981d5c979-utilities\") pod \"9ef50a75-3fb0-4a65-ae6a-45f981d5c979\" (UID: \"9ef50a75-3fb0-4a65-ae6a-45f981d5c979\") " Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.141229 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ef50a75-3fb0-4a65-ae6a-45f981d5c979-catalog-content\") pod \"9ef50a75-3fb0-4a65-ae6a-45f981d5c979\" (UID: \"9ef50a75-3fb0-4a65-ae6a-45f981d5c979\") " Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.141276 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ec27be85-6d9d-4d96-b3b8-b78a7a941acb-marketplace-operator-metrics\") pod \"ec27be85-6d9d-4d96-b3b8-b78a7a941acb\" (UID: \"ec27be85-6d9d-4d96-b3b8-b78a7a941acb\") " Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.141313 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sstvn\" (UniqueName: \"kubernetes.io/projected/9ef50a75-3fb0-4a65-ae6a-45f981d5c979-kube-api-access-sstvn\") pod \"9ef50a75-3fb0-4a65-ae6a-45f981d5c979\" (UID: \"9ef50a75-3fb0-4a65-ae6a-45f981d5c979\") " Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.141394 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qc82p\" (UniqueName: \"kubernetes.io/projected/ec27be85-6d9d-4d96-b3b8-b78a7a941acb-kube-api-access-qc82p\") pod \"ec27be85-6d9d-4d96-b3b8-b78a7a941acb\" (UID: \"ec27be85-6d9d-4d96-b3b8-b78a7a941acb\") " Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.141418 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03be8283-9ddf-4ecb-a1e7-579c322ca5d6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "03be8283-9ddf-4ecb-a1e7-579c322ca5d6" (UID: "03be8283-9ddf-4ecb-a1e7-579c322ca5d6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.141638 4802 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03be8283-9ddf-4ecb-a1e7-579c322ca5d6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.141649 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gqtns\" (UniqueName: \"kubernetes.io/projected/edb8213d-6a0c-4051-a230-4ec109d6e578-kube-api-access-gqtns\") on node \"crc\" DevicePath \"\"" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.141659 4802 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3b86ac0-0a60-46f1-b82c-4bea67124a88-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.141668 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t5n77\" (UniqueName: \"kubernetes.io/projected/9b68d090-3526-4778-ad64-51da5e2191a4-kube-api-access-t5n77\") on node \"crc\" DevicePath \"\"" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.141677 4802 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3b86ac0-0a60-46f1-b82c-4bea67124a88-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.141702 4802 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/edb8213d-6a0c-4051-a230-4ec109d6e578-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.141711 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wr5tq\" (UniqueName: \"kubernetes.io/projected/f3b86ac0-0a60-46f1-b82c-4bea67124a88-kube-api-access-wr5tq\") on node \"crc\" DevicePath \"\"" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.141723 4802 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03be8283-9ddf-4ecb-a1e7-579c322ca5d6-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.141732 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wv7n5\" (UniqueName: \"kubernetes.io/projected/03be8283-9ddf-4ecb-a1e7-579c322ca5d6-kube-api-access-wv7n5\") on node \"crc\" DevicePath \"\"" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.141740 4802 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b68d090-3526-4778-ad64-51da5e2191a4-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.141781 4802 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b68d090-3526-4778-ad64-51da5e2191a4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.142550 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec27be85-6d9d-4d96-b3b8-b78a7a941acb-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "ec27be85-6d9d-4d96-b3b8-b78a7a941acb" (UID: "ec27be85-6d9d-4d96-b3b8-b78a7a941acb"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.143310 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ef50a75-3fb0-4a65-ae6a-45f981d5c979-utilities" (OuterVolumeSpecName: "utilities") pod "9ef50a75-3fb0-4a65-ae6a-45f981d5c979" (UID: "9ef50a75-3fb0-4a65-ae6a-45f981d5c979"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.145010 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec27be85-6d9d-4d96-b3b8-b78a7a941acb-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "ec27be85-6d9d-4d96-b3b8-b78a7a941acb" (UID: "ec27be85-6d9d-4d96-b3b8-b78a7a941acb"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.145618 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ef50a75-3fb0-4a65-ae6a-45f981d5c979-kube-api-access-sstvn" (OuterVolumeSpecName: "kube-api-access-sstvn") pod "9ef50a75-3fb0-4a65-ae6a-45f981d5c979" (UID: "9ef50a75-3fb0-4a65-ae6a-45f981d5c979"). InnerVolumeSpecName "kube-api-access-sstvn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.146336 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec27be85-6d9d-4d96-b3b8-b78a7a941acb-kube-api-access-qc82p" (OuterVolumeSpecName: "kube-api-access-qc82p") pod "ec27be85-6d9d-4d96-b3b8-b78a7a941acb" (UID: "ec27be85-6d9d-4d96-b3b8-b78a7a941acb"). InnerVolumeSpecName "kube-api-access-qc82p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.193518 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/edb8213d-6a0c-4051-a230-4ec109d6e578-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "edb8213d-6a0c-4051-a230-4ec109d6e578" (UID: "edb8213d-6a0c-4051-a230-4ec109d6e578"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.209729 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ef50a75-3fb0-4a65-ae6a-45f981d5c979-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9ef50a75-3fb0-4a65-ae6a-45f981d5c979" (UID: "9ef50a75-3fb0-4a65-ae6a-45f981d5c979"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.242700 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qc82p\" (UniqueName: \"kubernetes.io/projected/ec27be85-6d9d-4d96-b3b8-b78a7a941acb-kube-api-access-qc82p\") on node \"crc\" DevicePath \"\"" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.242769 4802 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ec27be85-6d9d-4d96-b3b8-b78a7a941acb-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.242780 4802 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ef50a75-3fb0-4a65-ae6a-45f981d5c979-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.242791 4802 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ef50a75-3fb0-4a65-ae6a-45f981d5c979-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.242799 4802 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ec27be85-6d9d-4d96-b3b8-b78a7a941acb-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.242807 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sstvn\" (UniqueName: \"kubernetes.io/projected/9ef50a75-3fb0-4a65-ae6a-45f981d5c979-kube-api-access-sstvn\") on node \"crc\" DevicePath \"\"" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.242816 4802 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/edb8213d-6a0c-4051-a230-4ec109d6e578-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.380113 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z29bw" event={"ID":"9b68d090-3526-4778-ad64-51da5e2191a4","Type":"ContainerDied","Data":"84396c83a7f184dacab7a87871a7a312d9938d7d51249d9ac392afdc1edf8da5"} Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.380167 4802 scope.go:117] "RemoveContainer" containerID="aa53a5441403a493a64d80cb023e5f5513aef3f032db63372ccab0cf61f3c480" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.380337 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z29bw" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.387212 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wxqsx" event={"ID":"cffba732-050e-470a-9876-5954b2854ba5","Type":"ContainerStarted","Data":"a9f56979c8496733ee3009335132f5f418b1a0604c223a174884b75660b9a3d5"} Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.387258 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wxqsx" event={"ID":"cffba732-050e-470a-9876-5954b2854ba5","Type":"ContainerStarted","Data":"6d18ca6e8c3123b61c77b78903d5ca9b0f00671f90ade65ffba2b82e76e41cbe"} Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.387956 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-wxqsx" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.396034 4802 scope.go:117] "RemoveContainer" containerID="eda8799283b13a2b7a49f02e60f9897e2b46fc5cd80385de5d493d3bafa10830" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.396110 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-wxqsx" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.397344 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sxwkm" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.397402 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sxwkm" event={"ID":"4a20041d-53aa-4179-b92e-9e7204c800ce","Type":"ContainerDied","Data":"1fe6950a93a2dd7bc463ee39b919bfdfd9b0431781316a96fe69e2867eb54bfa"} Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.401033 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ddpmx" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.401097 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ddpmx" event={"ID":"ec27be85-6d9d-4d96-b3b8-b78a7a941acb","Type":"ContainerDied","Data":"105affda22265051b61aca4aeaf6bc4c0dbb6cba1492e6b2f5c6a830948846bb"} Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.408424 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hqh24" event={"ID":"9ef50a75-3fb0-4a65-ae6a-45f981d5c979","Type":"ContainerDied","Data":"40a9127061e1d069851a74b67d1d9f7910ae3b9c533659f012828ec807360493"} Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.408527 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hqh24" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.416843 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-wxqsx" podStartSLOduration=2.416817072 podStartE2EDuration="2.416817072s" podCreationTimestamp="2025-12-06 03:44:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:44:02.413561453 +0000 UTC m=+235.285470615" watchObservedRunningTime="2025-12-06 03:44:02.416817072 +0000 UTC m=+235.288726234" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.419865 4802 scope.go:117] "RemoveContainer" containerID="397b292060e8e62c253746a3759e3dfa9f6e6b591b42c77fa0b30757636c00a7" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.420067 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h7kc8" event={"ID":"edb8213d-6a0c-4051-a230-4ec109d6e578","Type":"ContainerDied","Data":"6ebe08776ad723f0640f9ddbf39e0df2605443d52062262e5138be9988831d18"} Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.420140 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h7kc8" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.439279 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zzqgx" event={"ID":"03be8283-9ddf-4ecb-a1e7-579c322ca5d6","Type":"ContainerDied","Data":"c33bb4b3a3a9a725b1c15f3ef15e32e69d08dc350ddee82abc5c3c961451a763"} Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.439410 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zzqgx" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.444039 4802 scope.go:117] "RemoveContainer" containerID="01f107ede9c9ccf901332436b6fbe30835290ff8a074f09d367bacbcbd39ee7a" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.447508 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5smmc" event={"ID":"f3b86ac0-0a60-46f1-b82c-4bea67124a88","Type":"ContainerDied","Data":"7e224574fc248a6e6a948eabbf596ccfdea512b46261b84c98713eb75574aaee"} Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.447993 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5smmc" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.492648 4802 scope.go:117] "RemoveContainer" containerID="ee4f28d8dcd84e0f39cad21e47bf46a87f02d12b3516a7ffa1b02c0c1882a796" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.513516 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hqh24"] Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.516162 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hqh24"] Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.520902 4802 scope.go:117] "RemoveContainer" containerID="7f372faed4e069189425f6e681735ed7e7a60e44247259a3cb2291b9cd69a2eb" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.536037 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-sxwkm"] Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.536970 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-sxwkm"] Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.553675 4802 scope.go:117] "RemoveContainer" containerID="bf69b7afe30f3da1ad1bf2efa2ab232935a02b768403437dfc405d6d91f019da" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.559823 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-z29bw"] Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.564637 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-z29bw"] Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.569153 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-h7kc8"] Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.569531 4802 scope.go:117] "RemoveContainer" containerID="526d21aa930fe3d9a935b02bf3298f63b99e6fd460f906b2cb8ff06a5cc1ed1a" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.573183 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-h7kc8"] Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.575513 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zzqgx"] Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.577925 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zzqgx"] Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.589008 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5smmc"] Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.603103 4802 scope.go:117] "RemoveContainer" containerID="ffb979adb26fcbc6e9bd32c1fb2aca0c0a69106462aaedd90702d4fdbaf95047" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.603353 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5smmc"] Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.608549 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ddpmx"] Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.610947 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ddpmx"] Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.615928 4802 scope.go:117] "RemoveContainer" containerID="37ffe03bba19bb6171c5d04d99f9a3638b6e17e277bf6bf20a39cfe3baf7414d" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.627451 4802 scope.go:117] "RemoveContainer" containerID="90422ef940487979c738c9408947ea9a9c23db7a8748cbfb02434a53708b6a53" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.638532 4802 scope.go:117] "RemoveContainer" containerID="4493b583a803e12b2a5080c894164167f920ccc635ab9ad94afb11aecd4e0417" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.654865 4802 scope.go:117] "RemoveContainer" containerID="26b46463a973134ae8e6eed18bdbd6ce95260bb992ec56f67f8ab6877e1bf3f0" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.672545 4802 scope.go:117] "RemoveContainer" containerID="27f211beea61c680cd529158e01283a1070e9c6a850425273880787fd7750485" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.686620 4802 scope.go:117] "RemoveContainer" containerID="b5466e90518b9c52f5efd636a232d9dc07215fcbacbe49f3067e85ab55e8a114" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.700904 4802 scope.go:117] "RemoveContainer" containerID="9907ef4a9dc245ca170757969279ed86ad2bab18ef94fdcbd36bc3561c07b90d" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.711635 4802 scope.go:117] "RemoveContainer" containerID="7d3b07a31ed9406630f598bfd3d5db7a2d6d651c5b05ccadb53cf46f1c71b97b" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.723946 4802 scope.go:117] "RemoveContainer" containerID="7f766f943a6941f0810f824d83449e726b163fa01b79b4d68e64a07213508279" Dec 06 03:44:02 crc kubenswrapper[4802]: I1206 03:44:02.736845 4802 scope.go:117] "RemoveContainer" containerID="83129b5f5b11dadd2eb15c5a06f651ece3ab00cdf66ff0c69f51a4cde595e814" Dec 06 03:44:03 crc kubenswrapper[4802]: I1206 03:44:03.463677 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03be8283-9ddf-4ecb-a1e7-579c322ca5d6" path="/var/lib/kubelet/pods/03be8283-9ddf-4ecb-a1e7-579c322ca5d6/volumes" Dec 06 03:44:03 crc kubenswrapper[4802]: I1206 03:44:03.465065 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a20041d-53aa-4179-b92e-9e7204c800ce" path="/var/lib/kubelet/pods/4a20041d-53aa-4179-b92e-9e7204c800ce/volumes" Dec 06 03:44:03 crc kubenswrapper[4802]: I1206 03:44:03.466531 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b68d090-3526-4778-ad64-51da5e2191a4" path="/var/lib/kubelet/pods/9b68d090-3526-4778-ad64-51da5e2191a4/volumes" Dec 06 03:44:03 crc kubenswrapper[4802]: I1206 03:44:03.469047 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ef50a75-3fb0-4a65-ae6a-45f981d5c979" path="/var/lib/kubelet/pods/9ef50a75-3fb0-4a65-ae6a-45f981d5c979/volumes" Dec 06 03:44:03 crc kubenswrapper[4802]: I1206 03:44:03.470418 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec27be85-6d9d-4d96-b3b8-b78a7a941acb" path="/var/lib/kubelet/pods/ec27be85-6d9d-4d96-b3b8-b78a7a941acb/volumes" Dec 06 03:44:03 crc kubenswrapper[4802]: I1206 03:44:03.473091 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="edb8213d-6a0c-4051-a230-4ec109d6e578" path="/var/lib/kubelet/pods/edb8213d-6a0c-4051-a230-4ec109d6e578/volumes" Dec 06 03:44:03 crc kubenswrapper[4802]: I1206 03:44:03.474778 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3b86ac0-0a60-46f1-b82c-4bea67124a88" path="/var/lib/kubelet/pods/f3b86ac0-0a60-46f1-b82c-4bea67124a88/volumes" Dec 06 03:44:03 crc kubenswrapper[4802]: I1206 03:44:03.508934 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-skt5b"] Dec 06 03:44:03 crc kubenswrapper[4802]: E1206 03:44:03.509178 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03be8283-9ddf-4ecb-a1e7-579c322ca5d6" containerName="extract-content" Dec 06 03:44:03 crc kubenswrapper[4802]: I1206 03:44:03.509194 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="03be8283-9ddf-4ecb-a1e7-579c322ca5d6" containerName="extract-content" Dec 06 03:44:03 crc kubenswrapper[4802]: E1206 03:44:03.509210 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ef50a75-3fb0-4a65-ae6a-45f981d5c979" containerName="extract-content" Dec 06 03:44:03 crc kubenswrapper[4802]: I1206 03:44:03.509218 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ef50a75-3fb0-4a65-ae6a-45f981d5c979" containerName="extract-content" Dec 06 03:44:03 crc kubenswrapper[4802]: E1206 03:44:03.509229 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03be8283-9ddf-4ecb-a1e7-579c322ca5d6" containerName="registry-server" Dec 06 03:44:03 crc kubenswrapper[4802]: I1206 03:44:03.509236 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="03be8283-9ddf-4ecb-a1e7-579c322ca5d6" containerName="registry-server" Dec 06 03:44:03 crc kubenswrapper[4802]: E1206 03:44:03.509249 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b68d090-3526-4778-ad64-51da5e2191a4" containerName="extract-utilities" Dec 06 03:44:03 crc kubenswrapper[4802]: I1206 03:44:03.509257 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b68d090-3526-4778-ad64-51da5e2191a4" containerName="extract-utilities" Dec 06 03:44:03 crc kubenswrapper[4802]: E1206 03:44:03.509265 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a20041d-53aa-4179-b92e-9e7204c800ce" containerName="extract-utilities" Dec 06 03:44:03 crc kubenswrapper[4802]: I1206 03:44:03.509272 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a20041d-53aa-4179-b92e-9e7204c800ce" containerName="extract-utilities" Dec 06 03:44:03 crc kubenswrapper[4802]: E1206 03:44:03.509287 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03be8283-9ddf-4ecb-a1e7-579c322ca5d6" containerName="extract-utilities" Dec 06 03:44:03 crc kubenswrapper[4802]: I1206 03:44:03.509294 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="03be8283-9ddf-4ecb-a1e7-579c322ca5d6" containerName="extract-utilities" Dec 06 03:44:03 crc kubenswrapper[4802]: E1206 03:44:03.509303 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a20041d-53aa-4179-b92e-9e7204c800ce" containerName="registry-server" Dec 06 03:44:03 crc kubenswrapper[4802]: I1206 03:44:03.509310 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a20041d-53aa-4179-b92e-9e7204c800ce" containerName="registry-server" Dec 06 03:44:03 crc kubenswrapper[4802]: E1206 03:44:03.509319 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edb8213d-6a0c-4051-a230-4ec109d6e578" containerName="registry-server" Dec 06 03:44:03 crc kubenswrapper[4802]: I1206 03:44:03.509327 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="edb8213d-6a0c-4051-a230-4ec109d6e578" containerName="registry-server" Dec 06 03:44:03 crc kubenswrapper[4802]: E1206 03:44:03.509337 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edb8213d-6a0c-4051-a230-4ec109d6e578" containerName="extract-utilities" Dec 06 03:44:03 crc kubenswrapper[4802]: I1206 03:44:03.509345 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="edb8213d-6a0c-4051-a230-4ec109d6e578" containerName="extract-utilities" Dec 06 03:44:03 crc kubenswrapper[4802]: E1206 03:44:03.509357 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec27be85-6d9d-4d96-b3b8-b78a7a941acb" containerName="marketplace-operator" Dec 06 03:44:03 crc kubenswrapper[4802]: I1206 03:44:03.509364 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec27be85-6d9d-4d96-b3b8-b78a7a941acb" containerName="marketplace-operator" Dec 06 03:44:03 crc kubenswrapper[4802]: E1206 03:44:03.509372 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b68d090-3526-4778-ad64-51da5e2191a4" containerName="extract-content" Dec 06 03:44:03 crc kubenswrapper[4802]: I1206 03:44:03.509380 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b68d090-3526-4778-ad64-51da5e2191a4" containerName="extract-content" Dec 06 03:44:03 crc kubenswrapper[4802]: E1206 03:44:03.509395 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edb8213d-6a0c-4051-a230-4ec109d6e578" containerName="extract-content" Dec 06 03:44:03 crc kubenswrapper[4802]: I1206 03:44:03.509402 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="edb8213d-6a0c-4051-a230-4ec109d6e578" containerName="extract-content" Dec 06 03:44:03 crc kubenswrapper[4802]: E1206 03:44:03.509410 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a20041d-53aa-4179-b92e-9e7204c800ce" containerName="extract-content" Dec 06 03:44:03 crc kubenswrapper[4802]: I1206 03:44:03.509418 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a20041d-53aa-4179-b92e-9e7204c800ce" containerName="extract-content" Dec 06 03:44:03 crc kubenswrapper[4802]: E1206 03:44:03.509429 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3b86ac0-0a60-46f1-b82c-4bea67124a88" containerName="extract-utilities" Dec 06 03:44:03 crc kubenswrapper[4802]: I1206 03:44:03.509437 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3b86ac0-0a60-46f1-b82c-4bea67124a88" containerName="extract-utilities" Dec 06 03:44:03 crc kubenswrapper[4802]: E1206 03:44:03.509447 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b68d090-3526-4778-ad64-51da5e2191a4" containerName="registry-server" Dec 06 03:44:03 crc kubenswrapper[4802]: I1206 03:44:03.509455 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b68d090-3526-4778-ad64-51da5e2191a4" containerName="registry-server" Dec 06 03:44:03 crc kubenswrapper[4802]: E1206 03:44:03.509465 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3b86ac0-0a60-46f1-b82c-4bea67124a88" containerName="extract-content" Dec 06 03:44:03 crc kubenswrapper[4802]: I1206 03:44:03.509504 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3b86ac0-0a60-46f1-b82c-4bea67124a88" containerName="extract-content" Dec 06 03:44:03 crc kubenswrapper[4802]: E1206 03:44:03.509515 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ef50a75-3fb0-4a65-ae6a-45f981d5c979" containerName="extract-utilities" Dec 06 03:44:03 crc kubenswrapper[4802]: I1206 03:44:03.509524 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ef50a75-3fb0-4a65-ae6a-45f981d5c979" containerName="extract-utilities" Dec 06 03:44:03 crc kubenswrapper[4802]: E1206 03:44:03.509535 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3b86ac0-0a60-46f1-b82c-4bea67124a88" containerName="registry-server" Dec 06 03:44:03 crc kubenswrapper[4802]: I1206 03:44:03.509543 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3b86ac0-0a60-46f1-b82c-4bea67124a88" containerName="registry-server" Dec 06 03:44:03 crc kubenswrapper[4802]: E1206 03:44:03.509552 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ef50a75-3fb0-4a65-ae6a-45f981d5c979" containerName="registry-server" Dec 06 03:44:03 crc kubenswrapper[4802]: I1206 03:44:03.509560 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ef50a75-3fb0-4a65-ae6a-45f981d5c979" containerName="registry-server" Dec 06 03:44:03 crc kubenswrapper[4802]: I1206 03:44:03.509726 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b68d090-3526-4778-ad64-51da5e2191a4" containerName="registry-server" Dec 06 03:44:03 crc kubenswrapper[4802]: I1206 03:44:03.509742 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec27be85-6d9d-4d96-b3b8-b78a7a941acb" containerName="marketplace-operator" Dec 06 03:44:03 crc kubenswrapper[4802]: I1206 03:44:03.509800 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ef50a75-3fb0-4a65-ae6a-45f981d5c979" containerName="registry-server" Dec 06 03:44:03 crc kubenswrapper[4802]: I1206 03:44:03.509812 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="03be8283-9ddf-4ecb-a1e7-579c322ca5d6" containerName="registry-server" Dec 06 03:44:03 crc kubenswrapper[4802]: I1206 03:44:03.509820 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3b86ac0-0a60-46f1-b82c-4bea67124a88" containerName="registry-server" Dec 06 03:44:03 crc kubenswrapper[4802]: I1206 03:44:03.509835 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="edb8213d-6a0c-4051-a230-4ec109d6e578" containerName="registry-server" Dec 06 03:44:03 crc kubenswrapper[4802]: I1206 03:44:03.509847 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a20041d-53aa-4179-b92e-9e7204c800ce" containerName="registry-server" Dec 06 03:44:03 crc kubenswrapper[4802]: I1206 03:44:03.511087 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-skt5b" Dec 06 03:44:03 crc kubenswrapper[4802]: I1206 03:44:03.517676 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 06 03:44:03 crc kubenswrapper[4802]: I1206 03:44:03.519146 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-skt5b"] Dec 06 03:44:03 crc kubenswrapper[4802]: I1206 03:44:03.669647 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/042f382d-836e-4cc2-9065-352a210f6593-catalog-content\") pod \"certified-operators-skt5b\" (UID: \"042f382d-836e-4cc2-9065-352a210f6593\") " pod="openshift-marketplace/certified-operators-skt5b" Dec 06 03:44:03 crc kubenswrapper[4802]: I1206 03:44:03.669723 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/042f382d-836e-4cc2-9065-352a210f6593-utilities\") pod \"certified-operators-skt5b\" (UID: \"042f382d-836e-4cc2-9065-352a210f6593\") " pod="openshift-marketplace/certified-operators-skt5b" Dec 06 03:44:03 crc kubenswrapper[4802]: I1206 03:44:03.669794 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gz52b\" (UniqueName: \"kubernetes.io/projected/042f382d-836e-4cc2-9065-352a210f6593-kube-api-access-gz52b\") pod \"certified-operators-skt5b\" (UID: \"042f382d-836e-4cc2-9065-352a210f6593\") " pod="openshift-marketplace/certified-operators-skt5b" Dec 06 03:44:03 crc kubenswrapper[4802]: I1206 03:44:03.771572 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/042f382d-836e-4cc2-9065-352a210f6593-utilities\") pod \"certified-operators-skt5b\" (UID: \"042f382d-836e-4cc2-9065-352a210f6593\") " pod="openshift-marketplace/certified-operators-skt5b" Dec 06 03:44:03 crc kubenswrapper[4802]: I1206 03:44:03.772167 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/042f382d-836e-4cc2-9065-352a210f6593-utilities\") pod \"certified-operators-skt5b\" (UID: \"042f382d-836e-4cc2-9065-352a210f6593\") " pod="openshift-marketplace/certified-operators-skt5b" Dec 06 03:44:03 crc kubenswrapper[4802]: I1206 03:44:03.772352 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gz52b\" (UniqueName: \"kubernetes.io/projected/042f382d-836e-4cc2-9065-352a210f6593-kube-api-access-gz52b\") pod \"certified-operators-skt5b\" (UID: \"042f382d-836e-4cc2-9065-352a210f6593\") " pod="openshift-marketplace/certified-operators-skt5b" Dec 06 03:44:03 crc kubenswrapper[4802]: I1206 03:44:03.772563 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/042f382d-836e-4cc2-9065-352a210f6593-catalog-content\") pod \"certified-operators-skt5b\" (UID: \"042f382d-836e-4cc2-9065-352a210f6593\") " pod="openshift-marketplace/certified-operators-skt5b" Dec 06 03:44:03 crc kubenswrapper[4802]: I1206 03:44:03.773243 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/042f382d-836e-4cc2-9065-352a210f6593-catalog-content\") pod \"certified-operators-skt5b\" (UID: \"042f382d-836e-4cc2-9065-352a210f6593\") " pod="openshift-marketplace/certified-operators-skt5b" Dec 06 03:44:03 crc kubenswrapper[4802]: I1206 03:44:03.795338 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gz52b\" (UniqueName: \"kubernetes.io/projected/042f382d-836e-4cc2-9065-352a210f6593-kube-api-access-gz52b\") pod \"certified-operators-skt5b\" (UID: \"042f382d-836e-4cc2-9065-352a210f6593\") " pod="openshift-marketplace/certified-operators-skt5b" Dec 06 03:44:03 crc kubenswrapper[4802]: I1206 03:44:03.837651 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-skt5b" Dec 06 03:44:04 crc kubenswrapper[4802]: I1206 03:44:04.080695 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-skt5b"] Dec 06 03:44:04 crc kubenswrapper[4802]: I1206 03:44:04.106456 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rnrvp"] Dec 06 03:44:04 crc kubenswrapper[4802]: I1206 03:44:04.110326 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rnrvp" Dec 06 03:44:04 crc kubenswrapper[4802]: I1206 03:44:04.112343 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rnrvp"] Dec 06 03:44:04 crc kubenswrapper[4802]: I1206 03:44:04.114562 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 06 03:44:04 crc kubenswrapper[4802]: I1206 03:44:04.286517 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cae8cfbe-5644-45a5-92e6-0ce169cc47c5-utilities\") pod \"community-operators-rnrvp\" (UID: \"cae8cfbe-5644-45a5-92e6-0ce169cc47c5\") " pod="openshift-marketplace/community-operators-rnrvp" Dec 06 03:44:04 crc kubenswrapper[4802]: I1206 03:44:04.286690 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cae8cfbe-5644-45a5-92e6-0ce169cc47c5-catalog-content\") pod \"community-operators-rnrvp\" (UID: \"cae8cfbe-5644-45a5-92e6-0ce169cc47c5\") " pod="openshift-marketplace/community-operators-rnrvp" Dec 06 03:44:04 crc kubenswrapper[4802]: I1206 03:44:04.286730 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jw4j5\" (UniqueName: \"kubernetes.io/projected/cae8cfbe-5644-45a5-92e6-0ce169cc47c5-kube-api-access-jw4j5\") pod \"community-operators-rnrvp\" (UID: \"cae8cfbe-5644-45a5-92e6-0ce169cc47c5\") " pod="openshift-marketplace/community-operators-rnrvp" Dec 06 03:44:04 crc kubenswrapper[4802]: I1206 03:44:04.388958 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cae8cfbe-5644-45a5-92e6-0ce169cc47c5-catalog-content\") pod \"community-operators-rnrvp\" (UID: \"cae8cfbe-5644-45a5-92e6-0ce169cc47c5\") " pod="openshift-marketplace/community-operators-rnrvp" Dec 06 03:44:04 crc kubenswrapper[4802]: I1206 03:44:04.389055 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cae8cfbe-5644-45a5-92e6-0ce169cc47c5-catalog-content\") pod \"community-operators-rnrvp\" (UID: \"cae8cfbe-5644-45a5-92e6-0ce169cc47c5\") " pod="openshift-marketplace/community-operators-rnrvp" Dec 06 03:44:04 crc kubenswrapper[4802]: I1206 03:44:04.389145 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jw4j5\" (UniqueName: \"kubernetes.io/projected/cae8cfbe-5644-45a5-92e6-0ce169cc47c5-kube-api-access-jw4j5\") pod \"community-operators-rnrvp\" (UID: \"cae8cfbe-5644-45a5-92e6-0ce169cc47c5\") " pod="openshift-marketplace/community-operators-rnrvp" Dec 06 03:44:04 crc kubenswrapper[4802]: I1206 03:44:04.389201 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cae8cfbe-5644-45a5-92e6-0ce169cc47c5-utilities\") pod \"community-operators-rnrvp\" (UID: \"cae8cfbe-5644-45a5-92e6-0ce169cc47c5\") " pod="openshift-marketplace/community-operators-rnrvp" Dec 06 03:44:04 crc kubenswrapper[4802]: I1206 03:44:04.390290 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cae8cfbe-5644-45a5-92e6-0ce169cc47c5-utilities\") pod \"community-operators-rnrvp\" (UID: \"cae8cfbe-5644-45a5-92e6-0ce169cc47c5\") " pod="openshift-marketplace/community-operators-rnrvp" Dec 06 03:44:04 crc kubenswrapper[4802]: I1206 03:44:04.410544 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jw4j5\" (UniqueName: \"kubernetes.io/projected/cae8cfbe-5644-45a5-92e6-0ce169cc47c5-kube-api-access-jw4j5\") pod \"community-operators-rnrvp\" (UID: \"cae8cfbe-5644-45a5-92e6-0ce169cc47c5\") " pod="openshift-marketplace/community-operators-rnrvp" Dec 06 03:44:04 crc kubenswrapper[4802]: I1206 03:44:04.449637 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rnrvp" Dec 06 03:44:04 crc kubenswrapper[4802]: I1206 03:44:04.481381 4802 generic.go:334] "Generic (PLEG): container finished" podID="042f382d-836e-4cc2-9065-352a210f6593" containerID="8a55a7c7e29d61c89075613a7489c35146c61b48192c5beb1955f656acc79426" exitCode=0 Dec 06 03:44:04 crc kubenswrapper[4802]: I1206 03:44:04.482723 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-skt5b" event={"ID":"042f382d-836e-4cc2-9065-352a210f6593","Type":"ContainerDied","Data":"8a55a7c7e29d61c89075613a7489c35146c61b48192c5beb1955f656acc79426"} Dec 06 03:44:04 crc kubenswrapper[4802]: I1206 03:44:04.482773 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-skt5b" event={"ID":"042f382d-836e-4cc2-9065-352a210f6593","Type":"ContainerStarted","Data":"30d09c0325b032232dc7e8edaa0e8569cb70588c24111cc6a7c691dc596ef454"} Dec 06 03:44:04 crc kubenswrapper[4802]: I1206 03:44:04.861982 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rnrvp"] Dec 06 03:44:05 crc kubenswrapper[4802]: I1206 03:44:05.488550 4802 generic.go:334] "Generic (PLEG): container finished" podID="042f382d-836e-4cc2-9065-352a210f6593" containerID="0bd681dd43cd83bf375ebfb0268a3ed5ac6075e064c5377685ee66b444b0241f" exitCode=0 Dec 06 03:44:05 crc kubenswrapper[4802]: I1206 03:44:05.488592 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-skt5b" event={"ID":"042f382d-836e-4cc2-9065-352a210f6593","Type":"ContainerDied","Data":"0bd681dd43cd83bf375ebfb0268a3ed5ac6075e064c5377685ee66b444b0241f"} Dec 06 03:44:05 crc kubenswrapper[4802]: I1206 03:44:05.493210 4802 generic.go:334] "Generic (PLEG): container finished" podID="cae8cfbe-5644-45a5-92e6-0ce169cc47c5" containerID="90c9762845bdd20613718ae313a30735be1adc9963502b7d4988ade94a737e83" exitCode=0 Dec 06 03:44:05 crc kubenswrapper[4802]: I1206 03:44:05.493285 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rnrvp" event={"ID":"cae8cfbe-5644-45a5-92e6-0ce169cc47c5","Type":"ContainerDied","Data":"90c9762845bdd20613718ae313a30735be1adc9963502b7d4988ade94a737e83"} Dec 06 03:44:05 crc kubenswrapper[4802]: I1206 03:44:05.493335 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rnrvp" event={"ID":"cae8cfbe-5644-45a5-92e6-0ce169cc47c5","Type":"ContainerStarted","Data":"9fe1f6c0760dbf6607adb6efc9c1e6f044b45ec930ca38d3ee4fe4014b65b956"} Dec 06 03:44:05 crc kubenswrapper[4802]: I1206 03:44:05.912371 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2l4qq"] Dec 06 03:44:05 crc kubenswrapper[4802]: I1206 03:44:05.913609 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2l4qq" Dec 06 03:44:05 crc kubenswrapper[4802]: I1206 03:44:05.917636 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 06 03:44:05 crc kubenswrapper[4802]: I1206 03:44:05.917809 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2l4qq"] Dec 06 03:44:06 crc kubenswrapper[4802]: I1206 03:44:06.012398 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/199a835e-3d62-42ef-8df0-6f087ae5586e-catalog-content\") pod \"redhat-marketplace-2l4qq\" (UID: \"199a835e-3d62-42ef-8df0-6f087ae5586e\") " pod="openshift-marketplace/redhat-marketplace-2l4qq" Dec 06 03:44:06 crc kubenswrapper[4802]: I1206 03:44:06.012496 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/199a835e-3d62-42ef-8df0-6f087ae5586e-utilities\") pod \"redhat-marketplace-2l4qq\" (UID: \"199a835e-3d62-42ef-8df0-6f087ae5586e\") " pod="openshift-marketplace/redhat-marketplace-2l4qq" Dec 06 03:44:06 crc kubenswrapper[4802]: I1206 03:44:06.012532 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btqp4\" (UniqueName: \"kubernetes.io/projected/199a835e-3d62-42ef-8df0-6f087ae5586e-kube-api-access-btqp4\") pod \"redhat-marketplace-2l4qq\" (UID: \"199a835e-3d62-42ef-8df0-6f087ae5586e\") " pod="openshift-marketplace/redhat-marketplace-2l4qq" Dec 06 03:44:06 crc kubenswrapper[4802]: I1206 03:44:06.113303 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/199a835e-3d62-42ef-8df0-6f087ae5586e-catalog-content\") pod \"redhat-marketplace-2l4qq\" (UID: \"199a835e-3d62-42ef-8df0-6f087ae5586e\") " pod="openshift-marketplace/redhat-marketplace-2l4qq" Dec 06 03:44:06 crc kubenswrapper[4802]: I1206 03:44:06.113374 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/199a835e-3d62-42ef-8df0-6f087ae5586e-utilities\") pod \"redhat-marketplace-2l4qq\" (UID: \"199a835e-3d62-42ef-8df0-6f087ae5586e\") " pod="openshift-marketplace/redhat-marketplace-2l4qq" Dec 06 03:44:06 crc kubenswrapper[4802]: I1206 03:44:06.113404 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btqp4\" (UniqueName: \"kubernetes.io/projected/199a835e-3d62-42ef-8df0-6f087ae5586e-kube-api-access-btqp4\") pod \"redhat-marketplace-2l4qq\" (UID: \"199a835e-3d62-42ef-8df0-6f087ae5586e\") " pod="openshift-marketplace/redhat-marketplace-2l4qq" Dec 06 03:44:06 crc kubenswrapper[4802]: I1206 03:44:06.113883 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/199a835e-3d62-42ef-8df0-6f087ae5586e-catalog-content\") pod \"redhat-marketplace-2l4qq\" (UID: \"199a835e-3d62-42ef-8df0-6f087ae5586e\") " pod="openshift-marketplace/redhat-marketplace-2l4qq" Dec 06 03:44:06 crc kubenswrapper[4802]: I1206 03:44:06.113906 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/199a835e-3d62-42ef-8df0-6f087ae5586e-utilities\") pod \"redhat-marketplace-2l4qq\" (UID: \"199a835e-3d62-42ef-8df0-6f087ae5586e\") " pod="openshift-marketplace/redhat-marketplace-2l4qq" Dec 06 03:44:06 crc kubenswrapper[4802]: I1206 03:44:06.131156 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-btqp4\" (UniqueName: \"kubernetes.io/projected/199a835e-3d62-42ef-8df0-6f087ae5586e-kube-api-access-btqp4\") pod \"redhat-marketplace-2l4qq\" (UID: \"199a835e-3d62-42ef-8df0-6f087ae5586e\") " pod="openshift-marketplace/redhat-marketplace-2l4qq" Dec 06 03:44:06 crc kubenswrapper[4802]: I1206 03:44:06.228946 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2l4qq" Dec 06 03:44:06 crc kubenswrapper[4802]: I1206 03:44:06.500082 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-ldbdv"] Dec 06 03:44:06 crc kubenswrapper[4802]: I1206 03:44:06.501184 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ldbdv" Dec 06 03:44:06 crc kubenswrapper[4802]: I1206 03:44:06.503170 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 06 03:44:06 crc kubenswrapper[4802]: I1206 03:44:06.507586 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ldbdv"] Dec 06 03:44:06 crc kubenswrapper[4802]: I1206 03:44:06.508123 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-skt5b" event={"ID":"042f382d-836e-4cc2-9065-352a210f6593","Type":"ContainerStarted","Data":"54db7931473e9ad2df8a8f0c7a7906422b696fdb88ec606a0cb2564c032cc828"} Dec 06 03:44:06 crc kubenswrapper[4802]: I1206 03:44:06.510390 4802 generic.go:334] "Generic (PLEG): container finished" podID="cae8cfbe-5644-45a5-92e6-0ce169cc47c5" containerID="2c7b2ff8cbc9d6a795d262db2692bb16b4242ce9ed3b505ec6f8a15833dfc4f2" exitCode=0 Dec 06 03:44:06 crc kubenswrapper[4802]: I1206 03:44:06.510418 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rnrvp" event={"ID":"cae8cfbe-5644-45a5-92e6-0ce169cc47c5","Type":"ContainerDied","Data":"2c7b2ff8cbc9d6a795d262db2692bb16b4242ce9ed3b505ec6f8a15833dfc4f2"} Dec 06 03:44:06 crc kubenswrapper[4802]: I1206 03:44:06.549945 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-skt5b" podStartSLOduration=2.122249172 podStartE2EDuration="3.549928616s" podCreationTimestamp="2025-12-06 03:44:03 +0000 UTC" firstStartedPulling="2025-12-06 03:44:04.483309867 +0000 UTC m=+237.355219019" lastFinishedPulling="2025-12-06 03:44:05.910989291 +0000 UTC m=+238.782898463" observedRunningTime="2025-12-06 03:44:06.548487676 +0000 UTC m=+239.420396838" watchObservedRunningTime="2025-12-06 03:44:06.549928616 +0000 UTC m=+239.421837768" Dec 06 03:44:06 crc kubenswrapper[4802]: I1206 03:44:06.623181 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrc97\" (UniqueName: \"kubernetes.io/projected/a640d4db-76d9-4c21-b394-cbc18a3fe2c4-kube-api-access-qrc97\") pod \"redhat-operators-ldbdv\" (UID: \"a640d4db-76d9-4c21-b394-cbc18a3fe2c4\") " pod="openshift-marketplace/redhat-operators-ldbdv" Dec 06 03:44:06 crc kubenswrapper[4802]: I1206 03:44:06.623549 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a640d4db-76d9-4c21-b394-cbc18a3fe2c4-utilities\") pod \"redhat-operators-ldbdv\" (UID: \"a640d4db-76d9-4c21-b394-cbc18a3fe2c4\") " pod="openshift-marketplace/redhat-operators-ldbdv" Dec 06 03:44:06 crc kubenswrapper[4802]: I1206 03:44:06.623599 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a640d4db-76d9-4c21-b394-cbc18a3fe2c4-catalog-content\") pod \"redhat-operators-ldbdv\" (UID: \"a640d4db-76d9-4c21-b394-cbc18a3fe2c4\") " pod="openshift-marketplace/redhat-operators-ldbdv" Dec 06 03:44:06 crc kubenswrapper[4802]: I1206 03:44:06.639607 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2l4qq"] Dec 06 03:44:06 crc kubenswrapper[4802]: W1206 03:44:06.654791 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod199a835e_3d62_42ef_8df0_6f087ae5586e.slice/crio-f162aa5f074077f628e5b116838c0b51d354211c276f8a416efd30a3a7419673 WatchSource:0}: Error finding container f162aa5f074077f628e5b116838c0b51d354211c276f8a416efd30a3a7419673: Status 404 returned error can't find the container with id f162aa5f074077f628e5b116838c0b51d354211c276f8a416efd30a3a7419673 Dec 06 03:44:06 crc kubenswrapper[4802]: I1206 03:44:06.724646 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a640d4db-76d9-4c21-b394-cbc18a3fe2c4-utilities\") pod \"redhat-operators-ldbdv\" (UID: \"a640d4db-76d9-4c21-b394-cbc18a3fe2c4\") " pod="openshift-marketplace/redhat-operators-ldbdv" Dec 06 03:44:06 crc kubenswrapper[4802]: I1206 03:44:06.724692 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a640d4db-76d9-4c21-b394-cbc18a3fe2c4-catalog-content\") pod \"redhat-operators-ldbdv\" (UID: \"a640d4db-76d9-4c21-b394-cbc18a3fe2c4\") " pod="openshift-marketplace/redhat-operators-ldbdv" Dec 06 03:44:06 crc kubenswrapper[4802]: I1206 03:44:06.724735 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrc97\" (UniqueName: \"kubernetes.io/projected/a640d4db-76d9-4c21-b394-cbc18a3fe2c4-kube-api-access-qrc97\") pod \"redhat-operators-ldbdv\" (UID: \"a640d4db-76d9-4c21-b394-cbc18a3fe2c4\") " pod="openshift-marketplace/redhat-operators-ldbdv" Dec 06 03:44:06 crc kubenswrapper[4802]: I1206 03:44:06.725083 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a640d4db-76d9-4c21-b394-cbc18a3fe2c4-utilities\") pod \"redhat-operators-ldbdv\" (UID: \"a640d4db-76d9-4c21-b394-cbc18a3fe2c4\") " pod="openshift-marketplace/redhat-operators-ldbdv" Dec 06 03:44:06 crc kubenswrapper[4802]: I1206 03:44:06.725173 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a640d4db-76d9-4c21-b394-cbc18a3fe2c4-catalog-content\") pod \"redhat-operators-ldbdv\" (UID: \"a640d4db-76d9-4c21-b394-cbc18a3fe2c4\") " pod="openshift-marketplace/redhat-operators-ldbdv" Dec 06 03:44:06 crc kubenswrapper[4802]: I1206 03:44:06.752564 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrc97\" (UniqueName: \"kubernetes.io/projected/a640d4db-76d9-4c21-b394-cbc18a3fe2c4-kube-api-access-qrc97\") pod \"redhat-operators-ldbdv\" (UID: \"a640d4db-76d9-4c21-b394-cbc18a3fe2c4\") " pod="openshift-marketplace/redhat-operators-ldbdv" Dec 06 03:44:06 crc kubenswrapper[4802]: I1206 03:44:06.864004 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ldbdv" Dec 06 03:44:07 crc kubenswrapper[4802]: I1206 03:44:07.053460 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ldbdv"] Dec 06 03:44:07 crc kubenswrapper[4802]: W1206 03:44:07.064895 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda640d4db_76d9_4c21_b394_cbc18a3fe2c4.slice/crio-0046394ea416dbffb87675067479ae7a71ed31d0fa5c08e0d534dc707ed107c0 WatchSource:0}: Error finding container 0046394ea416dbffb87675067479ae7a71ed31d0fa5c08e0d534dc707ed107c0: Status 404 returned error can't find the container with id 0046394ea416dbffb87675067479ae7a71ed31d0fa5c08e0d534dc707ed107c0 Dec 06 03:44:07 crc kubenswrapper[4802]: I1206 03:44:07.517479 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rnrvp" event={"ID":"cae8cfbe-5644-45a5-92e6-0ce169cc47c5","Type":"ContainerStarted","Data":"cd4190ee007529b3c9b605f3e886660f56f5e1e666dbf62ee2b642da8845299f"} Dec 06 03:44:07 crc kubenswrapper[4802]: I1206 03:44:07.519007 4802 generic.go:334] "Generic (PLEG): container finished" podID="a640d4db-76d9-4c21-b394-cbc18a3fe2c4" containerID="04c57c7690d50458cf86cf2a6e9d4db48c5089c3ef1458902be41e39f966df2a" exitCode=0 Dec 06 03:44:07 crc kubenswrapper[4802]: I1206 03:44:07.519057 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ldbdv" event={"ID":"a640d4db-76d9-4c21-b394-cbc18a3fe2c4","Type":"ContainerDied","Data":"04c57c7690d50458cf86cf2a6e9d4db48c5089c3ef1458902be41e39f966df2a"} Dec 06 03:44:07 crc kubenswrapper[4802]: I1206 03:44:07.519072 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ldbdv" event={"ID":"a640d4db-76d9-4c21-b394-cbc18a3fe2c4","Type":"ContainerStarted","Data":"0046394ea416dbffb87675067479ae7a71ed31d0fa5c08e0d534dc707ed107c0"} Dec 06 03:44:07 crc kubenswrapper[4802]: I1206 03:44:07.540137 4802 generic.go:334] "Generic (PLEG): container finished" podID="199a835e-3d62-42ef-8df0-6f087ae5586e" containerID="f5a085aefc7994dbb6e0610f13231eef5a75ed5bc07023b4149db440f5fa9477" exitCode=0 Dec 06 03:44:07 crc kubenswrapper[4802]: I1206 03:44:07.540175 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2l4qq" event={"ID":"199a835e-3d62-42ef-8df0-6f087ae5586e","Type":"ContainerDied","Data":"f5a085aefc7994dbb6e0610f13231eef5a75ed5bc07023b4149db440f5fa9477"} Dec 06 03:44:07 crc kubenswrapper[4802]: I1206 03:44:07.540221 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2l4qq" event={"ID":"199a835e-3d62-42ef-8df0-6f087ae5586e","Type":"ContainerStarted","Data":"f162aa5f074077f628e5b116838c0b51d354211c276f8a416efd30a3a7419673"} Dec 06 03:44:07 crc kubenswrapper[4802]: I1206 03:44:07.545342 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rnrvp" podStartSLOduration=2.030864184 podStartE2EDuration="3.545320129s" podCreationTimestamp="2025-12-06 03:44:04 +0000 UTC" firstStartedPulling="2025-12-06 03:44:05.495211441 +0000 UTC m=+238.367120623" lastFinishedPulling="2025-12-06 03:44:07.009667406 +0000 UTC m=+239.881576568" observedRunningTime="2025-12-06 03:44:07.540391044 +0000 UTC m=+240.412300206" watchObservedRunningTime="2025-12-06 03:44:07.545320129 +0000 UTC m=+240.417229281" Dec 06 03:44:07 crc kubenswrapper[4802]: I1206 03:44:07.777858 4802 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 06 03:44:07 crc kubenswrapper[4802]: I1206 03:44:07.778904 4802 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 06 03:44:07 crc kubenswrapper[4802]: I1206 03:44:07.779082 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 03:44:07 crc kubenswrapper[4802]: I1206 03:44:07.779182 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://a895e0cc04880c3287ba725ac5484835235f1e066a94ce98fcf4b6133c5cbd2f" gracePeriod=15 Dec 06 03:44:07 crc kubenswrapper[4802]: I1206 03:44:07.779203 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://31752b84d86bda30792b29e4c6ed963753cc576c5de784395e8e6dcbebaced47" gracePeriod=15 Dec 06 03:44:07 crc kubenswrapper[4802]: I1206 03:44:07.779224 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://9a21a97385f856ec996105b597d0ae81c56ed28b625ef3f2905e1af6ec8f6a45" gracePeriod=15 Dec 06 03:44:07 crc kubenswrapper[4802]: I1206 03:44:07.779304 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://646e01a58908ba2e703aae5eef53e1b981b0f0c3d1d3202487f2a8b94ee3f55d" gracePeriod=15 Dec 06 03:44:07 crc kubenswrapper[4802]: I1206 03:44:07.779351 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://16a92d501c9e626fe91e4a6793e426d10eb292cfbab766db8b8cc72e5dae94ef" gracePeriod=15 Dec 06 03:44:07 crc kubenswrapper[4802]: I1206 03:44:07.779733 4802 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 06 03:44:07 crc kubenswrapper[4802]: E1206 03:44:07.779910 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 06 03:44:07 crc kubenswrapper[4802]: I1206 03:44:07.779927 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 06 03:44:07 crc kubenswrapper[4802]: E1206 03:44:07.779933 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 06 03:44:07 crc kubenswrapper[4802]: I1206 03:44:07.779940 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 06 03:44:07 crc kubenswrapper[4802]: E1206 03:44:07.779952 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 06 03:44:07 crc kubenswrapper[4802]: I1206 03:44:07.779958 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 06 03:44:07 crc kubenswrapper[4802]: E1206 03:44:07.779966 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 06 03:44:07 crc kubenswrapper[4802]: I1206 03:44:07.779973 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 06 03:44:07 crc kubenswrapper[4802]: E1206 03:44:07.779982 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 06 03:44:07 crc kubenswrapper[4802]: I1206 03:44:07.779987 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 06 03:44:07 crc kubenswrapper[4802]: E1206 03:44:07.779995 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 06 03:44:07 crc kubenswrapper[4802]: I1206 03:44:07.780001 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 06 03:44:07 crc kubenswrapper[4802]: E1206 03:44:07.780014 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 06 03:44:07 crc kubenswrapper[4802]: I1206 03:44:07.780021 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 06 03:44:07 crc kubenswrapper[4802]: E1206 03:44:07.780032 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 06 03:44:07 crc kubenswrapper[4802]: I1206 03:44:07.780039 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 06 03:44:07 crc kubenswrapper[4802]: I1206 03:44:07.780120 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 06 03:44:07 crc kubenswrapper[4802]: I1206 03:44:07.780135 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 06 03:44:07 crc kubenswrapper[4802]: I1206 03:44:07.780141 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 06 03:44:07 crc kubenswrapper[4802]: I1206 03:44:07.780148 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 06 03:44:07 crc kubenswrapper[4802]: I1206 03:44:07.780155 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 06 03:44:07 crc kubenswrapper[4802]: I1206 03:44:07.780162 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 06 03:44:07 crc kubenswrapper[4802]: I1206 03:44:07.780168 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 06 03:44:07 crc kubenswrapper[4802]: I1206 03:44:07.816782 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 06 03:44:07 crc kubenswrapper[4802]: I1206 03:44:07.844799 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 03:44:07 crc kubenswrapper[4802]: I1206 03:44:07.844897 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 03:44:07 crc kubenswrapper[4802]: I1206 03:44:07.844957 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 03:44:07 crc kubenswrapper[4802]: I1206 03:44:07.844991 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 03:44:07 crc kubenswrapper[4802]: I1206 03:44:07.845076 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 03:44:07 crc kubenswrapper[4802]: I1206 03:44:07.946590 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 03:44:07 crc kubenswrapper[4802]: I1206 03:44:07.946661 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 03:44:07 crc kubenswrapper[4802]: I1206 03:44:07.946688 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 03:44:07 crc kubenswrapper[4802]: I1206 03:44:07.946708 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:44:07 crc kubenswrapper[4802]: I1206 03:44:07.946728 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 03:44:07 crc kubenswrapper[4802]: I1206 03:44:07.946743 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 03:44:07 crc kubenswrapper[4802]: I1206 03:44:07.946812 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 03:44:07 crc kubenswrapper[4802]: I1206 03:44:07.946832 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 03:44:07 crc kubenswrapper[4802]: I1206 03:44:07.946891 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 03:44:07 crc kubenswrapper[4802]: I1206 03:44:07.946907 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 03:44:07 crc kubenswrapper[4802]: I1206 03:44:07.946924 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 03:44:07 crc kubenswrapper[4802]: I1206 03:44:07.946838 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:44:07 crc kubenswrapper[4802]: I1206 03:44:07.946966 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:44:08 crc kubenswrapper[4802]: I1206 03:44:08.048100 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:44:08 crc kubenswrapper[4802]: I1206 03:44:08.048291 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:44:08 crc kubenswrapper[4802]: I1206 03:44:08.048371 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:44:08 crc kubenswrapper[4802]: I1206 03:44:08.048444 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:44:08 crc kubenswrapper[4802]: I1206 03:44:08.048224 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:44:08 crc kubenswrapper[4802]: I1206 03:44:08.048498 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:44:08 crc kubenswrapper[4802]: E1206 03:44:08.069775 4802 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/events\": dial tcp 38.102.83.150:6443: connect: connection refused" event="&Event{ObjectMeta:{redhat-marketplace-2l4qq.187e83764b55d197 openshift-marketplace 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-marketplace,Name:redhat-marketplace-2l4qq,UID:199a835e-3d62-42ef-8df0-6f087ae5586e,APIVersion:v1,ResourceVersion:29462,FieldPath:spec.initContainers{extract-content},},Reason:Pulled,Message:Successfully pulled image \"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\" in 524ms (524ms including waiting). Image size: 1141987142 bytes.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-06 03:44:08.069108119 +0000 UTC m=+240.941017271,LastTimestamp:2025-12-06 03:44:08.069108119 +0000 UTC m=+240.941017271,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 06 03:44:08 crc kubenswrapper[4802]: I1206 03:44:08.115160 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 03:44:08 crc kubenswrapper[4802]: W1206 03:44:08.142689 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-35de19e67dfb5420fc048f2583f19392984b68e19cf14a8e138e7b8f5de8c2f6 WatchSource:0}: Error finding container 35de19e67dfb5420fc048f2583f19392984b68e19cf14a8e138e7b8f5de8c2f6: Status 404 returned error can't find the container with id 35de19e67dfb5420fc048f2583f19392984b68e19cf14a8e138e7b8f5de8c2f6 Dec 06 03:44:08 crc kubenswrapper[4802]: I1206 03:44:08.546467 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"ab648d5b2f1d8c98db0240df2881ef08dd4a7fac79eab166365e8f4c66285e3d"} Dec 06 03:44:08 crc kubenswrapper[4802]: I1206 03:44:08.546787 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"35de19e67dfb5420fc048f2583f19392984b68e19cf14a8e138e7b8f5de8c2f6"} Dec 06 03:44:08 crc kubenswrapper[4802]: I1206 03:44:08.548062 4802 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:08 crc kubenswrapper[4802]: I1206 03:44:08.548226 4802 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:08 crc kubenswrapper[4802]: I1206 03:44:08.549992 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ldbdv" event={"ID":"a640d4db-76d9-4c21-b394-cbc18a3fe2c4","Type":"ContainerStarted","Data":"675a5f831f1f94bd4e3fc3e716556df3abbda0eff08e5b126792668458e03a8f"} Dec 06 03:44:08 crc kubenswrapper[4802]: I1206 03:44:08.550548 4802 status_manager.go:851] "Failed to get status for pod" podUID="a640d4db-76d9-4c21-b394-cbc18a3fe2c4" pod="openshift-marketplace/redhat-operators-ldbdv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-ldbdv\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:08 crc kubenswrapper[4802]: I1206 03:44:08.550713 4802 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:08 crc kubenswrapper[4802]: I1206 03:44:08.550901 4802 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:08 crc kubenswrapper[4802]: I1206 03:44:08.552666 4802 generic.go:334] "Generic (PLEG): container finished" podID="199a835e-3d62-42ef-8df0-6f087ae5586e" containerID="c1d08a71e165e5c02ef6f1173b5f1d40359f7b4081f36afc9c5299b9f547443a" exitCode=0 Dec 06 03:44:08 crc kubenswrapper[4802]: I1206 03:44:08.552707 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2l4qq" event={"ID":"199a835e-3d62-42ef-8df0-6f087ae5586e","Type":"ContainerDied","Data":"c1d08a71e165e5c02ef6f1173b5f1d40359f7b4081f36afc9c5299b9f547443a"} Dec 06 03:44:08 crc kubenswrapper[4802]: I1206 03:44:08.553287 4802 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:08 crc kubenswrapper[4802]: I1206 03:44:08.553504 4802 status_manager.go:851] "Failed to get status for pod" podUID="199a835e-3d62-42ef-8df0-6f087ae5586e" pod="openshift-marketplace/redhat-marketplace-2l4qq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-2l4qq\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:08 crc kubenswrapper[4802]: I1206 03:44:08.554028 4802 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:08 crc kubenswrapper[4802]: I1206 03:44:08.554275 4802 status_manager.go:851] "Failed to get status for pod" podUID="a640d4db-76d9-4c21-b394-cbc18a3fe2c4" pod="openshift-marketplace/redhat-operators-ldbdv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-ldbdv\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:08 crc kubenswrapper[4802]: I1206 03:44:08.554768 4802 generic.go:334] "Generic (PLEG): container finished" podID="b58e10a8-7605-4863-bc97-91714add5409" containerID="af3ed61c6bef9dc4e55515ee327d8928fa08e228412d55611fd95b430c653929" exitCode=0 Dec 06 03:44:08 crc kubenswrapper[4802]: I1206 03:44:08.556269 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"b58e10a8-7605-4863-bc97-91714add5409","Type":"ContainerDied","Data":"af3ed61c6bef9dc4e55515ee327d8928fa08e228412d55611fd95b430c653929"} Dec 06 03:44:08 crc kubenswrapper[4802]: I1206 03:44:08.556981 4802 status_manager.go:851] "Failed to get status for pod" podUID="b58e10a8-7605-4863-bc97-91714add5409" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:08 crc kubenswrapper[4802]: I1206 03:44:08.557393 4802 status_manager.go:851] "Failed to get status for pod" podUID="199a835e-3d62-42ef-8df0-6f087ae5586e" pod="openshift-marketplace/redhat-marketplace-2l4qq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-2l4qq\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:08 crc kubenswrapper[4802]: I1206 03:44:08.557637 4802 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:08 crc kubenswrapper[4802]: I1206 03:44:08.558428 4802 status_manager.go:851] "Failed to get status for pod" podUID="a640d4db-76d9-4c21-b394-cbc18a3fe2c4" pod="openshift-marketplace/redhat-operators-ldbdv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-ldbdv\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:08 crc kubenswrapper[4802]: I1206 03:44:08.558691 4802 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:08 crc kubenswrapper[4802]: I1206 03:44:08.562854 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Dec 06 03:44:08 crc kubenswrapper[4802]: I1206 03:44:08.564307 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 06 03:44:08 crc kubenswrapper[4802]: I1206 03:44:08.564959 4802 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="31752b84d86bda30792b29e4c6ed963753cc576c5de784395e8e6dcbebaced47" exitCode=0 Dec 06 03:44:08 crc kubenswrapper[4802]: I1206 03:44:08.564987 4802 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="16a92d501c9e626fe91e4a6793e426d10eb292cfbab766db8b8cc72e5dae94ef" exitCode=0 Dec 06 03:44:08 crc kubenswrapper[4802]: I1206 03:44:08.564997 4802 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="9a21a97385f856ec996105b597d0ae81c56ed28b625ef3f2905e1af6ec8f6a45" exitCode=0 Dec 06 03:44:08 crc kubenswrapper[4802]: I1206 03:44:08.565006 4802 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="646e01a58908ba2e703aae5eef53e1b981b0f0c3d1d3202487f2a8b94ee3f55d" exitCode=2 Dec 06 03:44:08 crc kubenswrapper[4802]: I1206 03:44:08.565784 4802 scope.go:117] "RemoveContainer" containerID="2c08c3dcdf3d84355887c4855a13c5656dbcec0107852dbe0694ca970716fbbc" Dec 06 03:44:09 crc kubenswrapper[4802]: I1206 03:44:09.574211 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2l4qq" event={"ID":"199a835e-3d62-42ef-8df0-6f087ae5586e","Type":"ContainerStarted","Data":"5c82a43e5a4e4a5de3f334dd7af5c54247899931e2c55907feda1ac6a1c5b80e"} Dec 06 03:44:09 crc kubenswrapper[4802]: I1206 03:44:09.575127 4802 status_manager.go:851] "Failed to get status for pod" podUID="b58e10a8-7605-4863-bc97-91714add5409" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:09 crc kubenswrapper[4802]: I1206 03:44:09.575556 4802 status_manager.go:851] "Failed to get status for pod" podUID="199a835e-3d62-42ef-8df0-6f087ae5586e" pod="openshift-marketplace/redhat-marketplace-2l4qq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-2l4qq\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:09 crc kubenswrapper[4802]: I1206 03:44:09.576001 4802 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:09 crc kubenswrapper[4802]: I1206 03:44:09.576302 4802 status_manager.go:851] "Failed to get status for pod" podUID="a640d4db-76d9-4c21-b394-cbc18a3fe2c4" pod="openshift-marketplace/redhat-operators-ldbdv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-ldbdv\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:09 crc kubenswrapper[4802]: I1206 03:44:09.579876 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 06 03:44:09 crc kubenswrapper[4802]: I1206 03:44:09.583299 4802 generic.go:334] "Generic (PLEG): container finished" podID="a640d4db-76d9-4c21-b394-cbc18a3fe2c4" containerID="675a5f831f1f94bd4e3fc3e716556df3abbda0eff08e5b126792668458e03a8f" exitCode=0 Dec 06 03:44:09 crc kubenswrapper[4802]: I1206 03:44:09.583340 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ldbdv" event={"ID":"a640d4db-76d9-4c21-b394-cbc18a3fe2c4","Type":"ContainerDied","Data":"675a5f831f1f94bd4e3fc3e716556df3abbda0eff08e5b126792668458e03a8f"} Dec 06 03:44:09 crc kubenswrapper[4802]: I1206 03:44:09.583959 4802 status_manager.go:851] "Failed to get status for pod" podUID="b58e10a8-7605-4863-bc97-91714add5409" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:09 crc kubenswrapper[4802]: I1206 03:44:09.584285 4802 status_manager.go:851] "Failed to get status for pod" podUID="199a835e-3d62-42ef-8df0-6f087ae5586e" pod="openshift-marketplace/redhat-marketplace-2l4qq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-2l4qq\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:09 crc kubenswrapper[4802]: I1206 03:44:09.584696 4802 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:09 crc kubenswrapper[4802]: I1206 03:44:09.585077 4802 status_manager.go:851] "Failed to get status for pod" podUID="a640d4db-76d9-4c21-b394-cbc18a3fe2c4" pod="openshift-marketplace/redhat-operators-ldbdv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-ldbdv\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:10 crc kubenswrapper[4802]: I1206 03:44:10.595316 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"b58e10a8-7605-4863-bc97-91714add5409","Type":"ContainerDied","Data":"96a2340d0cb485091cf317bb1624e1303513182f8ee577fdf1ad9284815390f5"} Dec 06 03:44:10 crc kubenswrapper[4802]: I1206 03:44:10.595995 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="96a2340d0cb485091cf317bb1624e1303513182f8ee577fdf1ad9284815390f5" Dec 06 03:44:10 crc kubenswrapper[4802]: I1206 03:44:10.604385 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 06 03:44:10 crc kubenswrapper[4802]: I1206 03:44:10.605285 4802 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="a895e0cc04880c3287ba725ac5484835235f1e066a94ce98fcf4b6133c5cbd2f" exitCode=0 Dec 06 03:44:10 crc kubenswrapper[4802]: I1206 03:44:10.606306 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="85038e49ae134abd6863213a301321f9cfdc6ee76482529ff81ae788154bbfd0" Dec 06 03:44:11 crc kubenswrapper[4802]: I1206 03:44:11.033474 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 06 03:44:11 crc kubenswrapper[4802]: I1206 03:44:11.034071 4802 status_manager.go:851] "Failed to get status for pod" podUID="b58e10a8-7605-4863-bc97-91714add5409" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:11 crc kubenswrapper[4802]: I1206 03:44:11.034418 4802 status_manager.go:851] "Failed to get status for pod" podUID="199a835e-3d62-42ef-8df0-6f087ae5586e" pod="openshift-marketplace/redhat-marketplace-2l4qq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-2l4qq\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:11 crc kubenswrapper[4802]: I1206 03:44:11.035309 4802 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:11 crc kubenswrapper[4802]: I1206 03:44:11.035612 4802 status_manager.go:851] "Failed to get status for pod" podUID="a640d4db-76d9-4c21-b394-cbc18a3fe2c4" pod="openshift-marketplace/redhat-operators-ldbdv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-ldbdv\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:11 crc kubenswrapper[4802]: I1206 03:44:11.044542 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 06 03:44:11 crc kubenswrapper[4802]: I1206 03:44:11.045463 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:44:11 crc kubenswrapper[4802]: I1206 03:44:11.046263 4802 status_manager.go:851] "Failed to get status for pod" podUID="a640d4db-76d9-4c21-b394-cbc18a3fe2c4" pod="openshift-marketplace/redhat-operators-ldbdv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-ldbdv\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:11 crc kubenswrapper[4802]: I1206 03:44:11.046879 4802 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:11 crc kubenswrapper[4802]: I1206 03:44:11.047228 4802 status_manager.go:851] "Failed to get status for pod" podUID="b58e10a8-7605-4863-bc97-91714add5409" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:11 crc kubenswrapper[4802]: I1206 03:44:11.047535 4802 status_manager.go:851] "Failed to get status for pod" podUID="199a835e-3d62-42ef-8df0-6f087ae5586e" pod="openshift-marketplace/redhat-marketplace-2l4qq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-2l4qq\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:11 crc kubenswrapper[4802]: I1206 03:44:11.047720 4802 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:11 crc kubenswrapper[4802]: I1206 03:44:11.208925 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 06 03:44:11 crc kubenswrapper[4802]: I1206 03:44:11.209232 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b58e10a8-7605-4863-bc97-91714add5409-var-lock\") pod \"b58e10a8-7605-4863-bc97-91714add5409\" (UID: \"b58e10a8-7605-4863-bc97-91714add5409\") " Dec 06 03:44:11 crc kubenswrapper[4802]: I1206 03:44:11.209421 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b58e10a8-7605-4863-bc97-91714add5409-kubelet-dir\") pod \"b58e10a8-7605-4863-bc97-91714add5409\" (UID: \"b58e10a8-7605-4863-bc97-91714add5409\") " Dec 06 03:44:11 crc kubenswrapper[4802]: I1206 03:44:11.209594 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 06 03:44:11 crc kubenswrapper[4802]: I1206 03:44:11.209834 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 06 03:44:11 crc kubenswrapper[4802]: I1206 03:44:11.210005 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b58e10a8-7605-4863-bc97-91714add5409-kube-api-access\") pod \"b58e10a8-7605-4863-bc97-91714add5409\" (UID: \"b58e10a8-7605-4863-bc97-91714add5409\") " Dec 06 03:44:11 crc kubenswrapper[4802]: I1206 03:44:11.209645 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b58e10a8-7605-4863-bc97-91714add5409-var-lock" (OuterVolumeSpecName: "var-lock") pod "b58e10a8-7605-4863-bc97-91714add5409" (UID: "b58e10a8-7605-4863-bc97-91714add5409"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:44:11 crc kubenswrapper[4802]: I1206 03:44:11.209677 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:44:11 crc kubenswrapper[4802]: I1206 03:44:11.209699 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b58e10a8-7605-4863-bc97-91714add5409-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "b58e10a8-7605-4863-bc97-91714add5409" (UID: "b58e10a8-7605-4863-bc97-91714add5409"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:44:11 crc kubenswrapper[4802]: I1206 03:44:11.209773 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:44:11 crc kubenswrapper[4802]: I1206 03:44:11.210120 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:44:11 crc kubenswrapper[4802]: I1206 03:44:11.217174 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b58e10a8-7605-4863-bc97-91714add5409-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "b58e10a8-7605-4863-bc97-91714add5409" (UID: "b58e10a8-7605-4863-bc97-91714add5409"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:44:11 crc kubenswrapper[4802]: I1206 03:44:11.311918 4802 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b58e10a8-7605-4863-bc97-91714add5409-var-lock\") on node \"crc\" DevicePath \"\"" Dec 06 03:44:11 crc kubenswrapper[4802]: I1206 03:44:11.311957 4802 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b58e10a8-7605-4863-bc97-91714add5409-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 06 03:44:11 crc kubenswrapper[4802]: I1206 03:44:11.311966 4802 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 06 03:44:11 crc kubenswrapper[4802]: I1206 03:44:11.311974 4802 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 06 03:44:11 crc kubenswrapper[4802]: I1206 03:44:11.311989 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b58e10a8-7605-4863-bc97-91714add5409-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 03:44:11 crc kubenswrapper[4802]: I1206 03:44:11.311998 4802 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 06 03:44:11 crc kubenswrapper[4802]: E1206 03:44:11.421214 4802 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:11 crc kubenswrapper[4802]: E1206 03:44:11.422232 4802 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:11 crc kubenswrapper[4802]: E1206 03:44:11.422798 4802 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:11 crc kubenswrapper[4802]: E1206 03:44:11.423306 4802 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:11 crc kubenswrapper[4802]: E1206 03:44:11.423811 4802 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:11 crc kubenswrapper[4802]: I1206 03:44:11.423878 4802 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 06 03:44:11 crc kubenswrapper[4802]: E1206 03:44:11.424332 4802 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.150:6443: connect: connection refused" interval="200ms" Dec 06 03:44:11 crc kubenswrapper[4802]: I1206 03:44:11.456866 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 06 03:44:11 crc kubenswrapper[4802]: I1206 03:44:11.613796 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ldbdv" event={"ID":"a640d4db-76d9-4c21-b394-cbc18a3fe2c4","Type":"ContainerStarted","Data":"76e5cb3e7fc0ca6922ce58d4479d2fd465360f2006864ffa86f3dba2696ac0a6"} Dec 06 03:44:11 crc kubenswrapper[4802]: I1206 03:44:11.613846 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 06 03:44:11 crc kubenswrapper[4802]: I1206 03:44:11.613903 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:44:11 crc kubenswrapper[4802]: I1206 03:44:11.614710 4802 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:11 crc kubenswrapper[4802]: I1206 03:44:11.615146 4802 status_manager.go:851] "Failed to get status for pod" podUID="b58e10a8-7605-4863-bc97-91714add5409" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:11 crc kubenswrapper[4802]: I1206 03:44:11.615976 4802 status_manager.go:851] "Failed to get status for pod" podUID="199a835e-3d62-42ef-8df0-6f087ae5586e" pod="openshift-marketplace/redhat-marketplace-2l4qq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-2l4qq\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:11 crc kubenswrapper[4802]: I1206 03:44:11.616216 4802 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:11 crc kubenswrapper[4802]: I1206 03:44:11.616486 4802 status_manager.go:851] "Failed to get status for pod" podUID="a640d4db-76d9-4c21-b394-cbc18a3fe2c4" pod="openshift-marketplace/redhat-operators-ldbdv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-ldbdv\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:11 crc kubenswrapper[4802]: I1206 03:44:11.617006 4802 status_manager.go:851] "Failed to get status for pod" podUID="a640d4db-76d9-4c21-b394-cbc18a3fe2c4" pod="openshift-marketplace/redhat-operators-ldbdv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-ldbdv\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:11 crc kubenswrapper[4802]: I1206 03:44:11.617413 4802 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:11 crc kubenswrapper[4802]: I1206 03:44:11.617908 4802 status_manager.go:851] "Failed to get status for pod" podUID="b58e10a8-7605-4863-bc97-91714add5409" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:11 crc kubenswrapper[4802]: I1206 03:44:11.618165 4802 status_manager.go:851] "Failed to get status for pod" podUID="199a835e-3d62-42ef-8df0-6f087ae5586e" pod="openshift-marketplace/redhat-marketplace-2l4qq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-2l4qq\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:11 crc kubenswrapper[4802]: I1206 03:44:11.618412 4802 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:11 crc kubenswrapper[4802]: I1206 03:44:11.618738 4802 status_manager.go:851] "Failed to get status for pod" podUID="a640d4db-76d9-4c21-b394-cbc18a3fe2c4" pod="openshift-marketplace/redhat-operators-ldbdv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-ldbdv\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:11 crc kubenswrapper[4802]: I1206 03:44:11.619033 4802 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:11 crc kubenswrapper[4802]: I1206 03:44:11.619461 4802 status_manager.go:851] "Failed to get status for pod" podUID="b58e10a8-7605-4863-bc97-91714add5409" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:11 crc kubenswrapper[4802]: I1206 03:44:11.619738 4802 status_manager.go:851] "Failed to get status for pod" podUID="199a835e-3d62-42ef-8df0-6f087ae5586e" pod="openshift-marketplace/redhat-marketplace-2l4qq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-2l4qq\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:11 crc kubenswrapper[4802]: I1206 03:44:11.620132 4802 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:11 crc kubenswrapper[4802]: E1206 03:44:11.625546 4802 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.150:6443: connect: connection refused" interval="400ms" Dec 06 03:44:12 crc kubenswrapper[4802]: E1206 03:44:12.027087 4802 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.150:6443: connect: connection refused" interval="800ms" Dec 06 03:44:12 crc kubenswrapper[4802]: E1206 03:44:12.828317 4802 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.150:6443: connect: connection refused" interval="1.6s" Dec 06 03:44:13 crc kubenswrapper[4802]: I1206 03:44:13.838566 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-skt5b" Dec 06 03:44:13 crc kubenswrapper[4802]: I1206 03:44:13.838887 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-skt5b" Dec 06 03:44:13 crc kubenswrapper[4802]: I1206 03:44:13.881787 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-skt5b" Dec 06 03:44:13 crc kubenswrapper[4802]: I1206 03:44:13.882417 4802 status_manager.go:851] "Failed to get status for pod" podUID="042f382d-836e-4cc2-9065-352a210f6593" pod="openshift-marketplace/certified-operators-skt5b" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-skt5b\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:13 crc kubenswrapper[4802]: I1206 03:44:13.882916 4802 status_manager.go:851] "Failed to get status for pod" podUID="b58e10a8-7605-4863-bc97-91714add5409" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:13 crc kubenswrapper[4802]: I1206 03:44:13.883402 4802 status_manager.go:851] "Failed to get status for pod" podUID="199a835e-3d62-42ef-8df0-6f087ae5586e" pod="openshift-marketplace/redhat-marketplace-2l4qq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-2l4qq\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:13 crc kubenswrapper[4802]: I1206 03:44:13.883715 4802 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:13 crc kubenswrapper[4802]: I1206 03:44:13.884026 4802 status_manager.go:851] "Failed to get status for pod" podUID="a640d4db-76d9-4c21-b394-cbc18a3fe2c4" pod="openshift-marketplace/redhat-operators-ldbdv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-ldbdv\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:14 crc kubenswrapper[4802]: E1206 03:44:14.428874 4802 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.150:6443: connect: connection refused" interval="3.2s" Dec 06 03:44:14 crc kubenswrapper[4802]: I1206 03:44:14.449891 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rnrvp" Dec 06 03:44:14 crc kubenswrapper[4802]: I1206 03:44:14.449947 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rnrvp" Dec 06 03:44:14 crc kubenswrapper[4802]: I1206 03:44:14.492035 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rnrvp" Dec 06 03:44:14 crc kubenswrapper[4802]: I1206 03:44:14.492590 4802 status_manager.go:851] "Failed to get status for pod" podUID="042f382d-836e-4cc2-9065-352a210f6593" pod="openshift-marketplace/certified-operators-skt5b" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-skt5b\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:14 crc kubenswrapper[4802]: I1206 03:44:14.492954 4802 status_manager.go:851] "Failed to get status for pod" podUID="b58e10a8-7605-4863-bc97-91714add5409" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:14 crc kubenswrapper[4802]: I1206 03:44:14.493306 4802 status_manager.go:851] "Failed to get status for pod" podUID="199a835e-3d62-42ef-8df0-6f087ae5586e" pod="openshift-marketplace/redhat-marketplace-2l4qq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-2l4qq\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:14 crc kubenswrapper[4802]: I1206 03:44:14.493537 4802 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:14 crc kubenswrapper[4802]: I1206 03:44:14.493811 4802 status_manager.go:851] "Failed to get status for pod" podUID="cae8cfbe-5644-45a5-92e6-0ce169cc47c5" pod="openshift-marketplace/community-operators-rnrvp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rnrvp\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:14 crc kubenswrapper[4802]: I1206 03:44:14.494185 4802 status_manager.go:851] "Failed to get status for pod" podUID="a640d4db-76d9-4c21-b394-cbc18a3fe2c4" pod="openshift-marketplace/redhat-operators-ldbdv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-ldbdv\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:14 crc kubenswrapper[4802]: I1206 03:44:14.670470 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rnrvp" Dec 06 03:44:14 crc kubenswrapper[4802]: I1206 03:44:14.671404 4802 status_manager.go:851] "Failed to get status for pod" podUID="cae8cfbe-5644-45a5-92e6-0ce169cc47c5" pod="openshift-marketplace/community-operators-rnrvp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rnrvp\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:14 crc kubenswrapper[4802]: I1206 03:44:14.671801 4802 status_manager.go:851] "Failed to get status for pod" podUID="a640d4db-76d9-4c21-b394-cbc18a3fe2c4" pod="openshift-marketplace/redhat-operators-ldbdv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-ldbdv\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:14 crc kubenswrapper[4802]: I1206 03:44:14.671996 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-skt5b" Dec 06 03:44:14 crc kubenswrapper[4802]: I1206 03:44:14.672068 4802 status_manager.go:851] "Failed to get status for pod" podUID="042f382d-836e-4cc2-9065-352a210f6593" pod="openshift-marketplace/certified-operators-skt5b" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-skt5b\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:14 crc kubenswrapper[4802]: I1206 03:44:14.672387 4802 status_manager.go:851] "Failed to get status for pod" podUID="b58e10a8-7605-4863-bc97-91714add5409" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:14 crc kubenswrapper[4802]: I1206 03:44:14.672813 4802 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:14 crc kubenswrapper[4802]: I1206 03:44:14.673085 4802 status_manager.go:851] "Failed to get status for pod" podUID="199a835e-3d62-42ef-8df0-6f087ae5586e" pod="openshift-marketplace/redhat-marketplace-2l4qq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-2l4qq\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:14 crc kubenswrapper[4802]: I1206 03:44:14.673365 4802 status_manager.go:851] "Failed to get status for pod" podUID="199a835e-3d62-42ef-8df0-6f087ae5586e" pod="openshift-marketplace/redhat-marketplace-2l4qq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-2l4qq\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:14 crc kubenswrapper[4802]: I1206 03:44:14.673613 4802 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:14 crc kubenswrapper[4802]: I1206 03:44:14.673852 4802 status_manager.go:851] "Failed to get status for pod" podUID="cae8cfbe-5644-45a5-92e6-0ce169cc47c5" pod="openshift-marketplace/community-operators-rnrvp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rnrvp\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:14 crc kubenswrapper[4802]: I1206 03:44:14.674108 4802 status_manager.go:851] "Failed to get status for pod" podUID="a640d4db-76d9-4c21-b394-cbc18a3fe2c4" pod="openshift-marketplace/redhat-operators-ldbdv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-ldbdv\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:14 crc kubenswrapper[4802]: I1206 03:44:14.674378 4802 status_manager.go:851] "Failed to get status for pod" podUID="042f382d-836e-4cc2-9065-352a210f6593" pod="openshift-marketplace/certified-operators-skt5b" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-skt5b\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:14 crc kubenswrapper[4802]: I1206 03:44:14.674594 4802 status_manager.go:851] "Failed to get status for pod" podUID="b58e10a8-7605-4863-bc97-91714add5409" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:16 crc kubenswrapper[4802]: I1206 03:44:16.229669 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2l4qq" Dec 06 03:44:16 crc kubenswrapper[4802]: I1206 03:44:16.229774 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2l4qq" Dec 06 03:44:16 crc kubenswrapper[4802]: I1206 03:44:16.276226 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2l4qq" Dec 06 03:44:16 crc kubenswrapper[4802]: I1206 03:44:16.276833 4802 status_manager.go:851] "Failed to get status for pod" podUID="b58e10a8-7605-4863-bc97-91714add5409" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:16 crc kubenswrapper[4802]: I1206 03:44:16.277240 4802 status_manager.go:851] "Failed to get status for pod" podUID="199a835e-3d62-42ef-8df0-6f087ae5586e" pod="openshift-marketplace/redhat-marketplace-2l4qq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-2l4qq\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:16 crc kubenswrapper[4802]: I1206 03:44:16.277590 4802 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:16 crc kubenswrapper[4802]: I1206 03:44:16.277911 4802 status_manager.go:851] "Failed to get status for pod" podUID="a640d4db-76d9-4c21-b394-cbc18a3fe2c4" pod="openshift-marketplace/redhat-operators-ldbdv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-ldbdv\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:16 crc kubenswrapper[4802]: I1206 03:44:16.278154 4802 status_manager.go:851] "Failed to get status for pod" podUID="cae8cfbe-5644-45a5-92e6-0ce169cc47c5" pod="openshift-marketplace/community-operators-rnrvp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rnrvp\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:16 crc kubenswrapper[4802]: I1206 03:44:16.278427 4802 status_manager.go:851] "Failed to get status for pod" podUID="042f382d-836e-4cc2-9065-352a210f6593" pod="openshift-marketplace/certified-operators-skt5b" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-skt5b\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:16 crc kubenswrapper[4802]: E1206 03:44:16.316513 4802 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/events\": dial tcp 38.102.83.150:6443: connect: connection refused" event="&Event{ObjectMeta:{redhat-marketplace-2l4qq.187e83764b55d197 openshift-marketplace 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-marketplace,Name:redhat-marketplace-2l4qq,UID:199a835e-3d62-42ef-8df0-6f087ae5586e,APIVersion:v1,ResourceVersion:29462,FieldPath:spec.initContainers{extract-content},},Reason:Pulled,Message:Successfully pulled image \"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\" in 524ms (524ms including waiting). Image size: 1141987142 bytes.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-06 03:44:08.069108119 +0000 UTC m=+240.941017271,LastTimestamp:2025-12-06 03:44:08.069108119 +0000 UTC m=+240.941017271,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 06 03:44:16 crc kubenswrapper[4802]: I1206 03:44:16.702728 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2l4qq" Dec 06 03:44:16 crc kubenswrapper[4802]: I1206 03:44:16.703201 4802 status_manager.go:851] "Failed to get status for pod" podUID="042f382d-836e-4cc2-9065-352a210f6593" pod="openshift-marketplace/certified-operators-skt5b" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-skt5b\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:16 crc kubenswrapper[4802]: I1206 03:44:16.703609 4802 status_manager.go:851] "Failed to get status for pod" podUID="b58e10a8-7605-4863-bc97-91714add5409" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:16 crc kubenswrapper[4802]: I1206 03:44:16.703900 4802 status_manager.go:851] "Failed to get status for pod" podUID="199a835e-3d62-42ef-8df0-6f087ae5586e" pod="openshift-marketplace/redhat-marketplace-2l4qq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-2l4qq\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:16 crc kubenswrapper[4802]: I1206 03:44:16.704151 4802 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:16 crc kubenswrapper[4802]: I1206 03:44:16.704357 4802 status_manager.go:851] "Failed to get status for pod" podUID="cae8cfbe-5644-45a5-92e6-0ce169cc47c5" pod="openshift-marketplace/community-operators-rnrvp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rnrvp\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:16 crc kubenswrapper[4802]: I1206 03:44:16.704552 4802 status_manager.go:851] "Failed to get status for pod" podUID="a640d4db-76d9-4c21-b394-cbc18a3fe2c4" pod="openshift-marketplace/redhat-operators-ldbdv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-ldbdv\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:16 crc kubenswrapper[4802]: I1206 03:44:16.865257 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-ldbdv" Dec 06 03:44:16 crc kubenswrapper[4802]: I1206 03:44:16.865331 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-ldbdv" Dec 06 03:44:16 crc kubenswrapper[4802]: I1206 03:44:16.920553 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-ldbdv" Dec 06 03:44:16 crc kubenswrapper[4802]: I1206 03:44:16.921238 4802 status_manager.go:851] "Failed to get status for pod" podUID="199a835e-3d62-42ef-8df0-6f087ae5586e" pod="openshift-marketplace/redhat-marketplace-2l4qq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-2l4qq\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:16 crc kubenswrapper[4802]: I1206 03:44:16.921686 4802 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:16 crc kubenswrapper[4802]: I1206 03:44:16.922056 4802 status_manager.go:851] "Failed to get status for pod" podUID="a640d4db-76d9-4c21-b394-cbc18a3fe2c4" pod="openshift-marketplace/redhat-operators-ldbdv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-ldbdv\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:16 crc kubenswrapper[4802]: I1206 03:44:16.922389 4802 status_manager.go:851] "Failed to get status for pod" podUID="cae8cfbe-5644-45a5-92e6-0ce169cc47c5" pod="openshift-marketplace/community-operators-rnrvp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rnrvp\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:16 crc kubenswrapper[4802]: I1206 03:44:16.922712 4802 status_manager.go:851] "Failed to get status for pod" podUID="042f382d-836e-4cc2-9065-352a210f6593" pod="openshift-marketplace/certified-operators-skt5b" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-skt5b\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:16 crc kubenswrapper[4802]: I1206 03:44:16.923100 4802 status_manager.go:851] "Failed to get status for pod" podUID="b58e10a8-7605-4863-bc97-91714add5409" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:17 crc kubenswrapper[4802]: I1206 03:44:17.452441 4802 status_manager.go:851] "Failed to get status for pod" podUID="042f382d-836e-4cc2-9065-352a210f6593" pod="openshift-marketplace/certified-operators-skt5b" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-skt5b\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:17 crc kubenswrapper[4802]: I1206 03:44:17.453519 4802 status_manager.go:851] "Failed to get status for pod" podUID="b58e10a8-7605-4863-bc97-91714add5409" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:17 crc kubenswrapper[4802]: I1206 03:44:17.453892 4802 status_manager.go:851] "Failed to get status for pod" podUID="199a835e-3d62-42ef-8df0-6f087ae5586e" pod="openshift-marketplace/redhat-marketplace-2l4qq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-2l4qq\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:17 crc kubenswrapper[4802]: I1206 03:44:17.454381 4802 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:17 crc kubenswrapper[4802]: I1206 03:44:17.454810 4802 status_manager.go:851] "Failed to get status for pod" podUID="a640d4db-76d9-4c21-b394-cbc18a3fe2c4" pod="openshift-marketplace/redhat-operators-ldbdv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-ldbdv\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:17 crc kubenswrapper[4802]: I1206 03:44:17.455229 4802 status_manager.go:851] "Failed to get status for pod" podUID="cae8cfbe-5644-45a5-92e6-0ce169cc47c5" pod="openshift-marketplace/community-operators-rnrvp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rnrvp\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:17 crc kubenswrapper[4802]: E1206 03:44:17.629457 4802 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.150:6443: connect: connection refused" interval="6.4s" Dec 06 03:44:17 crc kubenswrapper[4802]: I1206 03:44:17.693091 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-ldbdv" Dec 06 03:44:17 crc kubenswrapper[4802]: I1206 03:44:17.693815 4802 status_manager.go:851] "Failed to get status for pod" podUID="b58e10a8-7605-4863-bc97-91714add5409" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:17 crc kubenswrapper[4802]: I1206 03:44:17.694193 4802 status_manager.go:851] "Failed to get status for pod" podUID="199a835e-3d62-42ef-8df0-6f087ae5586e" pod="openshift-marketplace/redhat-marketplace-2l4qq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-2l4qq\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:17 crc kubenswrapper[4802]: I1206 03:44:17.694555 4802 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:17 crc kubenswrapper[4802]: I1206 03:44:17.694913 4802 status_manager.go:851] "Failed to get status for pod" podUID="cae8cfbe-5644-45a5-92e6-0ce169cc47c5" pod="openshift-marketplace/community-operators-rnrvp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rnrvp\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:17 crc kubenswrapper[4802]: I1206 03:44:17.695218 4802 status_manager.go:851] "Failed to get status for pod" podUID="a640d4db-76d9-4c21-b394-cbc18a3fe2c4" pod="openshift-marketplace/redhat-operators-ldbdv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-ldbdv\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:17 crc kubenswrapper[4802]: I1206 03:44:17.695610 4802 status_manager.go:851] "Failed to get status for pod" podUID="042f382d-836e-4cc2-9065-352a210f6593" pod="openshift-marketplace/certified-operators-skt5b" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-skt5b\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:20 crc kubenswrapper[4802]: I1206 03:44:20.449253 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:44:20 crc kubenswrapper[4802]: I1206 03:44:20.450351 4802 status_manager.go:851] "Failed to get status for pod" podUID="cae8cfbe-5644-45a5-92e6-0ce169cc47c5" pod="openshift-marketplace/community-operators-rnrvp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rnrvp\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:20 crc kubenswrapper[4802]: I1206 03:44:20.450859 4802 status_manager.go:851] "Failed to get status for pod" podUID="a640d4db-76d9-4c21-b394-cbc18a3fe2c4" pod="openshift-marketplace/redhat-operators-ldbdv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-ldbdv\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:20 crc kubenswrapper[4802]: I1206 03:44:20.451324 4802 status_manager.go:851] "Failed to get status for pod" podUID="042f382d-836e-4cc2-9065-352a210f6593" pod="openshift-marketplace/certified-operators-skt5b" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-skt5b\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:20 crc kubenswrapper[4802]: I1206 03:44:20.451852 4802 status_manager.go:851] "Failed to get status for pod" podUID="b58e10a8-7605-4863-bc97-91714add5409" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:20 crc kubenswrapper[4802]: I1206 03:44:20.452293 4802 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:20 crc kubenswrapper[4802]: I1206 03:44:20.452820 4802 status_manager.go:851] "Failed to get status for pod" podUID="199a835e-3d62-42ef-8df0-6f087ae5586e" pod="openshift-marketplace/redhat-marketplace-2l4qq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-2l4qq\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:20 crc kubenswrapper[4802]: I1206 03:44:20.465573 4802 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1721e985-40bd-4b2d-a0e8-d4e365094ba6" Dec 06 03:44:20 crc kubenswrapper[4802]: I1206 03:44:20.465620 4802 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1721e985-40bd-4b2d-a0e8-d4e365094ba6" Dec 06 03:44:20 crc kubenswrapper[4802]: E1206 03:44:20.466097 4802 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:44:20 crc kubenswrapper[4802]: I1206 03:44:20.466578 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:44:20 crc kubenswrapper[4802]: W1206 03:44:20.488703 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-edad9f630c5e5c8d40e951b89e24271b68c131f204e21a2ab9075a09fa7ef3dd WatchSource:0}: Error finding container edad9f630c5e5c8d40e951b89e24271b68c131f204e21a2ab9075a09fa7ef3dd: Status 404 returned error can't find the container with id edad9f630c5e5c8d40e951b89e24271b68c131f204e21a2ab9075a09fa7ef3dd Dec 06 03:44:20 crc kubenswrapper[4802]: I1206 03:44:20.661689 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"edad9f630c5e5c8d40e951b89e24271b68c131f204e21a2ab9075a09fa7ef3dd"} Dec 06 03:44:23 crc kubenswrapper[4802]: I1206 03:44:23.685148 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"8e5e5de03cddcb5ec234b1c4e68a3c4ca8d6a7259d788ec0d93ea6afd5467bcd"} Dec 06 03:44:24 crc kubenswrapper[4802]: E1206 03:44:24.031413 4802 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.150:6443: connect: connection refused" interval="7s" Dec 06 03:44:24 crc kubenswrapper[4802]: I1206 03:44:24.695675 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 06 03:44:24 crc kubenswrapper[4802]: I1206 03:44:24.695809 4802 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="03004a29aefe7fe2452fb9f6663c0807afccb055d7ddbf6b907f305e6a19c81c" exitCode=1 Dec 06 03:44:24 crc kubenswrapper[4802]: I1206 03:44:24.695925 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"03004a29aefe7fe2452fb9f6663c0807afccb055d7ddbf6b907f305e6a19c81c"} Dec 06 03:44:24 crc kubenswrapper[4802]: I1206 03:44:24.696568 4802 scope.go:117] "RemoveContainer" containerID="03004a29aefe7fe2452fb9f6663c0807afccb055d7ddbf6b907f305e6a19c81c" Dec 06 03:44:24 crc kubenswrapper[4802]: I1206 03:44:24.697341 4802 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:24 crc kubenswrapper[4802]: I1206 03:44:24.698093 4802 status_manager.go:851] "Failed to get status for pod" podUID="cae8cfbe-5644-45a5-92e6-0ce169cc47c5" pod="openshift-marketplace/community-operators-rnrvp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rnrvp\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:24 crc kubenswrapper[4802]: I1206 03:44:24.698640 4802 status_manager.go:851] "Failed to get status for pod" podUID="a640d4db-76d9-4c21-b394-cbc18a3fe2c4" pod="openshift-marketplace/redhat-operators-ldbdv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-ldbdv\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:24 crc kubenswrapper[4802]: I1206 03:44:24.698984 4802 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="8e5e5de03cddcb5ec234b1c4e68a3c4ca8d6a7259d788ec0d93ea6afd5467bcd" exitCode=0 Dec 06 03:44:24 crc kubenswrapper[4802]: I1206 03:44:24.699090 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"8e5e5de03cddcb5ec234b1c4e68a3c4ca8d6a7259d788ec0d93ea6afd5467bcd"} Dec 06 03:44:24 crc kubenswrapper[4802]: I1206 03:44:24.699106 4802 status_manager.go:851] "Failed to get status for pod" podUID="042f382d-836e-4cc2-9065-352a210f6593" pod="openshift-marketplace/certified-operators-skt5b" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-skt5b\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:24 crc kubenswrapper[4802]: I1206 03:44:24.699475 4802 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1721e985-40bd-4b2d-a0e8-d4e365094ba6" Dec 06 03:44:24 crc kubenswrapper[4802]: I1206 03:44:24.699522 4802 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1721e985-40bd-4b2d-a0e8-d4e365094ba6" Dec 06 03:44:24 crc kubenswrapper[4802]: I1206 03:44:24.699670 4802 status_manager.go:851] "Failed to get status for pod" podUID="b58e10a8-7605-4863-bc97-91714add5409" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:24 crc kubenswrapper[4802]: E1206 03:44:24.700180 4802 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:44:24 crc kubenswrapper[4802]: I1206 03:44:24.700340 4802 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:24 crc kubenswrapper[4802]: I1206 03:44:24.701184 4802 status_manager.go:851] "Failed to get status for pod" podUID="199a835e-3d62-42ef-8df0-6f087ae5586e" pod="openshift-marketplace/redhat-marketplace-2l4qq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-2l4qq\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:24 crc kubenswrapper[4802]: I1206 03:44:24.702043 4802 status_manager.go:851] "Failed to get status for pod" podUID="cae8cfbe-5644-45a5-92e6-0ce169cc47c5" pod="openshift-marketplace/community-operators-rnrvp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rnrvp\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:24 crc kubenswrapper[4802]: I1206 03:44:24.702506 4802 status_manager.go:851] "Failed to get status for pod" podUID="a640d4db-76d9-4c21-b394-cbc18a3fe2c4" pod="openshift-marketplace/redhat-operators-ldbdv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-ldbdv\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:24 crc kubenswrapper[4802]: I1206 03:44:24.703049 4802 status_manager.go:851] "Failed to get status for pod" podUID="042f382d-836e-4cc2-9065-352a210f6593" pod="openshift-marketplace/certified-operators-skt5b" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-skt5b\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:24 crc kubenswrapper[4802]: I1206 03:44:24.703614 4802 status_manager.go:851] "Failed to get status for pod" podUID="b58e10a8-7605-4863-bc97-91714add5409" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:24 crc kubenswrapper[4802]: I1206 03:44:24.704346 4802 status_manager.go:851] "Failed to get status for pod" podUID="199a835e-3d62-42ef-8df0-6f087ae5586e" pod="openshift-marketplace/redhat-marketplace-2l4qq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-2l4qq\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:24 crc kubenswrapper[4802]: I1206 03:44:24.704852 4802 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:24 crc kubenswrapper[4802]: I1206 03:44:24.705330 4802 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 06 03:44:25 crc kubenswrapper[4802]: I1206 03:44:25.707804 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 06 03:44:25 crc kubenswrapper[4802]: I1206 03:44:25.708109 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"d800788811d0199918bc6e592d6762893f27a74a5f811c94ff3d28f20138e8ab"} Dec 06 03:44:25 crc kubenswrapper[4802]: I1206 03:44:25.710685 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"4660e3e3e274e44ce0fe7edb15d232fb8c79cce31a19ad896e807efc52d2b145"} Dec 06 03:44:26 crc kubenswrapper[4802]: I1206 03:44:26.277337 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-rqs9q" podUID="263721d5-fcca-411a-876e-0acfe0863d2d" containerName="oauth-openshift" containerID="cri-o://f6333b5fe7dbcf85d5cef40f1644c3754ad08959a1f3230bb14eaa606e8db655" gracePeriod=15 Dec 06 03:44:26 crc kubenswrapper[4802]: I1206 03:44:26.671771 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 03:44:26 crc kubenswrapper[4802]: I1206 03:44:26.717340 4802 generic.go:334] "Generic (PLEG): container finished" podID="263721d5-fcca-411a-876e-0acfe0863d2d" containerID="f6333b5fe7dbcf85d5cef40f1644c3754ad08959a1f3230bb14eaa606e8db655" exitCode=0 Dec 06 03:44:26 crc kubenswrapper[4802]: I1206 03:44:26.717471 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-rqs9q" event={"ID":"263721d5-fcca-411a-876e-0acfe0863d2d","Type":"ContainerDied","Data":"f6333b5fe7dbcf85d5cef40f1644c3754ad08959a1f3230bb14eaa606e8db655"} Dec 06 03:44:26 crc kubenswrapper[4802]: I1206 03:44:26.720887 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"d8a32d82bdbbd78a6271c9452522a21ec754f04de70a3033c3c35016e8242902"} Dec 06 03:44:26 crc kubenswrapper[4802]: I1206 03:44:26.720958 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"275f9a9c9b41b9ef6c94225d1b9676f207abd71ea38c4d1c4e8a664f1a92f698"} Dec 06 03:44:26 crc kubenswrapper[4802]: I1206 03:44:26.761211 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-rqs9q" Dec 06 03:44:27 crc kubenswrapper[4802]: I1206 03:44:27.728722 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-rqs9q" event={"ID":"263721d5-fcca-411a-876e-0acfe0863d2d","Type":"ContainerDied","Data":"19eb0b14b80bc6311964b5a7503f525c2e7d8fe43f506b84a274e59d8cbf29e5"} Dec 06 03:44:27 crc kubenswrapper[4802]: I1206 03:44:27.728786 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-rqs9q" Dec 06 03:44:27 crc kubenswrapper[4802]: I1206 03:44:27.728814 4802 scope.go:117] "RemoveContainer" containerID="f6333b5fe7dbcf85d5cef40f1644c3754ad08959a1f3230bb14eaa606e8db655" Dec 06 03:44:27 crc kubenswrapper[4802]: I1206 03:44:27.737275 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"5fa131fcadebf9071c36e87632f226b0a60488f579b1ae16523f29897285c78d"} Dec 06 03:44:27 crc kubenswrapper[4802]: I1206 03:44:27.737608 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:44:27 crc kubenswrapper[4802]: I1206 03:44:27.737623 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"9aeb4911a76aeb97a839ae8c3cec16b69f8acb2a62f30794ca6bc4e50280d1cc"} Dec 06 03:44:27 crc kubenswrapper[4802]: I1206 03:44:27.737881 4802 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1721e985-40bd-4b2d-a0e8-d4e365094ba6" Dec 06 03:44:27 crc kubenswrapper[4802]: I1206 03:44:27.737921 4802 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1721e985-40bd-4b2d-a0e8-d4e365094ba6" Dec 06 03:44:27 crc kubenswrapper[4802]: I1206 03:44:27.743710 4802 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:44:28 crc kubenswrapper[4802]: I1206 03:44:28.748309 4802 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1721e985-40bd-4b2d-a0e8-d4e365094ba6" Dec 06 03:44:28 crc kubenswrapper[4802]: I1206 03:44:28.748335 4802 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1721e985-40bd-4b2d-a0e8-d4e365094ba6" Dec 06 03:44:29 crc kubenswrapper[4802]: I1206 03:44:29.960252 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 03:44:29 crc kubenswrapper[4802]: I1206 03:44:29.967422 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 03:44:30 crc kubenswrapper[4802]: I1206 03:44:30.467062 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:44:30 crc kubenswrapper[4802]: I1206 03:44:30.467514 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:44:30 crc kubenswrapper[4802]: I1206 03:44:30.468352 4802 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1721e985-40bd-4b2d-a0e8-d4e365094ba6" Dec 06 03:44:30 crc kubenswrapper[4802]: I1206 03:44:30.468396 4802 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1721e985-40bd-4b2d-a0e8-d4e365094ba6" Dec 06 03:44:30 crc kubenswrapper[4802]: I1206 03:44:30.475649 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:44:30 crc kubenswrapper[4802]: I1206 03:44:30.760744 4802 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1721e985-40bd-4b2d-a0e8-d4e365094ba6" Dec 06 03:44:30 crc kubenswrapper[4802]: I1206 03:44:30.760825 4802 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1721e985-40bd-4b2d-a0e8-d4e365094ba6" Dec 06 03:44:30 crc kubenswrapper[4802]: I1206 03:44:30.767408 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:44:31 crc kubenswrapper[4802]: I1206 03:44:31.570666 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-system-service-ca\") pod \"263721d5-fcca-411a-876e-0acfe0863d2d\" (UID: \"263721d5-fcca-411a-876e-0acfe0863d2d\") " Dec 06 03:44:31 crc kubenswrapper[4802]: I1206 03:44:31.571605 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-system-trusted-ca-bundle\") pod \"263721d5-fcca-411a-876e-0acfe0863d2d\" (UID: \"263721d5-fcca-411a-876e-0acfe0863d2d\") " Dec 06 03:44:31 crc kubenswrapper[4802]: I1206 03:44:31.571479 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "263721d5-fcca-411a-876e-0acfe0863d2d" (UID: "263721d5-fcca-411a-876e-0acfe0863d2d"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:44:31 crc kubenswrapper[4802]: I1206 03:44:31.571784 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-system-session\") pod \"263721d5-fcca-411a-876e-0acfe0863d2d\" (UID: \"263721d5-fcca-411a-876e-0acfe0863d2d\") " Dec 06 03:44:31 crc kubenswrapper[4802]: I1206 03:44:31.571991 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "263721d5-fcca-411a-876e-0acfe0863d2d" (UID: "263721d5-fcca-411a-876e-0acfe0863d2d"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:44:31 crc kubenswrapper[4802]: I1206 03:44:31.572132 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-system-cliconfig\") pod \"263721d5-fcca-411a-876e-0acfe0863d2d\" (UID: \"263721d5-fcca-411a-876e-0acfe0863d2d\") " Dec 06 03:44:31 crc kubenswrapper[4802]: I1206 03:44:31.572300 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-system-serving-cert\") pod \"263721d5-fcca-411a-876e-0acfe0863d2d\" (UID: \"263721d5-fcca-411a-876e-0acfe0863d2d\") " Dec 06 03:44:31 crc kubenswrapper[4802]: I1206 03:44:31.572448 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-system-ocp-branding-template\") pod \"263721d5-fcca-411a-876e-0acfe0863d2d\" (UID: \"263721d5-fcca-411a-876e-0acfe0863d2d\") " Dec 06 03:44:31 crc kubenswrapper[4802]: I1206 03:44:31.572622 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/263721d5-fcca-411a-876e-0acfe0863d2d-audit-policies\") pod \"263721d5-fcca-411a-876e-0acfe0863d2d\" (UID: \"263721d5-fcca-411a-876e-0acfe0863d2d\") " Dec 06 03:44:31 crc kubenswrapper[4802]: I1206 03:44:31.572807 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-user-idp-0-file-data\") pod \"263721d5-fcca-411a-876e-0acfe0863d2d\" (UID: \"263721d5-fcca-411a-876e-0acfe0863d2d\") " Dec 06 03:44:31 crc kubenswrapper[4802]: I1206 03:44:31.572935 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/263721d5-fcca-411a-876e-0acfe0863d2d-audit-dir\") pod \"263721d5-fcca-411a-876e-0acfe0863d2d\" (UID: \"263721d5-fcca-411a-876e-0acfe0863d2d\") " Dec 06 03:44:31 crc kubenswrapper[4802]: I1206 03:44:31.573032 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bgrht\" (UniqueName: \"kubernetes.io/projected/263721d5-fcca-411a-876e-0acfe0863d2d-kube-api-access-bgrht\") pod \"263721d5-fcca-411a-876e-0acfe0863d2d\" (UID: \"263721d5-fcca-411a-876e-0acfe0863d2d\") " Dec 06 03:44:31 crc kubenswrapper[4802]: I1206 03:44:31.573164 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-user-template-login\") pod \"263721d5-fcca-411a-876e-0acfe0863d2d\" (UID: \"263721d5-fcca-411a-876e-0acfe0863d2d\") " Dec 06 03:44:31 crc kubenswrapper[4802]: I1206 03:44:31.572371 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "263721d5-fcca-411a-876e-0acfe0863d2d" (UID: "263721d5-fcca-411a-876e-0acfe0863d2d"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:44:31 crc kubenswrapper[4802]: I1206 03:44:31.572988 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/263721d5-fcca-411a-876e-0acfe0863d2d-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "263721d5-fcca-411a-876e-0acfe0863d2d" (UID: "263721d5-fcca-411a-876e-0acfe0863d2d"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:44:31 crc kubenswrapper[4802]: I1206 03:44:31.573028 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/263721d5-fcca-411a-876e-0acfe0863d2d-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "263721d5-fcca-411a-876e-0acfe0863d2d" (UID: "263721d5-fcca-411a-876e-0acfe0863d2d"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:44:31 crc kubenswrapper[4802]: I1206 03:44:31.573522 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-user-template-error\") pod \"263721d5-fcca-411a-876e-0acfe0863d2d\" (UID: \"263721d5-fcca-411a-876e-0acfe0863d2d\") " Dec 06 03:44:31 crc kubenswrapper[4802]: I1206 03:44:31.573652 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-system-router-certs\") pod \"263721d5-fcca-411a-876e-0acfe0863d2d\" (UID: \"263721d5-fcca-411a-876e-0acfe0863d2d\") " Dec 06 03:44:31 crc kubenswrapper[4802]: I1206 03:44:31.573823 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-user-template-provider-selection\") pod \"263721d5-fcca-411a-876e-0acfe0863d2d\" (UID: \"263721d5-fcca-411a-876e-0acfe0863d2d\") " Dec 06 03:44:31 crc kubenswrapper[4802]: I1206 03:44:31.574376 4802 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 06 03:44:31 crc kubenswrapper[4802]: I1206 03:44:31.574483 4802 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/263721d5-fcca-411a-876e-0acfe0863d2d-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 06 03:44:31 crc kubenswrapper[4802]: I1206 03:44:31.574566 4802 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/263721d5-fcca-411a-876e-0acfe0863d2d-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 06 03:44:31 crc kubenswrapper[4802]: I1206 03:44:31.576654 4802 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 06 03:44:31 crc kubenswrapper[4802]: I1206 03:44:31.576892 4802 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 03:44:31 crc kubenswrapper[4802]: I1206 03:44:31.600888 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "263721d5-fcca-411a-876e-0acfe0863d2d" (UID: "263721d5-fcca-411a-876e-0acfe0863d2d"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:44:31 crc kubenswrapper[4802]: I1206 03:44:31.601064 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "263721d5-fcca-411a-876e-0acfe0863d2d" (UID: "263721d5-fcca-411a-876e-0acfe0863d2d"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:44:31 crc kubenswrapper[4802]: I1206 03:44:31.601387 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "263721d5-fcca-411a-876e-0acfe0863d2d" (UID: "263721d5-fcca-411a-876e-0acfe0863d2d"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:44:31 crc kubenswrapper[4802]: I1206 03:44:31.602822 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "263721d5-fcca-411a-876e-0acfe0863d2d" (UID: "263721d5-fcca-411a-876e-0acfe0863d2d"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:44:31 crc kubenswrapper[4802]: I1206 03:44:31.603567 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "263721d5-fcca-411a-876e-0acfe0863d2d" (UID: "263721d5-fcca-411a-876e-0acfe0863d2d"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:44:31 crc kubenswrapper[4802]: I1206 03:44:31.603925 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/263721d5-fcca-411a-876e-0acfe0863d2d-kube-api-access-bgrht" (OuterVolumeSpecName: "kube-api-access-bgrht") pod "263721d5-fcca-411a-876e-0acfe0863d2d" (UID: "263721d5-fcca-411a-876e-0acfe0863d2d"). InnerVolumeSpecName "kube-api-access-bgrht". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:44:31 crc kubenswrapper[4802]: I1206 03:44:31.604159 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "263721d5-fcca-411a-876e-0acfe0863d2d" (UID: "263721d5-fcca-411a-876e-0acfe0863d2d"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:44:31 crc kubenswrapper[4802]: I1206 03:44:31.608062 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "263721d5-fcca-411a-876e-0acfe0863d2d" (UID: "263721d5-fcca-411a-876e-0acfe0863d2d"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:44:31 crc kubenswrapper[4802]: I1206 03:44:31.614072 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "263721d5-fcca-411a-876e-0acfe0863d2d" (UID: "263721d5-fcca-411a-876e-0acfe0863d2d"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:44:31 crc kubenswrapper[4802]: I1206 03:44:31.647829 4802 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="9453d406-16aa-4c7f-ac2b-a38a68b8744c" Dec 06 03:44:31 crc kubenswrapper[4802]: I1206 03:44:31.678793 4802 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 06 03:44:31 crc kubenswrapper[4802]: I1206 03:44:31.678857 4802 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 06 03:44:31 crc kubenswrapper[4802]: I1206 03:44:31.680845 4802 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 06 03:44:31 crc kubenswrapper[4802]: I1206 03:44:31.680960 4802 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:44:31 crc kubenswrapper[4802]: I1206 03:44:31.681015 4802 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 06 03:44:31 crc kubenswrapper[4802]: I1206 03:44:31.681065 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bgrht\" (UniqueName: \"kubernetes.io/projected/263721d5-fcca-411a-876e-0acfe0863d2d-kube-api-access-bgrht\") on node \"crc\" DevicePath \"\"" Dec 06 03:44:31 crc kubenswrapper[4802]: I1206 03:44:31.681150 4802 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 06 03:44:31 crc kubenswrapper[4802]: I1206 03:44:31.681205 4802 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 06 03:44:31 crc kubenswrapper[4802]: I1206 03:44:31.681299 4802 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/263721d5-fcca-411a-876e-0acfe0863d2d-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 06 03:44:31 crc kubenswrapper[4802]: I1206 03:44:31.763632 4802 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1721e985-40bd-4b2d-a0e8-d4e365094ba6" Dec 06 03:44:31 crc kubenswrapper[4802]: I1206 03:44:31.763921 4802 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1721e985-40bd-4b2d-a0e8-d4e365094ba6" Dec 06 03:44:31 crc kubenswrapper[4802]: I1206 03:44:31.766857 4802 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="9453d406-16aa-4c7f-ac2b-a38a68b8744c" Dec 06 03:44:32 crc kubenswrapper[4802]: E1206 03:44:32.364672 4802 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"v4-0-config-user-template-provider-selection\": Failed to watch *v1.Secret: unknown (get secrets)" logger="UnhandledError" Dec 06 03:44:36 crc kubenswrapper[4802]: I1206 03:44:36.675491 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 03:44:41 crc kubenswrapper[4802]: I1206 03:44:41.657939 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 06 03:44:41 crc kubenswrapper[4802]: I1206 03:44:41.905971 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 06 03:44:42 crc kubenswrapper[4802]: I1206 03:44:42.595029 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 06 03:44:42 crc kubenswrapper[4802]: I1206 03:44:42.620521 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 06 03:44:42 crc kubenswrapper[4802]: I1206 03:44:42.654176 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 06 03:44:42 crc kubenswrapper[4802]: I1206 03:44:42.884313 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 06 03:44:43 crc kubenswrapper[4802]: I1206 03:44:43.372964 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 06 03:44:43 crc kubenswrapper[4802]: I1206 03:44:43.389482 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 06 03:44:43 crc kubenswrapper[4802]: I1206 03:44:43.421266 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 06 03:44:43 crc kubenswrapper[4802]: I1206 03:44:43.943028 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 06 03:44:44 crc kubenswrapper[4802]: I1206 03:44:44.134181 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 06 03:44:44 crc kubenswrapper[4802]: I1206 03:44:44.249253 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 06 03:44:44 crc kubenswrapper[4802]: I1206 03:44:44.339958 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 06 03:44:44 crc kubenswrapper[4802]: I1206 03:44:44.381330 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 06 03:44:44 crc kubenswrapper[4802]: I1206 03:44:44.416936 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 06 03:44:44 crc kubenswrapper[4802]: I1206 03:44:44.463194 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 06 03:44:44 crc kubenswrapper[4802]: I1206 03:44:44.799112 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 06 03:44:44 crc kubenswrapper[4802]: I1206 03:44:44.891867 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 06 03:44:44 crc kubenswrapper[4802]: I1206 03:44:44.979716 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 06 03:44:45 crc kubenswrapper[4802]: I1206 03:44:45.005080 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 06 03:44:45 crc kubenswrapper[4802]: I1206 03:44:45.277349 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 06 03:44:45 crc kubenswrapper[4802]: I1206 03:44:45.280968 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 06 03:44:45 crc kubenswrapper[4802]: I1206 03:44:45.394153 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 06 03:44:45 crc kubenswrapper[4802]: I1206 03:44:45.460564 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 06 03:44:45 crc kubenswrapper[4802]: I1206 03:44:45.482944 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 06 03:44:45 crc kubenswrapper[4802]: I1206 03:44:45.549299 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 06 03:44:45 crc kubenswrapper[4802]: I1206 03:44:45.594693 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 06 03:44:45 crc kubenswrapper[4802]: I1206 03:44:45.661065 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 06 03:44:45 crc kubenswrapper[4802]: I1206 03:44:45.707902 4802 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 06 03:44:45 crc kubenswrapper[4802]: I1206 03:44:45.740668 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 06 03:44:45 crc kubenswrapper[4802]: I1206 03:44:45.788374 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 06 03:44:45 crc kubenswrapper[4802]: I1206 03:44:45.813555 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 06 03:44:45 crc kubenswrapper[4802]: I1206 03:44:45.813854 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 06 03:44:45 crc kubenswrapper[4802]: I1206 03:44:45.845659 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 06 03:44:45 crc kubenswrapper[4802]: I1206 03:44:45.893932 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 06 03:44:45 crc kubenswrapper[4802]: I1206 03:44:45.961697 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 06 03:44:46 crc kubenswrapper[4802]: I1206 03:44:46.011703 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 06 03:44:46 crc kubenswrapper[4802]: I1206 03:44:46.231128 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 06 03:44:46 crc kubenswrapper[4802]: I1206 03:44:46.254715 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 06 03:44:46 crc kubenswrapper[4802]: I1206 03:44:46.328821 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 06 03:44:46 crc kubenswrapper[4802]: I1206 03:44:46.352545 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 06 03:44:46 crc kubenswrapper[4802]: I1206 03:44:46.423297 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 06 03:44:46 crc kubenswrapper[4802]: I1206 03:44:46.533557 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 06 03:44:46 crc kubenswrapper[4802]: I1206 03:44:46.683100 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 06 03:44:46 crc kubenswrapper[4802]: I1206 03:44:46.897542 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 06 03:44:46 crc kubenswrapper[4802]: I1206 03:44:46.948126 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 06 03:44:47 crc kubenswrapper[4802]: I1206 03:44:47.236724 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 06 03:44:47 crc kubenswrapper[4802]: I1206 03:44:47.241359 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 06 03:44:47 crc kubenswrapper[4802]: I1206 03:44:47.263899 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 06 03:44:47 crc kubenswrapper[4802]: I1206 03:44:47.278068 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 06 03:44:47 crc kubenswrapper[4802]: I1206 03:44:47.321027 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 06 03:44:47 crc kubenswrapper[4802]: I1206 03:44:47.323149 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 06 03:44:47 crc kubenswrapper[4802]: I1206 03:44:47.327445 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 06 03:44:47 crc kubenswrapper[4802]: I1206 03:44:47.376977 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 06 03:44:47 crc kubenswrapper[4802]: I1206 03:44:47.419811 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 06 03:44:47 crc kubenswrapper[4802]: I1206 03:44:47.498162 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 06 03:44:47 crc kubenswrapper[4802]: I1206 03:44:47.552590 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 06 03:44:47 crc kubenswrapper[4802]: I1206 03:44:47.599386 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 06 03:44:47 crc kubenswrapper[4802]: I1206 03:44:47.599684 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 06 03:44:47 crc kubenswrapper[4802]: I1206 03:44:47.661291 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 06 03:44:47 crc kubenswrapper[4802]: I1206 03:44:47.723547 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 06 03:44:47 crc kubenswrapper[4802]: I1206 03:44:47.798976 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 06 03:44:47 crc kubenswrapper[4802]: I1206 03:44:47.850149 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 06 03:44:48 crc kubenswrapper[4802]: I1206 03:44:48.019560 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 06 03:44:48 crc kubenswrapper[4802]: I1206 03:44:48.044561 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 06 03:44:48 crc kubenswrapper[4802]: I1206 03:44:48.089598 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 06 03:44:48 crc kubenswrapper[4802]: I1206 03:44:48.304371 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 06 03:44:48 crc kubenswrapper[4802]: I1206 03:44:48.321205 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 06 03:44:48 crc kubenswrapper[4802]: I1206 03:44:48.338537 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 06 03:44:48 crc kubenswrapper[4802]: I1206 03:44:48.355782 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 06 03:44:48 crc kubenswrapper[4802]: I1206 03:44:48.399869 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 06 03:44:48 crc kubenswrapper[4802]: I1206 03:44:48.405518 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 06 03:44:48 crc kubenswrapper[4802]: I1206 03:44:48.442358 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 06 03:44:48 crc kubenswrapper[4802]: I1206 03:44:48.459232 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 06 03:44:48 crc kubenswrapper[4802]: I1206 03:44:48.488490 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 06 03:44:48 crc kubenswrapper[4802]: I1206 03:44:48.492568 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 06 03:44:48 crc kubenswrapper[4802]: I1206 03:44:48.503506 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 06 03:44:48 crc kubenswrapper[4802]: I1206 03:44:48.533378 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 06 03:44:48 crc kubenswrapper[4802]: I1206 03:44:48.587336 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 06 03:44:48 crc kubenswrapper[4802]: I1206 03:44:48.602314 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 06 03:44:48 crc kubenswrapper[4802]: I1206 03:44:48.619163 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 06 03:44:48 crc kubenswrapper[4802]: I1206 03:44:48.744896 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 06 03:44:48 crc kubenswrapper[4802]: I1206 03:44:48.859473 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 06 03:44:48 crc kubenswrapper[4802]: I1206 03:44:48.860782 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 06 03:44:48 crc kubenswrapper[4802]: I1206 03:44:48.975928 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 06 03:44:48 crc kubenswrapper[4802]: I1206 03:44:48.977995 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 06 03:44:49 crc kubenswrapper[4802]: I1206 03:44:49.092982 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 06 03:44:49 crc kubenswrapper[4802]: I1206 03:44:49.103425 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 06 03:44:49 crc kubenswrapper[4802]: I1206 03:44:49.172829 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 06 03:44:49 crc kubenswrapper[4802]: I1206 03:44:49.219612 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 06 03:44:49 crc kubenswrapper[4802]: I1206 03:44:49.308991 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 06 03:44:49 crc kubenswrapper[4802]: I1206 03:44:49.355286 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 06 03:44:49 crc kubenswrapper[4802]: I1206 03:44:49.404038 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 06 03:44:49 crc kubenswrapper[4802]: I1206 03:44:49.457837 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 06 03:44:49 crc kubenswrapper[4802]: I1206 03:44:49.470071 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 06 03:44:49 crc kubenswrapper[4802]: I1206 03:44:49.563962 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 06 03:44:49 crc kubenswrapper[4802]: I1206 03:44:49.642468 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 06 03:44:49 crc kubenswrapper[4802]: I1206 03:44:49.723829 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 06 03:44:49 crc kubenswrapper[4802]: I1206 03:44:49.735716 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 06 03:44:49 crc kubenswrapper[4802]: I1206 03:44:49.776700 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 06 03:44:49 crc kubenswrapper[4802]: I1206 03:44:49.822214 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 06 03:44:49 crc kubenswrapper[4802]: I1206 03:44:49.860183 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 06 03:44:49 crc kubenswrapper[4802]: I1206 03:44:49.894307 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 06 03:44:49 crc kubenswrapper[4802]: I1206 03:44:49.912462 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 06 03:44:49 crc kubenswrapper[4802]: I1206 03:44:49.946303 4802 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 06 03:44:50 crc kubenswrapper[4802]: I1206 03:44:50.072149 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 06 03:44:50 crc kubenswrapper[4802]: I1206 03:44:50.112628 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 06 03:44:50 crc kubenswrapper[4802]: I1206 03:44:50.239160 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 06 03:44:50 crc kubenswrapper[4802]: I1206 03:44:50.268443 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 06 03:44:50 crc kubenswrapper[4802]: I1206 03:44:50.381905 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 06 03:44:50 crc kubenswrapper[4802]: I1206 03:44:50.416919 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 06 03:44:50 crc kubenswrapper[4802]: I1206 03:44:50.507367 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 06 03:44:50 crc kubenswrapper[4802]: I1206 03:44:50.524656 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 06 03:44:50 crc kubenswrapper[4802]: I1206 03:44:50.544459 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 06 03:44:50 crc kubenswrapper[4802]: I1206 03:44:50.607408 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 06 03:44:50 crc kubenswrapper[4802]: I1206 03:44:50.631094 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 06 03:44:50 crc kubenswrapper[4802]: I1206 03:44:50.695848 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 06 03:44:50 crc kubenswrapper[4802]: I1206 03:44:50.718328 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 06 03:44:50 crc kubenswrapper[4802]: I1206 03:44:50.749910 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 06 03:44:50 crc kubenswrapper[4802]: I1206 03:44:50.838484 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 06 03:44:50 crc kubenswrapper[4802]: I1206 03:44:50.882974 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 06 03:44:50 crc kubenswrapper[4802]: I1206 03:44:50.933648 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 06 03:44:50 crc kubenswrapper[4802]: I1206 03:44:50.982248 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 06 03:44:50 crc kubenswrapper[4802]: I1206 03:44:50.990371 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 06 03:44:51 crc kubenswrapper[4802]: I1206 03:44:51.328438 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 06 03:44:51 crc kubenswrapper[4802]: I1206 03:44:51.333198 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 06 03:44:51 crc kubenswrapper[4802]: I1206 03:44:51.554986 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 06 03:44:51 crc kubenswrapper[4802]: I1206 03:44:51.811845 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 06 03:44:51 crc kubenswrapper[4802]: I1206 03:44:51.892085 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 06 03:44:51 crc kubenswrapper[4802]: I1206 03:44:51.909215 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 06 03:44:52 crc kubenswrapper[4802]: I1206 03:44:52.094672 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 06 03:44:52 crc kubenswrapper[4802]: I1206 03:44:52.194473 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 06 03:44:52 crc kubenswrapper[4802]: I1206 03:44:52.240291 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 06 03:44:52 crc kubenswrapper[4802]: I1206 03:44:52.272739 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 06 03:44:52 crc kubenswrapper[4802]: I1206 03:44:52.276941 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 06 03:44:52 crc kubenswrapper[4802]: I1206 03:44:52.348625 4802 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 06 03:44:52 crc kubenswrapper[4802]: I1206 03:44:52.396135 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 06 03:44:52 crc kubenswrapper[4802]: I1206 03:44:52.443731 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 06 03:44:52 crc kubenswrapper[4802]: I1206 03:44:52.459921 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 06 03:44:52 crc kubenswrapper[4802]: I1206 03:44:52.494034 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 06 03:44:52 crc kubenswrapper[4802]: I1206 03:44:52.498455 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 06 03:44:52 crc kubenswrapper[4802]: I1206 03:44:52.505241 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 06 03:44:52 crc kubenswrapper[4802]: I1206 03:44:52.506359 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 06 03:44:52 crc kubenswrapper[4802]: I1206 03:44:52.586806 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 06 03:44:52 crc kubenswrapper[4802]: I1206 03:44:52.596285 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 06 03:44:52 crc kubenswrapper[4802]: I1206 03:44:52.609518 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 06 03:44:52 crc kubenswrapper[4802]: I1206 03:44:52.651405 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 06 03:44:52 crc kubenswrapper[4802]: I1206 03:44:52.662991 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 06 03:44:52 crc kubenswrapper[4802]: I1206 03:44:52.717389 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 06 03:44:52 crc kubenswrapper[4802]: I1206 03:44:52.758296 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 06 03:44:52 crc kubenswrapper[4802]: I1206 03:44:52.769949 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 06 03:44:52 crc kubenswrapper[4802]: I1206 03:44:52.836402 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 06 03:44:52 crc kubenswrapper[4802]: I1206 03:44:52.957127 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 06 03:44:52 crc kubenswrapper[4802]: I1206 03:44:52.967692 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 06 03:44:53 crc kubenswrapper[4802]: I1206 03:44:53.092326 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 06 03:44:53 crc kubenswrapper[4802]: I1206 03:44:53.121809 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 06 03:44:53 crc kubenswrapper[4802]: I1206 03:44:53.238309 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 06 03:44:53 crc kubenswrapper[4802]: I1206 03:44:53.245242 4802 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 06 03:44:53 crc kubenswrapper[4802]: I1206 03:44:53.376250 4802 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 06 03:44:53 crc kubenswrapper[4802]: I1206 03:44:53.376781 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2l4qq" podStartSLOduration=46.96469262 podStartE2EDuration="48.376759327s" podCreationTimestamp="2025-12-06 03:44:05 +0000 UTC" firstStartedPulling="2025-12-06 03:44:07.54462813 +0000 UTC m=+240.416537302" lastFinishedPulling="2025-12-06 03:44:08.956694857 +0000 UTC m=+241.828604009" observedRunningTime="2025-12-06 03:44:31.560714476 +0000 UTC m=+264.432623648" watchObservedRunningTime="2025-12-06 03:44:53.376759327 +0000 UTC m=+286.248668479" Dec 06 03:44:53 crc kubenswrapper[4802]: I1206 03:44:53.377501 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=46.377496098 podStartE2EDuration="46.377496098s" podCreationTimestamp="2025-12-06 03:44:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:44:31.574374747 +0000 UTC m=+264.446283929" watchObservedRunningTime="2025-12-06 03:44:53.377496098 +0000 UTC m=+286.249405270" Dec 06 03:44:53 crc kubenswrapper[4802]: I1206 03:44:53.380120 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-ldbdv" podStartSLOduration=44.885852098 podStartE2EDuration="47.380111409s" podCreationTimestamp="2025-12-06 03:44:06 +0000 UTC" firstStartedPulling="2025-12-06 03:44:07.521259542 +0000 UTC m=+240.393168694" lastFinishedPulling="2025-12-06 03:44:10.015518853 +0000 UTC m=+242.887428005" observedRunningTime="2025-12-06 03:44:31.610886359 +0000 UTC m=+264.482795521" watchObservedRunningTime="2025-12-06 03:44:53.380111409 +0000 UTC m=+286.252020561" Dec 06 03:44:53 crc kubenswrapper[4802]: I1206 03:44:53.381334 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-rqs9q","openshift-kube-apiserver/kube-apiserver-crc"] Dec 06 03:44:53 crc kubenswrapper[4802]: I1206 03:44:53.381399 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 06 03:44:53 crc kubenswrapper[4802]: I1206 03:44:53.397000 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:44:53 crc kubenswrapper[4802]: I1206 03:44:53.398184 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 06 03:44:53 crc kubenswrapper[4802]: I1206 03:44:53.407996 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=26.407974776 podStartE2EDuration="26.407974776s" podCreationTimestamp="2025-12-06 03:44:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:44:53.405783136 +0000 UTC m=+286.277692298" watchObservedRunningTime="2025-12-06 03:44:53.407974776 +0000 UTC m=+286.279883938" Dec 06 03:44:53 crc kubenswrapper[4802]: I1206 03:44:53.434388 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 06 03:44:53 crc kubenswrapper[4802]: I1206 03:44:53.460294 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="263721d5-fcca-411a-876e-0acfe0863d2d" path="/var/lib/kubelet/pods/263721d5-fcca-411a-876e-0acfe0863d2d/volumes" Dec 06 03:44:53 crc kubenswrapper[4802]: I1206 03:44:53.540231 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 06 03:44:53 crc kubenswrapper[4802]: I1206 03:44:53.559390 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 06 03:44:53 crc kubenswrapper[4802]: I1206 03:44:53.604363 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 06 03:44:53 crc kubenswrapper[4802]: I1206 03:44:53.613120 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 06 03:44:53 crc kubenswrapper[4802]: I1206 03:44:53.636485 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 06 03:44:53 crc kubenswrapper[4802]: I1206 03:44:53.647701 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 06 03:44:53 crc kubenswrapper[4802]: I1206 03:44:53.662591 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 06 03:44:53 crc kubenswrapper[4802]: I1206 03:44:53.681290 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 06 03:44:53 crc kubenswrapper[4802]: I1206 03:44:53.707674 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 06 03:44:53 crc kubenswrapper[4802]: I1206 03:44:53.791117 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 06 03:44:53 crc kubenswrapper[4802]: I1206 03:44:53.852439 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 06 03:44:53 crc kubenswrapper[4802]: I1206 03:44:53.955792 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 06 03:44:53 crc kubenswrapper[4802]: I1206 03:44:53.985945 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.082447 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.085324 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-666545c866-5zpjj"] Dec 06 03:44:54 crc kubenswrapper[4802]: E1206 03:44:54.085594 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b58e10a8-7605-4863-bc97-91714add5409" containerName="installer" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.085618 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="b58e10a8-7605-4863-bc97-91714add5409" containerName="installer" Dec 06 03:44:54 crc kubenswrapper[4802]: E1206 03:44:54.085636 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="263721d5-fcca-411a-876e-0acfe0863d2d" containerName="oauth-openshift" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.085645 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="263721d5-fcca-411a-876e-0acfe0863d2d" containerName="oauth-openshift" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.085814 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="b58e10a8-7605-4863-bc97-91714add5409" containerName="installer" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.085841 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="263721d5-fcca-411a-876e-0acfe0863d2d" containerName="oauth-openshift" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.086346 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-666545c866-5zpjj" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.089185 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.090124 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.090832 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.090914 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.091098 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.091321 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.091438 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.091454 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.092818 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.093351 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.093901 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.093988 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.099498 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.104003 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.117624 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.180172 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/fa2c8b27-30df-4963-a1a8-a2bb986ad696-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-666545c866-5zpjj\" (UID: \"fa2c8b27-30df-4963-a1a8-a2bb986ad696\") " pod="openshift-authentication/oauth-openshift-666545c866-5zpjj" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.180246 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/fa2c8b27-30df-4963-a1a8-a2bb986ad696-audit-policies\") pod \"oauth-openshift-666545c866-5zpjj\" (UID: \"fa2c8b27-30df-4963-a1a8-a2bb986ad696\") " pod="openshift-authentication/oauth-openshift-666545c866-5zpjj" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.180280 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fa2c8b27-30df-4963-a1a8-a2bb986ad696-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-666545c866-5zpjj\" (UID: \"fa2c8b27-30df-4963-a1a8-a2bb986ad696\") " pod="openshift-authentication/oauth-openshift-666545c866-5zpjj" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.180308 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/fa2c8b27-30df-4963-a1a8-a2bb986ad696-v4-0-config-system-session\") pod \"oauth-openshift-666545c866-5zpjj\" (UID: \"fa2c8b27-30df-4963-a1a8-a2bb986ad696\") " pod="openshift-authentication/oauth-openshift-666545c866-5zpjj" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.180341 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/fa2c8b27-30df-4963-a1a8-a2bb986ad696-v4-0-config-system-router-certs\") pod \"oauth-openshift-666545c866-5zpjj\" (UID: \"fa2c8b27-30df-4963-a1a8-a2bb986ad696\") " pod="openshift-authentication/oauth-openshift-666545c866-5zpjj" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.180364 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/fa2c8b27-30df-4963-a1a8-a2bb986ad696-v4-0-config-system-service-ca\") pod \"oauth-openshift-666545c866-5zpjj\" (UID: \"fa2c8b27-30df-4963-a1a8-a2bb986ad696\") " pod="openshift-authentication/oauth-openshift-666545c866-5zpjj" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.180395 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/fa2c8b27-30df-4963-a1a8-a2bb986ad696-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-666545c866-5zpjj\" (UID: \"fa2c8b27-30df-4963-a1a8-a2bb986ad696\") " pod="openshift-authentication/oauth-openshift-666545c866-5zpjj" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.180416 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/fa2c8b27-30df-4963-a1a8-a2bb986ad696-v4-0-config-system-serving-cert\") pod \"oauth-openshift-666545c866-5zpjj\" (UID: \"fa2c8b27-30df-4963-a1a8-a2bb986ad696\") " pod="openshift-authentication/oauth-openshift-666545c866-5zpjj" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.180480 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/fa2c8b27-30df-4963-a1a8-a2bb986ad696-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-666545c866-5zpjj\" (UID: \"fa2c8b27-30df-4963-a1a8-a2bb986ad696\") " pod="openshift-authentication/oauth-openshift-666545c866-5zpjj" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.180496 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/fa2c8b27-30df-4963-a1a8-a2bb986ad696-v4-0-config-system-cliconfig\") pod \"oauth-openshift-666545c866-5zpjj\" (UID: \"fa2c8b27-30df-4963-a1a8-a2bb986ad696\") " pod="openshift-authentication/oauth-openshift-666545c866-5zpjj" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.180516 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/fa2c8b27-30df-4963-a1a8-a2bb986ad696-v4-0-config-user-template-login\") pod \"oauth-openshift-666545c866-5zpjj\" (UID: \"fa2c8b27-30df-4963-a1a8-a2bb986ad696\") " pod="openshift-authentication/oauth-openshift-666545c866-5zpjj" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.180534 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/fa2c8b27-30df-4963-a1a8-a2bb986ad696-audit-dir\") pod \"oauth-openshift-666545c866-5zpjj\" (UID: \"fa2c8b27-30df-4963-a1a8-a2bb986ad696\") " pod="openshift-authentication/oauth-openshift-666545c866-5zpjj" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.180551 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcslh\" (UniqueName: \"kubernetes.io/projected/fa2c8b27-30df-4963-a1a8-a2bb986ad696-kube-api-access-rcslh\") pod \"oauth-openshift-666545c866-5zpjj\" (UID: \"fa2c8b27-30df-4963-a1a8-a2bb986ad696\") " pod="openshift-authentication/oauth-openshift-666545c866-5zpjj" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.180650 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/fa2c8b27-30df-4963-a1a8-a2bb986ad696-v4-0-config-user-template-error\") pod \"oauth-openshift-666545c866-5zpjj\" (UID: \"fa2c8b27-30df-4963-a1a8-a2bb986ad696\") " pod="openshift-authentication/oauth-openshift-666545c866-5zpjj" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.250008 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.262500 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.282386 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/fa2c8b27-30df-4963-a1a8-a2bb986ad696-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-666545c866-5zpjj\" (UID: \"fa2c8b27-30df-4963-a1a8-a2bb986ad696\") " pod="openshift-authentication/oauth-openshift-666545c866-5zpjj" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.282440 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/fa2c8b27-30df-4963-a1a8-a2bb986ad696-v4-0-config-system-cliconfig\") pod \"oauth-openshift-666545c866-5zpjj\" (UID: \"fa2c8b27-30df-4963-a1a8-a2bb986ad696\") " pod="openshift-authentication/oauth-openshift-666545c866-5zpjj" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.282471 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/fa2c8b27-30df-4963-a1a8-a2bb986ad696-v4-0-config-user-template-login\") pod \"oauth-openshift-666545c866-5zpjj\" (UID: \"fa2c8b27-30df-4963-a1a8-a2bb986ad696\") " pod="openshift-authentication/oauth-openshift-666545c866-5zpjj" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.282496 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/fa2c8b27-30df-4963-a1a8-a2bb986ad696-audit-dir\") pod \"oauth-openshift-666545c866-5zpjj\" (UID: \"fa2c8b27-30df-4963-a1a8-a2bb986ad696\") " pod="openshift-authentication/oauth-openshift-666545c866-5zpjj" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.282518 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcslh\" (UniqueName: \"kubernetes.io/projected/fa2c8b27-30df-4963-a1a8-a2bb986ad696-kube-api-access-rcslh\") pod \"oauth-openshift-666545c866-5zpjj\" (UID: \"fa2c8b27-30df-4963-a1a8-a2bb986ad696\") " pod="openshift-authentication/oauth-openshift-666545c866-5zpjj" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.282540 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/fa2c8b27-30df-4963-a1a8-a2bb986ad696-v4-0-config-user-template-error\") pod \"oauth-openshift-666545c866-5zpjj\" (UID: \"fa2c8b27-30df-4963-a1a8-a2bb986ad696\") " pod="openshift-authentication/oauth-openshift-666545c866-5zpjj" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.282592 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/fa2c8b27-30df-4963-a1a8-a2bb986ad696-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-666545c866-5zpjj\" (UID: \"fa2c8b27-30df-4963-a1a8-a2bb986ad696\") " pod="openshift-authentication/oauth-openshift-666545c866-5zpjj" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.282627 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/fa2c8b27-30df-4963-a1a8-a2bb986ad696-audit-policies\") pod \"oauth-openshift-666545c866-5zpjj\" (UID: \"fa2c8b27-30df-4963-a1a8-a2bb986ad696\") " pod="openshift-authentication/oauth-openshift-666545c866-5zpjj" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.282655 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fa2c8b27-30df-4963-a1a8-a2bb986ad696-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-666545c866-5zpjj\" (UID: \"fa2c8b27-30df-4963-a1a8-a2bb986ad696\") " pod="openshift-authentication/oauth-openshift-666545c866-5zpjj" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.282677 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/fa2c8b27-30df-4963-a1a8-a2bb986ad696-v4-0-config-system-session\") pod \"oauth-openshift-666545c866-5zpjj\" (UID: \"fa2c8b27-30df-4963-a1a8-a2bb986ad696\") " pod="openshift-authentication/oauth-openshift-666545c866-5zpjj" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.282704 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/fa2c8b27-30df-4963-a1a8-a2bb986ad696-v4-0-config-system-router-certs\") pod \"oauth-openshift-666545c866-5zpjj\" (UID: \"fa2c8b27-30df-4963-a1a8-a2bb986ad696\") " pod="openshift-authentication/oauth-openshift-666545c866-5zpjj" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.282726 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/fa2c8b27-30df-4963-a1a8-a2bb986ad696-v4-0-config-system-service-ca\") pod \"oauth-openshift-666545c866-5zpjj\" (UID: \"fa2c8b27-30df-4963-a1a8-a2bb986ad696\") " pod="openshift-authentication/oauth-openshift-666545c866-5zpjj" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.282771 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/fa2c8b27-30df-4963-a1a8-a2bb986ad696-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-666545c866-5zpjj\" (UID: \"fa2c8b27-30df-4963-a1a8-a2bb986ad696\") " pod="openshift-authentication/oauth-openshift-666545c866-5zpjj" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.282795 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/fa2c8b27-30df-4963-a1a8-a2bb986ad696-v4-0-config-system-serving-cert\") pod \"oauth-openshift-666545c866-5zpjj\" (UID: \"fa2c8b27-30df-4963-a1a8-a2bb986ad696\") " pod="openshift-authentication/oauth-openshift-666545c866-5zpjj" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.284377 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/fa2c8b27-30df-4963-a1a8-a2bb986ad696-v4-0-config-system-cliconfig\") pod \"oauth-openshift-666545c866-5zpjj\" (UID: \"fa2c8b27-30df-4963-a1a8-a2bb986ad696\") " pod="openshift-authentication/oauth-openshift-666545c866-5zpjj" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.284900 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/fa2c8b27-30df-4963-a1a8-a2bb986ad696-audit-policies\") pod \"oauth-openshift-666545c866-5zpjj\" (UID: \"fa2c8b27-30df-4963-a1a8-a2bb986ad696\") " pod="openshift-authentication/oauth-openshift-666545c866-5zpjj" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.285983 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/fa2c8b27-30df-4963-a1a8-a2bb986ad696-audit-dir\") pod \"oauth-openshift-666545c866-5zpjj\" (UID: \"fa2c8b27-30df-4963-a1a8-a2bb986ad696\") " pod="openshift-authentication/oauth-openshift-666545c866-5zpjj" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.285998 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fa2c8b27-30df-4963-a1a8-a2bb986ad696-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-666545c866-5zpjj\" (UID: \"fa2c8b27-30df-4963-a1a8-a2bb986ad696\") " pod="openshift-authentication/oauth-openshift-666545c866-5zpjj" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.286572 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/fa2c8b27-30df-4963-a1a8-a2bb986ad696-v4-0-config-system-service-ca\") pod \"oauth-openshift-666545c866-5zpjj\" (UID: \"fa2c8b27-30df-4963-a1a8-a2bb986ad696\") " pod="openshift-authentication/oauth-openshift-666545c866-5zpjj" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.288397 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/fa2c8b27-30df-4963-a1a8-a2bb986ad696-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-666545c866-5zpjj\" (UID: \"fa2c8b27-30df-4963-a1a8-a2bb986ad696\") " pod="openshift-authentication/oauth-openshift-666545c866-5zpjj" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.289127 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/fa2c8b27-30df-4963-a1a8-a2bb986ad696-v4-0-config-system-router-certs\") pod \"oauth-openshift-666545c866-5zpjj\" (UID: \"fa2c8b27-30df-4963-a1a8-a2bb986ad696\") " pod="openshift-authentication/oauth-openshift-666545c866-5zpjj" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.289777 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/fa2c8b27-30df-4963-a1a8-a2bb986ad696-v4-0-config-system-serving-cert\") pod \"oauth-openshift-666545c866-5zpjj\" (UID: \"fa2c8b27-30df-4963-a1a8-a2bb986ad696\") " pod="openshift-authentication/oauth-openshift-666545c866-5zpjj" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.290585 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/fa2c8b27-30df-4963-a1a8-a2bb986ad696-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-666545c866-5zpjj\" (UID: \"fa2c8b27-30df-4963-a1a8-a2bb986ad696\") " pod="openshift-authentication/oauth-openshift-666545c866-5zpjj" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.291113 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/fa2c8b27-30df-4963-a1a8-a2bb986ad696-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-666545c866-5zpjj\" (UID: \"fa2c8b27-30df-4963-a1a8-a2bb986ad696\") " pod="openshift-authentication/oauth-openshift-666545c866-5zpjj" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.291724 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/fa2c8b27-30df-4963-a1a8-a2bb986ad696-v4-0-config-user-template-login\") pod \"oauth-openshift-666545c866-5zpjj\" (UID: \"fa2c8b27-30df-4963-a1a8-a2bb986ad696\") " pod="openshift-authentication/oauth-openshift-666545c866-5zpjj" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.291971 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/fa2c8b27-30df-4963-a1a8-a2bb986ad696-v4-0-config-user-template-error\") pod \"oauth-openshift-666545c866-5zpjj\" (UID: \"fa2c8b27-30df-4963-a1a8-a2bb986ad696\") " pod="openshift-authentication/oauth-openshift-666545c866-5zpjj" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.300072 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/fa2c8b27-30df-4963-a1a8-a2bb986ad696-v4-0-config-system-session\") pod \"oauth-openshift-666545c866-5zpjj\" (UID: \"fa2c8b27-30df-4963-a1a8-a2bb986ad696\") " pod="openshift-authentication/oauth-openshift-666545c866-5zpjj" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.303673 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rcslh\" (UniqueName: \"kubernetes.io/projected/fa2c8b27-30df-4963-a1a8-a2bb986ad696-kube-api-access-rcslh\") pod \"oauth-openshift-666545c866-5zpjj\" (UID: \"fa2c8b27-30df-4963-a1a8-a2bb986ad696\") " pod="openshift-authentication/oauth-openshift-666545c866-5zpjj" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.366242 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.366413 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.373058 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.402917 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-666545c866-5zpjj" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.428842 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.444305 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.446108 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-666545c866-5zpjj"] Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.610170 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.874632 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.904384 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 06 03:44:54 crc kubenswrapper[4802]: I1206 03:44:54.923996 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 06 03:44:55 crc kubenswrapper[4802]: I1206 03:44:55.099420 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 06 03:44:55 crc kubenswrapper[4802]: I1206 03:44:55.326808 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-666545c866-5zpjj"] Dec 06 03:44:55 crc kubenswrapper[4802]: I1206 03:44:55.333030 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 06 03:44:55 crc kubenswrapper[4802]: I1206 03:44:55.355419 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 06 03:44:55 crc kubenswrapper[4802]: I1206 03:44:55.376617 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 06 03:44:55 crc kubenswrapper[4802]: I1206 03:44:55.384778 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 06 03:44:55 crc kubenswrapper[4802]: I1206 03:44:55.398225 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 06 03:44:55 crc kubenswrapper[4802]: I1206 03:44:55.540595 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 06 03:44:55 crc kubenswrapper[4802]: I1206 03:44:55.833828 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 06 03:44:55 crc kubenswrapper[4802]: I1206 03:44:55.930604 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-666545c866-5zpjj" event={"ID":"fa2c8b27-30df-4963-a1a8-a2bb986ad696","Type":"ContainerStarted","Data":"98ac55e109aebe13e472b5f2230aa1e434d92d336bf4074779bea73d43a4a295"} Dec 06 03:44:55 crc kubenswrapper[4802]: I1206 03:44:55.930946 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-666545c866-5zpjj" Dec 06 03:44:55 crc kubenswrapper[4802]: I1206 03:44:55.931092 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-666545c866-5zpjj" event={"ID":"fa2c8b27-30df-4963-a1a8-a2bb986ad696","Type":"ContainerStarted","Data":"d765721cd9b3cc014528411abc2e8b61bf3b5d2d26950bc8b1a4f02171a5fdaf"} Dec 06 03:44:55 crc kubenswrapper[4802]: I1206 03:44:55.966390 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-666545c866-5zpjj" podStartSLOduration=54.966362561 podStartE2EDuration="54.966362561s" podCreationTimestamp="2025-12-06 03:44:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:44:55.96189905 +0000 UTC m=+288.833808242" watchObservedRunningTime="2025-12-06 03:44:55.966362561 +0000 UTC m=+288.838271723" Dec 06 03:44:56 crc kubenswrapper[4802]: I1206 03:44:56.040068 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 06 03:44:56 crc kubenswrapper[4802]: I1206 03:44:56.103560 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-666545c866-5zpjj" Dec 06 03:44:56 crc kubenswrapper[4802]: I1206 03:44:56.107675 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 06 03:44:56 crc kubenswrapper[4802]: I1206 03:44:56.164803 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 06 03:44:56 crc kubenswrapper[4802]: I1206 03:44:56.172284 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 06 03:44:56 crc kubenswrapper[4802]: I1206 03:44:56.176261 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 06 03:44:56 crc kubenswrapper[4802]: I1206 03:44:56.223025 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 06 03:44:56 crc kubenswrapper[4802]: I1206 03:44:56.358327 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 06 03:44:56 crc kubenswrapper[4802]: I1206 03:44:56.402529 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 06 03:44:56 crc kubenswrapper[4802]: I1206 03:44:56.461157 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 06 03:44:56 crc kubenswrapper[4802]: I1206 03:44:56.542462 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 06 03:44:56 crc kubenswrapper[4802]: I1206 03:44:56.610642 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 06 03:44:56 crc kubenswrapper[4802]: I1206 03:44:56.662674 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 06 03:44:56 crc kubenswrapper[4802]: I1206 03:44:56.889625 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 06 03:44:56 crc kubenswrapper[4802]: I1206 03:44:56.938969 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 06 03:45:00 crc kubenswrapper[4802]: I1206 03:45:00.194500 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416545-9z9qr"] Dec 06 03:45:00 crc kubenswrapper[4802]: I1206 03:45:00.195969 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416545-9z9qr" Dec 06 03:45:00 crc kubenswrapper[4802]: I1206 03:45:00.198103 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 03:45:00 crc kubenswrapper[4802]: I1206 03:45:00.198202 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 03:45:00 crc kubenswrapper[4802]: I1206 03:45:00.199581 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416545-9z9qr"] Dec 06 03:45:00 crc kubenswrapper[4802]: I1206 03:45:00.280105 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/db461f4d-dd3b-48ff-9d07-f686dce48122-secret-volume\") pod \"collect-profiles-29416545-9z9qr\" (UID: \"db461f4d-dd3b-48ff-9d07-f686dce48122\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416545-9z9qr" Dec 06 03:45:00 crc kubenswrapper[4802]: I1206 03:45:00.280196 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pg782\" (UniqueName: \"kubernetes.io/projected/db461f4d-dd3b-48ff-9d07-f686dce48122-kube-api-access-pg782\") pod \"collect-profiles-29416545-9z9qr\" (UID: \"db461f4d-dd3b-48ff-9d07-f686dce48122\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416545-9z9qr" Dec 06 03:45:00 crc kubenswrapper[4802]: I1206 03:45:00.280296 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/db461f4d-dd3b-48ff-9d07-f686dce48122-config-volume\") pod \"collect-profiles-29416545-9z9qr\" (UID: \"db461f4d-dd3b-48ff-9d07-f686dce48122\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416545-9z9qr" Dec 06 03:45:00 crc kubenswrapper[4802]: I1206 03:45:00.381284 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/db461f4d-dd3b-48ff-9d07-f686dce48122-secret-volume\") pod \"collect-profiles-29416545-9z9qr\" (UID: \"db461f4d-dd3b-48ff-9d07-f686dce48122\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416545-9z9qr" Dec 06 03:45:00 crc kubenswrapper[4802]: I1206 03:45:00.381620 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pg782\" (UniqueName: \"kubernetes.io/projected/db461f4d-dd3b-48ff-9d07-f686dce48122-kube-api-access-pg782\") pod \"collect-profiles-29416545-9z9qr\" (UID: \"db461f4d-dd3b-48ff-9d07-f686dce48122\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416545-9z9qr" Dec 06 03:45:00 crc kubenswrapper[4802]: I1206 03:45:00.381815 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/db461f4d-dd3b-48ff-9d07-f686dce48122-config-volume\") pod \"collect-profiles-29416545-9z9qr\" (UID: \"db461f4d-dd3b-48ff-9d07-f686dce48122\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416545-9z9qr" Dec 06 03:45:00 crc kubenswrapper[4802]: I1206 03:45:00.383248 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/db461f4d-dd3b-48ff-9d07-f686dce48122-config-volume\") pod \"collect-profiles-29416545-9z9qr\" (UID: \"db461f4d-dd3b-48ff-9d07-f686dce48122\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416545-9z9qr" Dec 06 03:45:00 crc kubenswrapper[4802]: I1206 03:45:00.392067 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/db461f4d-dd3b-48ff-9d07-f686dce48122-secret-volume\") pod \"collect-profiles-29416545-9z9qr\" (UID: \"db461f4d-dd3b-48ff-9d07-f686dce48122\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416545-9z9qr" Dec 06 03:45:00 crc kubenswrapper[4802]: I1206 03:45:00.409993 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pg782\" (UniqueName: \"kubernetes.io/projected/db461f4d-dd3b-48ff-9d07-f686dce48122-kube-api-access-pg782\") pod \"collect-profiles-29416545-9z9qr\" (UID: \"db461f4d-dd3b-48ff-9d07-f686dce48122\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416545-9z9qr" Dec 06 03:45:00 crc kubenswrapper[4802]: I1206 03:45:00.511414 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416545-9z9qr" Dec 06 03:45:00 crc kubenswrapper[4802]: I1206 03:45:00.916852 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416545-9z9qr"] Dec 06 03:45:00 crc kubenswrapper[4802]: W1206 03:45:00.921429 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddb461f4d_dd3b_48ff_9d07_f686dce48122.slice/crio-90012456ebcb5f2f4a35a9cf000adea3ee48934c2bd85f4ee0e37bae5d344907 WatchSource:0}: Error finding container 90012456ebcb5f2f4a35a9cf000adea3ee48934c2bd85f4ee0e37bae5d344907: Status 404 returned error can't find the container with id 90012456ebcb5f2f4a35a9cf000adea3ee48934c2bd85f4ee0e37bae5d344907 Dec 06 03:45:00 crc kubenswrapper[4802]: I1206 03:45:00.974536 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416545-9z9qr" event={"ID":"db461f4d-dd3b-48ff-9d07-f686dce48122","Type":"ContainerStarted","Data":"90012456ebcb5f2f4a35a9cf000adea3ee48934c2bd85f4ee0e37bae5d344907"} Dec 06 03:45:01 crc kubenswrapper[4802]: I1206 03:45:01.981950 4802 generic.go:334] "Generic (PLEG): container finished" podID="db461f4d-dd3b-48ff-9d07-f686dce48122" containerID="5e5b040efcfd42d0060dedb8ab5af64340411a989af338fdb19b891bf7d03056" exitCode=0 Dec 06 03:45:01 crc kubenswrapper[4802]: I1206 03:45:01.982041 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416545-9z9qr" event={"ID":"db461f4d-dd3b-48ff-9d07-f686dce48122","Type":"ContainerDied","Data":"5e5b040efcfd42d0060dedb8ab5af64340411a989af338fdb19b891bf7d03056"} Dec 06 03:45:03 crc kubenswrapper[4802]: I1206 03:45:03.312526 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416545-9z9qr" Dec 06 03:45:03 crc kubenswrapper[4802]: I1206 03:45:03.422009 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/db461f4d-dd3b-48ff-9d07-f686dce48122-config-volume\") pod \"db461f4d-dd3b-48ff-9d07-f686dce48122\" (UID: \"db461f4d-dd3b-48ff-9d07-f686dce48122\") " Dec 06 03:45:03 crc kubenswrapper[4802]: I1206 03:45:03.422108 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pg782\" (UniqueName: \"kubernetes.io/projected/db461f4d-dd3b-48ff-9d07-f686dce48122-kube-api-access-pg782\") pod \"db461f4d-dd3b-48ff-9d07-f686dce48122\" (UID: \"db461f4d-dd3b-48ff-9d07-f686dce48122\") " Dec 06 03:45:03 crc kubenswrapper[4802]: I1206 03:45:03.422169 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/db461f4d-dd3b-48ff-9d07-f686dce48122-secret-volume\") pod \"db461f4d-dd3b-48ff-9d07-f686dce48122\" (UID: \"db461f4d-dd3b-48ff-9d07-f686dce48122\") " Dec 06 03:45:03 crc kubenswrapper[4802]: I1206 03:45:03.423381 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db461f4d-dd3b-48ff-9d07-f686dce48122-config-volume" (OuterVolumeSpecName: "config-volume") pod "db461f4d-dd3b-48ff-9d07-f686dce48122" (UID: "db461f4d-dd3b-48ff-9d07-f686dce48122"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:45:03 crc kubenswrapper[4802]: I1206 03:45:03.428654 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db461f4d-dd3b-48ff-9d07-f686dce48122-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "db461f4d-dd3b-48ff-9d07-f686dce48122" (UID: "db461f4d-dd3b-48ff-9d07-f686dce48122"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:45:03 crc kubenswrapper[4802]: I1206 03:45:03.428823 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db461f4d-dd3b-48ff-9d07-f686dce48122-kube-api-access-pg782" (OuterVolumeSpecName: "kube-api-access-pg782") pod "db461f4d-dd3b-48ff-9d07-f686dce48122" (UID: "db461f4d-dd3b-48ff-9d07-f686dce48122"). InnerVolumeSpecName "kube-api-access-pg782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:45:03 crc kubenswrapper[4802]: I1206 03:45:03.523930 4802 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/db461f4d-dd3b-48ff-9d07-f686dce48122-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 03:45:03 crc kubenswrapper[4802]: I1206 03:45:03.524014 4802 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/db461f4d-dd3b-48ff-9d07-f686dce48122-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 03:45:03 crc kubenswrapper[4802]: I1206 03:45:03.524035 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pg782\" (UniqueName: \"kubernetes.io/projected/db461f4d-dd3b-48ff-9d07-f686dce48122-kube-api-access-pg782\") on node \"crc\" DevicePath \"\"" Dec 06 03:45:03 crc kubenswrapper[4802]: I1206 03:45:03.999075 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416545-9z9qr" event={"ID":"db461f4d-dd3b-48ff-9d07-f686dce48122","Type":"ContainerDied","Data":"90012456ebcb5f2f4a35a9cf000adea3ee48934c2bd85f4ee0e37bae5d344907"} Dec 06 03:45:03 crc kubenswrapper[4802]: I1206 03:45:03.999137 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="90012456ebcb5f2f4a35a9cf000adea3ee48934c2bd85f4ee0e37bae5d344907" Dec 06 03:45:03 crc kubenswrapper[4802]: I1206 03:45:03.999176 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416545-9z9qr" Dec 06 03:45:04 crc kubenswrapper[4802]: I1206 03:45:04.237637 4802 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 06 03:45:04 crc kubenswrapper[4802]: I1206 03:45:04.238245 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://ab648d5b2f1d8c98db0240df2881ef08dd4a7fac79eab166365e8f4c66285e3d" gracePeriod=5 Dec 06 03:45:05 crc kubenswrapper[4802]: I1206 03:45:05.856770 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 06 03:45:07 crc kubenswrapper[4802]: I1206 03:45:07.297450 4802 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Dec 06 03:45:09 crc kubenswrapper[4802]: I1206 03:45:09.843815 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 06 03:45:09 crc kubenswrapper[4802]: I1206 03:45:09.844457 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 03:45:09 crc kubenswrapper[4802]: I1206 03:45:09.933046 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 06 03:45:09 crc kubenswrapper[4802]: I1206 03:45:09.933128 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 06 03:45:09 crc kubenswrapper[4802]: I1206 03:45:09.933159 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:45:09 crc kubenswrapper[4802]: I1206 03:45:09.933208 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 06 03:45:09 crc kubenswrapper[4802]: I1206 03:45:09.933240 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:45:09 crc kubenswrapper[4802]: I1206 03:45:09.933280 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 06 03:45:09 crc kubenswrapper[4802]: I1206 03:45:09.933373 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 06 03:45:09 crc kubenswrapper[4802]: I1206 03:45:09.933455 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:45:09 crc kubenswrapper[4802]: I1206 03:45:09.933614 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:45:09 crc kubenswrapper[4802]: I1206 03:45:09.933799 4802 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 06 03:45:09 crc kubenswrapper[4802]: I1206 03:45:09.933822 4802 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 06 03:45:09 crc kubenswrapper[4802]: I1206 03:45:09.933840 4802 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 06 03:45:09 crc kubenswrapper[4802]: I1206 03:45:09.933857 4802 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 06 03:45:09 crc kubenswrapper[4802]: I1206 03:45:09.947059 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:45:10 crc kubenswrapper[4802]: I1206 03:45:10.034970 4802 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 06 03:45:10 crc kubenswrapper[4802]: I1206 03:45:10.040389 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 06 03:45:10 crc kubenswrapper[4802]: I1206 03:45:10.040480 4802 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="ab648d5b2f1d8c98db0240df2881ef08dd4a7fac79eab166365e8f4c66285e3d" exitCode=137 Dec 06 03:45:10 crc kubenswrapper[4802]: I1206 03:45:10.040547 4802 scope.go:117] "RemoveContainer" containerID="ab648d5b2f1d8c98db0240df2881ef08dd4a7fac79eab166365e8f4c66285e3d" Dec 06 03:45:10 crc kubenswrapper[4802]: I1206 03:45:10.040596 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 03:45:10 crc kubenswrapper[4802]: I1206 03:45:10.068152 4802 scope.go:117] "RemoveContainer" containerID="ab648d5b2f1d8c98db0240df2881ef08dd4a7fac79eab166365e8f4c66285e3d" Dec 06 03:45:10 crc kubenswrapper[4802]: E1206 03:45:10.068850 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab648d5b2f1d8c98db0240df2881ef08dd4a7fac79eab166365e8f4c66285e3d\": container with ID starting with ab648d5b2f1d8c98db0240df2881ef08dd4a7fac79eab166365e8f4c66285e3d not found: ID does not exist" containerID="ab648d5b2f1d8c98db0240df2881ef08dd4a7fac79eab166365e8f4c66285e3d" Dec 06 03:45:10 crc kubenswrapper[4802]: I1206 03:45:10.068931 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab648d5b2f1d8c98db0240df2881ef08dd4a7fac79eab166365e8f4c66285e3d"} err="failed to get container status \"ab648d5b2f1d8c98db0240df2881ef08dd4a7fac79eab166365e8f4c66285e3d\": rpc error: code = NotFound desc = could not find container \"ab648d5b2f1d8c98db0240df2881ef08dd4a7fac79eab166365e8f4c66285e3d\": container with ID starting with ab648d5b2f1d8c98db0240df2881ef08dd4a7fac79eab166365e8f4c66285e3d not found: ID does not exist" Dec 06 03:45:11 crc kubenswrapper[4802]: I1206 03:45:11.138349 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 06 03:45:11 crc kubenswrapper[4802]: I1206 03:45:11.462251 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 06 03:45:11 crc kubenswrapper[4802]: I1206 03:45:11.462718 4802 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Dec 06 03:45:11 crc kubenswrapper[4802]: I1206 03:45:11.476820 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 06 03:45:11 crc kubenswrapper[4802]: I1206 03:45:11.476921 4802 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="eb1d012e-259a-4030-8fd3-aa574698f288" Dec 06 03:45:11 crc kubenswrapper[4802]: I1206 03:45:11.482895 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 06 03:45:11 crc kubenswrapper[4802]: I1206 03:45:11.482953 4802 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="eb1d012e-259a-4030-8fd3-aa574698f288" Dec 06 03:45:11 crc kubenswrapper[4802]: I1206 03:45:11.633793 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 06 03:45:13 crc kubenswrapper[4802]: I1206 03:45:13.687843 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 06 03:45:14 crc kubenswrapper[4802]: I1206 03:45:14.034192 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 06 03:45:15 crc kubenswrapper[4802]: I1206 03:45:15.158010 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 06 03:45:15 crc kubenswrapper[4802]: I1206 03:45:15.219691 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 06 03:45:17 crc kubenswrapper[4802]: I1206 03:45:17.507498 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 06 03:45:18 crc kubenswrapper[4802]: I1206 03:45:18.618373 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-cgpdh"] Dec 06 03:45:18 crc kubenswrapper[4802]: I1206 03:45:18.618575 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-cgpdh" podUID="a68fc0dd-997c-4d12-9a20-5345dfe98a41" containerName="controller-manager" containerID="cri-o://f9fc0a9a62f254e815cc5ad499da3ac6e027b333f613cfd4d44e4a88fe48a39a" gracePeriod=30 Dec 06 03:45:18 crc kubenswrapper[4802]: I1206 03:45:18.716678 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-9v855"] Dec 06 03:45:18 crc kubenswrapper[4802]: I1206 03:45:18.717226 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9v855" podUID="db213df1-bfa2-4edd-a3fc-246e35e20585" containerName="route-controller-manager" containerID="cri-o://941b27608e31e3ffeb9e31e27d6baeca5f2247a255b2a9d98a41883c2d37c3b0" gracePeriod=30 Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.032964 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9v855" Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.099001 4802 generic.go:334] "Generic (PLEG): container finished" podID="db213df1-bfa2-4edd-a3fc-246e35e20585" containerID="941b27608e31e3ffeb9e31e27d6baeca5f2247a255b2a9d98a41883c2d37c3b0" exitCode=0 Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.099090 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9v855" Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.099499 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9v855" event={"ID":"db213df1-bfa2-4edd-a3fc-246e35e20585","Type":"ContainerDied","Data":"941b27608e31e3ffeb9e31e27d6baeca5f2247a255b2a9d98a41883c2d37c3b0"} Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.099527 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9v855" event={"ID":"db213df1-bfa2-4edd-a3fc-246e35e20585","Type":"ContainerDied","Data":"01a3cf31ddd14c1ad48b68874e13595b105e4c77bfdcd50de7a9d301d58afad2"} Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.099546 4802 scope.go:117] "RemoveContainer" containerID="941b27608e31e3ffeb9e31e27d6baeca5f2247a255b2a9d98a41883c2d37c3b0" Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.100917 4802 generic.go:334] "Generic (PLEG): container finished" podID="a68fc0dd-997c-4d12-9a20-5345dfe98a41" containerID="f9fc0a9a62f254e815cc5ad499da3ac6e027b333f613cfd4d44e4a88fe48a39a" exitCode=0 Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.100935 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-cgpdh" event={"ID":"a68fc0dd-997c-4d12-9a20-5345dfe98a41","Type":"ContainerDied","Data":"f9fc0a9a62f254e815cc5ad499da3ac6e027b333f613cfd4d44e4a88fe48a39a"} Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.120215 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.126577 4802 scope.go:117] "RemoveContainer" containerID="941b27608e31e3ffeb9e31e27d6baeca5f2247a255b2a9d98a41883c2d37c3b0" Dec 06 03:45:19 crc kubenswrapper[4802]: E1206 03:45:19.127035 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"941b27608e31e3ffeb9e31e27d6baeca5f2247a255b2a9d98a41883c2d37c3b0\": container with ID starting with 941b27608e31e3ffeb9e31e27d6baeca5f2247a255b2a9d98a41883c2d37c3b0 not found: ID does not exist" containerID="941b27608e31e3ffeb9e31e27d6baeca5f2247a255b2a9d98a41883c2d37c3b0" Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.127077 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"941b27608e31e3ffeb9e31e27d6baeca5f2247a255b2a9d98a41883c2d37c3b0"} err="failed to get container status \"941b27608e31e3ffeb9e31e27d6baeca5f2247a255b2a9d98a41883c2d37c3b0\": rpc error: code = NotFound desc = could not find container \"941b27608e31e3ffeb9e31e27d6baeca5f2247a255b2a9d98a41883c2d37c3b0\": container with ID starting with 941b27608e31e3ffeb9e31e27d6baeca5f2247a255b2a9d98a41883c2d37c3b0 not found: ID does not exist" Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.159629 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/db213df1-bfa2-4edd-a3fc-246e35e20585-client-ca\") pod \"db213df1-bfa2-4edd-a3fc-246e35e20585\" (UID: \"db213df1-bfa2-4edd-a3fc-246e35e20585\") " Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.159723 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/db213df1-bfa2-4edd-a3fc-246e35e20585-serving-cert\") pod \"db213df1-bfa2-4edd-a3fc-246e35e20585\" (UID: \"db213df1-bfa2-4edd-a3fc-246e35e20585\") " Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.159782 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fpqvr\" (UniqueName: \"kubernetes.io/projected/db213df1-bfa2-4edd-a3fc-246e35e20585-kube-api-access-fpqvr\") pod \"db213df1-bfa2-4edd-a3fc-246e35e20585\" (UID: \"db213df1-bfa2-4edd-a3fc-246e35e20585\") " Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.159886 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db213df1-bfa2-4edd-a3fc-246e35e20585-config\") pod \"db213df1-bfa2-4edd-a3fc-246e35e20585\" (UID: \"db213df1-bfa2-4edd-a3fc-246e35e20585\") " Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.160720 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db213df1-bfa2-4edd-a3fc-246e35e20585-client-ca" (OuterVolumeSpecName: "client-ca") pod "db213df1-bfa2-4edd-a3fc-246e35e20585" (UID: "db213df1-bfa2-4edd-a3fc-246e35e20585"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.160727 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db213df1-bfa2-4edd-a3fc-246e35e20585-config" (OuterVolumeSpecName: "config") pod "db213df1-bfa2-4edd-a3fc-246e35e20585" (UID: "db213df1-bfa2-4edd-a3fc-246e35e20585"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.160943 4802 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/db213df1-bfa2-4edd-a3fc-246e35e20585-client-ca\") on node \"crc\" DevicePath \"\"" Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.160963 4802 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db213df1-bfa2-4edd-a3fc-246e35e20585-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.167834 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db213df1-bfa2-4edd-a3fc-246e35e20585-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "db213df1-bfa2-4edd-a3fc-246e35e20585" (UID: "db213df1-bfa2-4edd-a3fc-246e35e20585"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.170621 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db213df1-bfa2-4edd-a3fc-246e35e20585-kube-api-access-fpqvr" (OuterVolumeSpecName: "kube-api-access-fpqvr") pod "db213df1-bfa2-4edd-a3fc-246e35e20585" (UID: "db213df1-bfa2-4edd-a3fc-246e35e20585"). InnerVolumeSpecName "kube-api-access-fpqvr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.262279 4802 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/db213df1-bfa2-4edd-a3fc-246e35e20585-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.262316 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fpqvr\" (UniqueName: \"kubernetes.io/projected/db213df1-bfa2-4edd-a3fc-246e35e20585-kube-api-access-fpqvr\") on node \"crc\" DevicePath \"\"" Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.436435 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-9v855"] Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.439206 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-9v855"] Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.459100 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-cgpdh" Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.460026 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db213df1-bfa2-4edd-a3fc-246e35e20585" path="/var/lib/kubelet/pods/db213df1-bfa2-4edd-a3fc-246e35e20585/volumes" Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.569139 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a68fc0dd-997c-4d12-9a20-5345dfe98a41-client-ca\") pod \"a68fc0dd-997c-4d12-9a20-5345dfe98a41\" (UID: \"a68fc0dd-997c-4d12-9a20-5345dfe98a41\") " Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.569211 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a68fc0dd-997c-4d12-9a20-5345dfe98a41-config\") pod \"a68fc0dd-997c-4d12-9a20-5345dfe98a41\" (UID: \"a68fc0dd-997c-4d12-9a20-5345dfe98a41\") " Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.569229 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a68fc0dd-997c-4d12-9a20-5345dfe98a41-serving-cert\") pod \"a68fc0dd-997c-4d12-9a20-5345dfe98a41\" (UID: \"a68fc0dd-997c-4d12-9a20-5345dfe98a41\") " Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.569253 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ccnj9\" (UniqueName: \"kubernetes.io/projected/a68fc0dd-997c-4d12-9a20-5345dfe98a41-kube-api-access-ccnj9\") pod \"a68fc0dd-997c-4d12-9a20-5345dfe98a41\" (UID: \"a68fc0dd-997c-4d12-9a20-5345dfe98a41\") " Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.569299 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a68fc0dd-997c-4d12-9a20-5345dfe98a41-proxy-ca-bundles\") pod \"a68fc0dd-997c-4d12-9a20-5345dfe98a41\" (UID: \"a68fc0dd-997c-4d12-9a20-5345dfe98a41\") " Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.570593 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a68fc0dd-997c-4d12-9a20-5345dfe98a41-client-ca" (OuterVolumeSpecName: "client-ca") pod "a68fc0dd-997c-4d12-9a20-5345dfe98a41" (UID: "a68fc0dd-997c-4d12-9a20-5345dfe98a41"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.570605 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a68fc0dd-997c-4d12-9a20-5345dfe98a41-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "a68fc0dd-997c-4d12-9a20-5345dfe98a41" (UID: "a68fc0dd-997c-4d12-9a20-5345dfe98a41"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.570704 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a68fc0dd-997c-4d12-9a20-5345dfe98a41-config" (OuterVolumeSpecName: "config") pod "a68fc0dd-997c-4d12-9a20-5345dfe98a41" (UID: "a68fc0dd-997c-4d12-9a20-5345dfe98a41"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.573355 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a68fc0dd-997c-4d12-9a20-5345dfe98a41-kube-api-access-ccnj9" (OuterVolumeSpecName: "kube-api-access-ccnj9") pod "a68fc0dd-997c-4d12-9a20-5345dfe98a41" (UID: "a68fc0dd-997c-4d12-9a20-5345dfe98a41"). InnerVolumeSpecName "kube-api-access-ccnj9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.573392 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a68fc0dd-997c-4d12-9a20-5345dfe98a41-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "a68fc0dd-997c-4d12-9a20-5345dfe98a41" (UID: "a68fc0dd-997c-4d12-9a20-5345dfe98a41"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.671074 4802 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a68fc0dd-997c-4d12-9a20-5345dfe98a41-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.671944 4802 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a68fc0dd-997c-4d12-9a20-5345dfe98a41-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.672044 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ccnj9\" (UniqueName: \"kubernetes.io/projected/a68fc0dd-997c-4d12-9a20-5345dfe98a41-kube-api-access-ccnj9\") on node \"crc\" DevicePath \"\"" Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.672114 4802 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a68fc0dd-997c-4d12-9a20-5345dfe98a41-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.672186 4802 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a68fc0dd-997c-4d12-9a20-5345dfe98a41-client-ca\") on node \"crc\" DevicePath \"\"" Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.889211 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-ff6dfd758-fh9rf"] Dec 06 03:45:19 crc kubenswrapper[4802]: E1206 03:45:19.890084 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db213df1-bfa2-4edd-a3fc-246e35e20585" containerName="route-controller-manager" Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.890259 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="db213df1-bfa2-4edd-a3fc-246e35e20585" containerName="route-controller-manager" Dec 06 03:45:19 crc kubenswrapper[4802]: E1206 03:45:19.890479 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.890606 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 06 03:45:19 crc kubenswrapper[4802]: E1206 03:45:19.890801 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a68fc0dd-997c-4d12-9a20-5345dfe98a41" containerName="controller-manager" Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.890983 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="a68fc0dd-997c-4d12-9a20-5345dfe98a41" containerName="controller-manager" Dec 06 03:45:19 crc kubenswrapper[4802]: E1206 03:45:19.891145 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db461f4d-dd3b-48ff-9d07-f686dce48122" containerName="collect-profiles" Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.891277 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="db461f4d-dd3b-48ff-9d07-f686dce48122" containerName="collect-profiles" Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.891651 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.891846 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="db213df1-bfa2-4edd-a3fc-246e35e20585" containerName="route-controller-manager" Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.892003 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="db461f4d-dd3b-48ff-9d07-f686dce48122" containerName="collect-profiles" Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.892139 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="a68fc0dd-997c-4d12-9a20-5345dfe98a41" containerName="controller-manager" Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.892947 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-ff6dfd758-fh9rf" Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.896494 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-575ccfcfc8-ct5bx"] Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.896732 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.897396 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-575ccfcfc8-ct5bx" Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.902605 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.902888 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.902959 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.903089 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.903113 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.907968 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-ff6dfd758-fh9rf"] Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.932866 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-575ccfcfc8-ct5bx"] Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.976517 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nv2gv\" (UniqueName: \"kubernetes.io/projected/f053daeb-b5e4-444e-a580-89a049e64beb-kube-api-access-nv2gv\") pod \"controller-manager-575ccfcfc8-ct5bx\" (UID: \"f053daeb-b5e4-444e-a580-89a049e64beb\") " pod="openshift-controller-manager/controller-manager-575ccfcfc8-ct5bx" Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.976596 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b8b92a83-a81a-4b3c-875a-65bd70b86a12-serving-cert\") pod \"route-controller-manager-ff6dfd758-fh9rf\" (UID: \"b8b92a83-a81a-4b3c-875a-65bd70b86a12\") " pod="openshift-route-controller-manager/route-controller-manager-ff6dfd758-fh9rf" Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.976689 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b8b92a83-a81a-4b3c-875a-65bd70b86a12-client-ca\") pod \"route-controller-manager-ff6dfd758-fh9rf\" (UID: \"b8b92a83-a81a-4b3c-875a-65bd70b86a12\") " pod="openshift-route-controller-manager/route-controller-manager-ff6dfd758-fh9rf" Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.976775 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b8b92a83-a81a-4b3c-875a-65bd70b86a12-config\") pod \"route-controller-manager-ff6dfd758-fh9rf\" (UID: \"b8b92a83-a81a-4b3c-875a-65bd70b86a12\") " pod="openshift-route-controller-manager/route-controller-manager-ff6dfd758-fh9rf" Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.976809 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f053daeb-b5e4-444e-a580-89a049e64beb-config\") pod \"controller-manager-575ccfcfc8-ct5bx\" (UID: \"f053daeb-b5e4-444e-a580-89a049e64beb\") " pod="openshift-controller-manager/controller-manager-575ccfcfc8-ct5bx" Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.976842 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f053daeb-b5e4-444e-a580-89a049e64beb-serving-cert\") pod \"controller-manager-575ccfcfc8-ct5bx\" (UID: \"f053daeb-b5e4-444e-a580-89a049e64beb\") " pod="openshift-controller-manager/controller-manager-575ccfcfc8-ct5bx" Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.976867 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f053daeb-b5e4-444e-a580-89a049e64beb-proxy-ca-bundles\") pod \"controller-manager-575ccfcfc8-ct5bx\" (UID: \"f053daeb-b5e4-444e-a580-89a049e64beb\") " pod="openshift-controller-manager/controller-manager-575ccfcfc8-ct5bx" Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.976902 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzmlz\" (UniqueName: \"kubernetes.io/projected/b8b92a83-a81a-4b3c-875a-65bd70b86a12-kube-api-access-fzmlz\") pod \"route-controller-manager-ff6dfd758-fh9rf\" (UID: \"b8b92a83-a81a-4b3c-875a-65bd70b86a12\") " pod="openshift-route-controller-manager/route-controller-manager-ff6dfd758-fh9rf" Dec 06 03:45:19 crc kubenswrapper[4802]: I1206 03:45:19.977002 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f053daeb-b5e4-444e-a580-89a049e64beb-client-ca\") pod \"controller-manager-575ccfcfc8-ct5bx\" (UID: \"f053daeb-b5e4-444e-a580-89a049e64beb\") " pod="openshift-controller-manager/controller-manager-575ccfcfc8-ct5bx" Dec 06 03:45:20 crc kubenswrapper[4802]: I1206 03:45:20.078329 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nv2gv\" (UniqueName: \"kubernetes.io/projected/f053daeb-b5e4-444e-a580-89a049e64beb-kube-api-access-nv2gv\") pod \"controller-manager-575ccfcfc8-ct5bx\" (UID: \"f053daeb-b5e4-444e-a580-89a049e64beb\") " pod="openshift-controller-manager/controller-manager-575ccfcfc8-ct5bx" Dec 06 03:45:20 crc kubenswrapper[4802]: I1206 03:45:20.078493 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b8b92a83-a81a-4b3c-875a-65bd70b86a12-serving-cert\") pod \"route-controller-manager-ff6dfd758-fh9rf\" (UID: \"b8b92a83-a81a-4b3c-875a-65bd70b86a12\") " pod="openshift-route-controller-manager/route-controller-manager-ff6dfd758-fh9rf" Dec 06 03:45:20 crc kubenswrapper[4802]: I1206 03:45:20.078567 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b8b92a83-a81a-4b3c-875a-65bd70b86a12-client-ca\") pod \"route-controller-manager-ff6dfd758-fh9rf\" (UID: \"b8b92a83-a81a-4b3c-875a-65bd70b86a12\") " pod="openshift-route-controller-manager/route-controller-manager-ff6dfd758-fh9rf" Dec 06 03:45:20 crc kubenswrapper[4802]: I1206 03:45:20.078652 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b8b92a83-a81a-4b3c-875a-65bd70b86a12-config\") pod \"route-controller-manager-ff6dfd758-fh9rf\" (UID: \"b8b92a83-a81a-4b3c-875a-65bd70b86a12\") " pod="openshift-route-controller-manager/route-controller-manager-ff6dfd758-fh9rf" Dec 06 03:45:20 crc kubenswrapper[4802]: I1206 03:45:20.078709 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f053daeb-b5e4-444e-a580-89a049e64beb-config\") pod \"controller-manager-575ccfcfc8-ct5bx\" (UID: \"f053daeb-b5e4-444e-a580-89a049e64beb\") " pod="openshift-controller-manager/controller-manager-575ccfcfc8-ct5bx" Dec 06 03:45:20 crc kubenswrapper[4802]: I1206 03:45:20.078792 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f053daeb-b5e4-444e-a580-89a049e64beb-serving-cert\") pod \"controller-manager-575ccfcfc8-ct5bx\" (UID: \"f053daeb-b5e4-444e-a580-89a049e64beb\") " pod="openshift-controller-manager/controller-manager-575ccfcfc8-ct5bx" Dec 06 03:45:20 crc kubenswrapper[4802]: I1206 03:45:20.078868 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f053daeb-b5e4-444e-a580-89a049e64beb-proxy-ca-bundles\") pod \"controller-manager-575ccfcfc8-ct5bx\" (UID: \"f053daeb-b5e4-444e-a580-89a049e64beb\") " pod="openshift-controller-manager/controller-manager-575ccfcfc8-ct5bx" Dec 06 03:45:20 crc kubenswrapper[4802]: I1206 03:45:20.078925 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzmlz\" (UniqueName: \"kubernetes.io/projected/b8b92a83-a81a-4b3c-875a-65bd70b86a12-kube-api-access-fzmlz\") pod \"route-controller-manager-ff6dfd758-fh9rf\" (UID: \"b8b92a83-a81a-4b3c-875a-65bd70b86a12\") " pod="openshift-route-controller-manager/route-controller-manager-ff6dfd758-fh9rf" Dec 06 03:45:20 crc kubenswrapper[4802]: I1206 03:45:20.079044 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f053daeb-b5e4-444e-a580-89a049e64beb-client-ca\") pod \"controller-manager-575ccfcfc8-ct5bx\" (UID: \"f053daeb-b5e4-444e-a580-89a049e64beb\") " pod="openshift-controller-manager/controller-manager-575ccfcfc8-ct5bx" Dec 06 03:45:20 crc kubenswrapper[4802]: I1206 03:45:20.080215 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b8b92a83-a81a-4b3c-875a-65bd70b86a12-client-ca\") pod \"route-controller-manager-ff6dfd758-fh9rf\" (UID: \"b8b92a83-a81a-4b3c-875a-65bd70b86a12\") " pod="openshift-route-controller-manager/route-controller-manager-ff6dfd758-fh9rf" Dec 06 03:45:20 crc kubenswrapper[4802]: I1206 03:45:20.081084 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f053daeb-b5e4-444e-a580-89a049e64beb-client-ca\") pod \"controller-manager-575ccfcfc8-ct5bx\" (UID: \"f053daeb-b5e4-444e-a580-89a049e64beb\") " pod="openshift-controller-manager/controller-manager-575ccfcfc8-ct5bx" Dec 06 03:45:20 crc kubenswrapper[4802]: I1206 03:45:20.081377 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f053daeb-b5e4-444e-a580-89a049e64beb-config\") pod \"controller-manager-575ccfcfc8-ct5bx\" (UID: \"f053daeb-b5e4-444e-a580-89a049e64beb\") " pod="openshift-controller-manager/controller-manager-575ccfcfc8-ct5bx" Dec 06 03:45:20 crc kubenswrapper[4802]: I1206 03:45:20.081399 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f053daeb-b5e4-444e-a580-89a049e64beb-proxy-ca-bundles\") pod \"controller-manager-575ccfcfc8-ct5bx\" (UID: \"f053daeb-b5e4-444e-a580-89a049e64beb\") " pod="openshift-controller-manager/controller-manager-575ccfcfc8-ct5bx" Dec 06 03:45:20 crc kubenswrapper[4802]: I1206 03:45:20.081887 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b8b92a83-a81a-4b3c-875a-65bd70b86a12-config\") pod \"route-controller-manager-ff6dfd758-fh9rf\" (UID: \"b8b92a83-a81a-4b3c-875a-65bd70b86a12\") " pod="openshift-route-controller-manager/route-controller-manager-ff6dfd758-fh9rf" Dec 06 03:45:20 crc kubenswrapper[4802]: I1206 03:45:20.083159 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b8b92a83-a81a-4b3c-875a-65bd70b86a12-serving-cert\") pod \"route-controller-manager-ff6dfd758-fh9rf\" (UID: \"b8b92a83-a81a-4b3c-875a-65bd70b86a12\") " pod="openshift-route-controller-manager/route-controller-manager-ff6dfd758-fh9rf" Dec 06 03:45:20 crc kubenswrapper[4802]: I1206 03:45:20.087736 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f053daeb-b5e4-444e-a580-89a049e64beb-serving-cert\") pod \"controller-manager-575ccfcfc8-ct5bx\" (UID: \"f053daeb-b5e4-444e-a580-89a049e64beb\") " pod="openshift-controller-manager/controller-manager-575ccfcfc8-ct5bx" Dec 06 03:45:20 crc kubenswrapper[4802]: I1206 03:45:20.104986 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzmlz\" (UniqueName: \"kubernetes.io/projected/b8b92a83-a81a-4b3c-875a-65bd70b86a12-kube-api-access-fzmlz\") pod \"route-controller-manager-ff6dfd758-fh9rf\" (UID: \"b8b92a83-a81a-4b3c-875a-65bd70b86a12\") " pod="openshift-route-controller-manager/route-controller-manager-ff6dfd758-fh9rf" Dec 06 03:45:20 crc kubenswrapper[4802]: I1206 03:45:20.105098 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nv2gv\" (UniqueName: \"kubernetes.io/projected/f053daeb-b5e4-444e-a580-89a049e64beb-kube-api-access-nv2gv\") pod \"controller-manager-575ccfcfc8-ct5bx\" (UID: \"f053daeb-b5e4-444e-a580-89a049e64beb\") " pod="openshift-controller-manager/controller-manager-575ccfcfc8-ct5bx" Dec 06 03:45:20 crc kubenswrapper[4802]: I1206 03:45:20.112715 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-cgpdh" event={"ID":"a68fc0dd-997c-4d12-9a20-5345dfe98a41","Type":"ContainerDied","Data":"197163bad5e7b4ac89d2ca3dc0e289138a45c852df8c982644c9ef597896be89"} Dec 06 03:45:20 crc kubenswrapper[4802]: I1206 03:45:20.112931 4802 scope.go:117] "RemoveContainer" containerID="f9fc0a9a62f254e815cc5ad499da3ac6e027b333f613cfd4d44e4a88fe48a39a" Dec 06 03:45:20 crc kubenswrapper[4802]: I1206 03:45:20.113437 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-cgpdh" Dec 06 03:45:20 crc kubenswrapper[4802]: I1206 03:45:20.155226 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-cgpdh"] Dec 06 03:45:20 crc kubenswrapper[4802]: I1206 03:45:20.158553 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-cgpdh"] Dec 06 03:45:20 crc kubenswrapper[4802]: I1206 03:45:20.215998 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-ff6dfd758-fh9rf" Dec 06 03:45:20 crc kubenswrapper[4802]: I1206 03:45:20.220216 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-575ccfcfc8-ct5bx" Dec 06 03:45:20 crc kubenswrapper[4802]: I1206 03:45:20.412331 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-ff6dfd758-fh9rf"] Dec 06 03:45:20 crc kubenswrapper[4802]: I1206 03:45:20.593500 4802 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 06 03:45:20 crc kubenswrapper[4802]: I1206 03:45:20.685876 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-575ccfcfc8-ct5bx"] Dec 06 03:45:20 crc kubenswrapper[4802]: W1206 03:45:20.687788 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf053daeb_b5e4_444e_a580_89a049e64beb.slice/crio-667062484bce48da96ad6348f6756471a061bf8265a87e1d934c156155724831 WatchSource:0}: Error finding container 667062484bce48da96ad6348f6756471a061bf8265a87e1d934c156155724831: Status 404 returned error can't find the container with id 667062484bce48da96ad6348f6756471a061bf8265a87e1d934c156155724831 Dec 06 03:45:21 crc kubenswrapper[4802]: I1206 03:45:21.118972 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-575ccfcfc8-ct5bx" event={"ID":"f053daeb-b5e4-444e-a580-89a049e64beb","Type":"ContainerStarted","Data":"f9edb365fd08f287b7bfddf5c09936e7acd026f108b1e4c7ffcb9934238c43fb"} Dec 06 03:45:21 crc kubenswrapper[4802]: I1206 03:45:21.119266 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-575ccfcfc8-ct5bx" Dec 06 03:45:21 crc kubenswrapper[4802]: I1206 03:45:21.119280 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-575ccfcfc8-ct5bx" event={"ID":"f053daeb-b5e4-444e-a580-89a049e64beb","Type":"ContainerStarted","Data":"667062484bce48da96ad6348f6756471a061bf8265a87e1d934c156155724831"} Dec 06 03:45:21 crc kubenswrapper[4802]: I1206 03:45:21.121466 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-ff6dfd758-fh9rf" event={"ID":"b8b92a83-a81a-4b3c-875a-65bd70b86a12","Type":"ContainerStarted","Data":"40851f37ccafbac8d005d368d0060f82adad5099f0b949faf7642d3964a74ae8"} Dec 06 03:45:21 crc kubenswrapper[4802]: I1206 03:45:21.121490 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-ff6dfd758-fh9rf" event={"ID":"b8b92a83-a81a-4b3c-875a-65bd70b86a12","Type":"ContainerStarted","Data":"ba65eb261632325af341fa05a6e29bbe3c2bf56295a7936c6d3e98ef2e266178"} Dec 06 03:45:21 crc kubenswrapper[4802]: I1206 03:45:21.121792 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-ff6dfd758-fh9rf" Dec 06 03:45:21 crc kubenswrapper[4802]: I1206 03:45:21.125838 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-ff6dfd758-fh9rf" Dec 06 03:45:21 crc kubenswrapper[4802]: I1206 03:45:21.126444 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-575ccfcfc8-ct5bx" Dec 06 03:45:21 crc kubenswrapper[4802]: I1206 03:45:21.141292 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-575ccfcfc8-ct5bx" podStartSLOduration=3.141273868 podStartE2EDuration="3.141273868s" podCreationTimestamp="2025-12-06 03:45:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:45:21.137199248 +0000 UTC m=+314.009108400" watchObservedRunningTime="2025-12-06 03:45:21.141273868 +0000 UTC m=+314.013183030" Dec 06 03:45:21 crc kubenswrapper[4802]: I1206 03:45:21.460102 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a68fc0dd-997c-4d12-9a20-5345dfe98a41" path="/var/lib/kubelet/pods/a68fc0dd-997c-4d12-9a20-5345dfe98a41/volumes" Dec 06 03:45:21 crc kubenswrapper[4802]: I1206 03:45:21.721387 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 06 03:45:21 crc kubenswrapper[4802]: I1206 03:45:21.874635 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 06 03:45:22 crc kubenswrapper[4802]: I1206 03:45:22.613328 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-ff6dfd758-fh9rf" podStartSLOduration=4.6133103890000005 podStartE2EDuration="4.613310389s" podCreationTimestamp="2025-12-06 03:45:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:45:21.178419977 +0000 UTC m=+314.050329139" watchObservedRunningTime="2025-12-06 03:45:22.613310389 +0000 UTC m=+315.485219541" Dec 06 03:45:22 crc kubenswrapper[4802]: I1206 03:45:22.615987 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-575ccfcfc8-ct5bx"] Dec 06 03:45:22 crc kubenswrapper[4802]: I1206 03:45:22.637253 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-ff6dfd758-fh9rf"] Dec 06 03:45:22 crc kubenswrapper[4802]: I1206 03:45:22.883012 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 06 03:45:24 crc kubenswrapper[4802]: I1206 03:45:24.074668 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 06 03:45:24 crc kubenswrapper[4802]: I1206 03:45:24.146415 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-575ccfcfc8-ct5bx" podUID="f053daeb-b5e4-444e-a580-89a049e64beb" containerName="controller-manager" containerID="cri-o://f9edb365fd08f287b7bfddf5c09936e7acd026f108b1e4c7ffcb9934238c43fb" gracePeriod=30 Dec 06 03:45:24 crc kubenswrapper[4802]: I1206 03:45:24.146649 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-ff6dfd758-fh9rf" podUID="b8b92a83-a81a-4b3c-875a-65bd70b86a12" containerName="route-controller-manager" containerID="cri-o://40851f37ccafbac8d005d368d0060f82adad5099f0b949faf7642d3964a74ae8" gracePeriod=30 Dec 06 03:45:24 crc kubenswrapper[4802]: I1206 03:45:24.227386 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 06 03:45:24 crc kubenswrapper[4802]: I1206 03:45:24.431555 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 06 03:45:25 crc kubenswrapper[4802]: I1206 03:45:25.155263 4802 generic.go:334] "Generic (PLEG): container finished" podID="f053daeb-b5e4-444e-a580-89a049e64beb" containerID="f9edb365fd08f287b7bfddf5c09936e7acd026f108b1e4c7ffcb9934238c43fb" exitCode=0 Dec 06 03:45:25 crc kubenswrapper[4802]: I1206 03:45:25.155361 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-575ccfcfc8-ct5bx" event={"ID":"f053daeb-b5e4-444e-a580-89a049e64beb","Type":"ContainerDied","Data":"f9edb365fd08f287b7bfddf5c09936e7acd026f108b1e4c7ffcb9934238c43fb"} Dec 06 03:45:25 crc kubenswrapper[4802]: I1206 03:45:25.160137 4802 generic.go:334] "Generic (PLEG): container finished" podID="b8b92a83-a81a-4b3c-875a-65bd70b86a12" containerID="40851f37ccafbac8d005d368d0060f82adad5099f0b949faf7642d3964a74ae8" exitCode=0 Dec 06 03:45:25 crc kubenswrapper[4802]: I1206 03:45:25.160618 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-ff6dfd758-fh9rf" event={"ID":"b8b92a83-a81a-4b3c-875a-65bd70b86a12","Type":"ContainerDied","Data":"40851f37ccafbac8d005d368d0060f82adad5099f0b949faf7642d3964a74ae8"} Dec 06 03:45:25 crc kubenswrapper[4802]: I1206 03:45:25.266035 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 06 03:45:25 crc kubenswrapper[4802]: I1206 03:45:25.308348 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-575ccfcfc8-ct5bx" Dec 06 03:45:25 crc kubenswrapper[4802]: I1206 03:45:25.356080 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-d5d5747-vgjmr"] Dec 06 03:45:25 crc kubenswrapper[4802]: E1206 03:45:25.356540 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f053daeb-b5e4-444e-a580-89a049e64beb" containerName="controller-manager" Dec 06 03:45:25 crc kubenswrapper[4802]: I1206 03:45:25.356557 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="f053daeb-b5e4-444e-a580-89a049e64beb" containerName="controller-manager" Dec 06 03:45:25 crc kubenswrapper[4802]: I1206 03:45:25.356911 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="f053daeb-b5e4-444e-a580-89a049e64beb" containerName="controller-manager" Dec 06 03:45:25 crc kubenswrapper[4802]: I1206 03:45:25.358041 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-d5d5747-vgjmr" Dec 06 03:45:25 crc kubenswrapper[4802]: I1206 03:45:25.358093 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-d5d5747-vgjmr"] Dec 06 03:45:25 crc kubenswrapper[4802]: I1206 03:45:25.364041 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-ff6dfd758-fh9rf" Dec 06 03:45:25 crc kubenswrapper[4802]: I1206 03:45:25.457678 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nv2gv\" (UniqueName: \"kubernetes.io/projected/f053daeb-b5e4-444e-a580-89a049e64beb-kube-api-access-nv2gv\") pod \"f053daeb-b5e4-444e-a580-89a049e64beb\" (UID: \"f053daeb-b5e4-444e-a580-89a049e64beb\") " Dec 06 03:45:25 crc kubenswrapper[4802]: I1206 03:45:25.457729 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f053daeb-b5e4-444e-a580-89a049e64beb-client-ca\") pod \"f053daeb-b5e4-444e-a580-89a049e64beb\" (UID: \"f053daeb-b5e4-444e-a580-89a049e64beb\") " Dec 06 03:45:25 crc kubenswrapper[4802]: I1206 03:45:25.457809 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f053daeb-b5e4-444e-a580-89a049e64beb-config\") pod \"f053daeb-b5e4-444e-a580-89a049e64beb\" (UID: \"f053daeb-b5e4-444e-a580-89a049e64beb\") " Dec 06 03:45:25 crc kubenswrapper[4802]: I1206 03:45:25.457827 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b8b92a83-a81a-4b3c-875a-65bd70b86a12-config\") pod \"b8b92a83-a81a-4b3c-875a-65bd70b86a12\" (UID: \"b8b92a83-a81a-4b3c-875a-65bd70b86a12\") " Dec 06 03:45:25 crc kubenswrapper[4802]: I1206 03:45:25.457863 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b8b92a83-a81a-4b3c-875a-65bd70b86a12-client-ca\") pod \"b8b92a83-a81a-4b3c-875a-65bd70b86a12\" (UID: \"b8b92a83-a81a-4b3c-875a-65bd70b86a12\") " Dec 06 03:45:25 crc kubenswrapper[4802]: I1206 03:45:25.457885 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f053daeb-b5e4-444e-a580-89a049e64beb-serving-cert\") pod \"f053daeb-b5e4-444e-a580-89a049e64beb\" (UID: \"f053daeb-b5e4-444e-a580-89a049e64beb\") " Dec 06 03:45:25 crc kubenswrapper[4802]: I1206 03:45:25.457908 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f053daeb-b5e4-444e-a580-89a049e64beb-proxy-ca-bundles\") pod \"f053daeb-b5e4-444e-a580-89a049e64beb\" (UID: \"f053daeb-b5e4-444e-a580-89a049e64beb\") " Dec 06 03:45:25 crc kubenswrapper[4802]: I1206 03:45:25.457929 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzmlz\" (UniqueName: \"kubernetes.io/projected/b8b92a83-a81a-4b3c-875a-65bd70b86a12-kube-api-access-fzmlz\") pod \"b8b92a83-a81a-4b3c-875a-65bd70b86a12\" (UID: \"b8b92a83-a81a-4b3c-875a-65bd70b86a12\") " Dec 06 03:45:25 crc kubenswrapper[4802]: I1206 03:45:25.457982 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b8b92a83-a81a-4b3c-875a-65bd70b86a12-serving-cert\") pod \"b8b92a83-a81a-4b3c-875a-65bd70b86a12\" (UID: \"b8b92a83-a81a-4b3c-875a-65bd70b86a12\") " Dec 06 03:45:25 crc kubenswrapper[4802]: I1206 03:45:25.458113 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1eed2052-fc62-4552-9ca6-3c23175b784a-proxy-ca-bundles\") pod \"controller-manager-d5d5747-vgjmr\" (UID: \"1eed2052-fc62-4552-9ca6-3c23175b784a\") " pod="openshift-controller-manager/controller-manager-d5d5747-vgjmr" Dec 06 03:45:25 crc kubenswrapper[4802]: I1206 03:45:25.458164 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fv8nn\" (UniqueName: \"kubernetes.io/projected/1eed2052-fc62-4552-9ca6-3c23175b784a-kube-api-access-fv8nn\") pod \"controller-manager-d5d5747-vgjmr\" (UID: \"1eed2052-fc62-4552-9ca6-3c23175b784a\") " pod="openshift-controller-manager/controller-manager-d5d5747-vgjmr" Dec 06 03:45:25 crc kubenswrapper[4802]: I1206 03:45:25.458229 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1eed2052-fc62-4552-9ca6-3c23175b784a-serving-cert\") pod \"controller-manager-d5d5747-vgjmr\" (UID: \"1eed2052-fc62-4552-9ca6-3c23175b784a\") " pod="openshift-controller-manager/controller-manager-d5d5747-vgjmr" Dec 06 03:45:25 crc kubenswrapper[4802]: I1206 03:45:25.458252 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1eed2052-fc62-4552-9ca6-3c23175b784a-config\") pod \"controller-manager-d5d5747-vgjmr\" (UID: \"1eed2052-fc62-4552-9ca6-3c23175b784a\") " pod="openshift-controller-manager/controller-manager-d5d5747-vgjmr" Dec 06 03:45:25 crc kubenswrapper[4802]: I1206 03:45:25.458275 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1eed2052-fc62-4552-9ca6-3c23175b784a-client-ca\") pod \"controller-manager-d5d5747-vgjmr\" (UID: \"1eed2052-fc62-4552-9ca6-3c23175b784a\") " pod="openshift-controller-manager/controller-manager-d5d5747-vgjmr" Dec 06 03:45:25 crc kubenswrapper[4802]: I1206 03:45:25.459085 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b8b92a83-a81a-4b3c-875a-65bd70b86a12-config" (OuterVolumeSpecName: "config") pod "b8b92a83-a81a-4b3c-875a-65bd70b86a12" (UID: "b8b92a83-a81a-4b3c-875a-65bd70b86a12"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:45:25 crc kubenswrapper[4802]: I1206 03:45:25.459215 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b8b92a83-a81a-4b3c-875a-65bd70b86a12-client-ca" (OuterVolumeSpecName: "client-ca") pod "b8b92a83-a81a-4b3c-875a-65bd70b86a12" (UID: "b8b92a83-a81a-4b3c-875a-65bd70b86a12"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:45:25 crc kubenswrapper[4802]: I1206 03:45:25.459850 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f053daeb-b5e4-444e-a580-89a049e64beb-client-ca" (OuterVolumeSpecName: "client-ca") pod "f053daeb-b5e4-444e-a580-89a049e64beb" (UID: "f053daeb-b5e4-444e-a580-89a049e64beb"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:45:25 crc kubenswrapper[4802]: I1206 03:45:25.460390 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f053daeb-b5e4-444e-a580-89a049e64beb-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "f053daeb-b5e4-444e-a580-89a049e64beb" (UID: "f053daeb-b5e4-444e-a580-89a049e64beb"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:45:25 crc kubenswrapper[4802]: I1206 03:45:25.460439 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f053daeb-b5e4-444e-a580-89a049e64beb-config" (OuterVolumeSpecName: "config") pod "f053daeb-b5e4-444e-a580-89a049e64beb" (UID: "f053daeb-b5e4-444e-a580-89a049e64beb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:45:25 crc kubenswrapper[4802]: I1206 03:45:25.464035 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f053daeb-b5e4-444e-a580-89a049e64beb-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "f053daeb-b5e4-444e-a580-89a049e64beb" (UID: "f053daeb-b5e4-444e-a580-89a049e64beb"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:45:25 crc kubenswrapper[4802]: I1206 03:45:25.464036 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f053daeb-b5e4-444e-a580-89a049e64beb-kube-api-access-nv2gv" (OuterVolumeSpecName: "kube-api-access-nv2gv") pod "f053daeb-b5e4-444e-a580-89a049e64beb" (UID: "f053daeb-b5e4-444e-a580-89a049e64beb"). InnerVolumeSpecName "kube-api-access-nv2gv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:45:25 crc kubenswrapper[4802]: I1206 03:45:25.464326 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8b92a83-a81a-4b3c-875a-65bd70b86a12-kube-api-access-fzmlz" (OuterVolumeSpecName: "kube-api-access-fzmlz") pod "b8b92a83-a81a-4b3c-875a-65bd70b86a12" (UID: "b8b92a83-a81a-4b3c-875a-65bd70b86a12"). InnerVolumeSpecName "kube-api-access-fzmlz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:45:25 crc kubenswrapper[4802]: I1206 03:45:25.464426 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8b92a83-a81a-4b3c-875a-65bd70b86a12-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "b8b92a83-a81a-4b3c-875a-65bd70b86a12" (UID: "b8b92a83-a81a-4b3c-875a-65bd70b86a12"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:45:25 crc kubenswrapper[4802]: I1206 03:45:25.559225 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fv8nn\" (UniqueName: \"kubernetes.io/projected/1eed2052-fc62-4552-9ca6-3c23175b784a-kube-api-access-fv8nn\") pod \"controller-manager-d5d5747-vgjmr\" (UID: \"1eed2052-fc62-4552-9ca6-3c23175b784a\") " pod="openshift-controller-manager/controller-manager-d5d5747-vgjmr" Dec 06 03:45:25 crc kubenswrapper[4802]: I1206 03:45:25.559316 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1eed2052-fc62-4552-9ca6-3c23175b784a-serving-cert\") pod \"controller-manager-d5d5747-vgjmr\" (UID: \"1eed2052-fc62-4552-9ca6-3c23175b784a\") " pod="openshift-controller-manager/controller-manager-d5d5747-vgjmr" Dec 06 03:45:25 crc kubenswrapper[4802]: I1206 03:45:25.559364 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1eed2052-fc62-4552-9ca6-3c23175b784a-config\") pod \"controller-manager-d5d5747-vgjmr\" (UID: \"1eed2052-fc62-4552-9ca6-3c23175b784a\") " pod="openshift-controller-manager/controller-manager-d5d5747-vgjmr" Dec 06 03:45:25 crc kubenswrapper[4802]: I1206 03:45:25.559434 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1eed2052-fc62-4552-9ca6-3c23175b784a-client-ca\") pod \"controller-manager-d5d5747-vgjmr\" (UID: \"1eed2052-fc62-4552-9ca6-3c23175b784a\") " pod="openshift-controller-manager/controller-manager-d5d5747-vgjmr" Dec 06 03:45:25 crc kubenswrapper[4802]: I1206 03:45:25.559490 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1eed2052-fc62-4552-9ca6-3c23175b784a-proxy-ca-bundles\") pod \"controller-manager-d5d5747-vgjmr\" (UID: \"1eed2052-fc62-4552-9ca6-3c23175b784a\") " pod="openshift-controller-manager/controller-manager-d5d5747-vgjmr" Dec 06 03:45:25 crc kubenswrapper[4802]: I1206 03:45:25.559954 4802 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b8b92a83-a81a-4b3c-875a-65bd70b86a12-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:45:25 crc kubenswrapper[4802]: I1206 03:45:25.559992 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nv2gv\" (UniqueName: \"kubernetes.io/projected/f053daeb-b5e4-444e-a580-89a049e64beb-kube-api-access-nv2gv\") on node \"crc\" DevicePath \"\"" Dec 06 03:45:25 crc kubenswrapper[4802]: I1206 03:45:25.560023 4802 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f053daeb-b5e4-444e-a580-89a049e64beb-client-ca\") on node \"crc\" DevicePath \"\"" Dec 06 03:45:25 crc kubenswrapper[4802]: I1206 03:45:25.560048 4802 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b8b92a83-a81a-4b3c-875a-65bd70b86a12-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:45:25 crc kubenswrapper[4802]: I1206 03:45:25.560075 4802 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f053daeb-b5e4-444e-a580-89a049e64beb-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:45:25 crc kubenswrapper[4802]: I1206 03:45:25.560099 4802 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b8b92a83-a81a-4b3c-875a-65bd70b86a12-client-ca\") on node \"crc\" DevicePath \"\"" Dec 06 03:45:25 crc kubenswrapper[4802]: I1206 03:45:25.560127 4802 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f053daeb-b5e4-444e-a580-89a049e64beb-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:45:25 crc kubenswrapper[4802]: I1206 03:45:25.560150 4802 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f053daeb-b5e4-444e-a580-89a049e64beb-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 06 03:45:25 crc kubenswrapper[4802]: I1206 03:45:25.560173 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzmlz\" (UniqueName: \"kubernetes.io/projected/b8b92a83-a81a-4b3c-875a-65bd70b86a12-kube-api-access-fzmlz\") on node \"crc\" DevicePath \"\"" Dec 06 03:45:25 crc kubenswrapper[4802]: I1206 03:45:25.560906 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1eed2052-fc62-4552-9ca6-3c23175b784a-config\") pod \"controller-manager-d5d5747-vgjmr\" (UID: \"1eed2052-fc62-4552-9ca6-3c23175b784a\") " pod="openshift-controller-manager/controller-manager-d5d5747-vgjmr" Dec 06 03:45:25 crc kubenswrapper[4802]: I1206 03:45:25.561499 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1eed2052-fc62-4552-9ca6-3c23175b784a-client-ca\") pod \"controller-manager-d5d5747-vgjmr\" (UID: \"1eed2052-fc62-4552-9ca6-3c23175b784a\") " pod="openshift-controller-manager/controller-manager-d5d5747-vgjmr" Dec 06 03:45:25 crc kubenswrapper[4802]: I1206 03:45:25.562880 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1eed2052-fc62-4552-9ca6-3c23175b784a-proxy-ca-bundles\") pod \"controller-manager-d5d5747-vgjmr\" (UID: \"1eed2052-fc62-4552-9ca6-3c23175b784a\") " pod="openshift-controller-manager/controller-manager-d5d5747-vgjmr" Dec 06 03:45:25 crc kubenswrapper[4802]: I1206 03:45:25.566087 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1eed2052-fc62-4552-9ca6-3c23175b784a-serving-cert\") pod \"controller-manager-d5d5747-vgjmr\" (UID: \"1eed2052-fc62-4552-9ca6-3c23175b784a\") " pod="openshift-controller-manager/controller-manager-d5d5747-vgjmr" Dec 06 03:45:25 crc kubenswrapper[4802]: I1206 03:45:25.586971 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fv8nn\" (UniqueName: \"kubernetes.io/projected/1eed2052-fc62-4552-9ca6-3c23175b784a-kube-api-access-fv8nn\") pod \"controller-manager-d5d5747-vgjmr\" (UID: \"1eed2052-fc62-4552-9ca6-3c23175b784a\") " pod="openshift-controller-manager/controller-manager-d5d5747-vgjmr" Dec 06 03:45:25 crc kubenswrapper[4802]: I1206 03:45:25.679204 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-d5d5747-vgjmr" Dec 06 03:45:25 crc kubenswrapper[4802]: I1206 03:45:25.887293 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-d5d5747-vgjmr"] Dec 06 03:45:26 crc kubenswrapper[4802]: I1206 03:45:26.167727 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-575ccfcfc8-ct5bx" event={"ID":"f053daeb-b5e4-444e-a580-89a049e64beb","Type":"ContainerDied","Data":"667062484bce48da96ad6348f6756471a061bf8265a87e1d934c156155724831"} Dec 06 03:45:26 crc kubenswrapper[4802]: I1206 03:45:26.167779 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-575ccfcfc8-ct5bx" Dec 06 03:45:26 crc kubenswrapper[4802]: I1206 03:45:26.168127 4802 scope.go:117] "RemoveContainer" containerID="f9edb365fd08f287b7bfddf5c09936e7acd026f108b1e4c7ffcb9934238c43fb" Dec 06 03:45:26 crc kubenswrapper[4802]: I1206 03:45:26.169691 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-ff6dfd758-fh9rf" Dec 06 03:45:26 crc kubenswrapper[4802]: I1206 03:45:26.169721 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-ff6dfd758-fh9rf" event={"ID":"b8b92a83-a81a-4b3c-875a-65bd70b86a12","Type":"ContainerDied","Data":"ba65eb261632325af341fa05a6e29bbe3c2bf56295a7936c6d3e98ef2e266178"} Dec 06 03:45:26 crc kubenswrapper[4802]: I1206 03:45:26.171606 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-d5d5747-vgjmr" event={"ID":"1eed2052-fc62-4552-9ca6-3c23175b784a","Type":"ContainerStarted","Data":"de14875ffed3a903daba61d2fa8c74c65ee77330d42b3ca8eea6c3181a0770db"} Dec 06 03:45:26 crc kubenswrapper[4802]: I1206 03:45:26.171667 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-d5d5747-vgjmr" event={"ID":"1eed2052-fc62-4552-9ca6-3c23175b784a","Type":"ContainerStarted","Data":"c406bc8b7652cc175943a27d0c3e34b0b56a2c9d16e8d08f6564d0dff66eae59"} Dec 06 03:45:26 crc kubenswrapper[4802]: I1206 03:45:26.171788 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-d5d5747-vgjmr" Dec 06 03:45:26 crc kubenswrapper[4802]: I1206 03:45:26.175849 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-d5d5747-vgjmr" Dec 06 03:45:26 crc kubenswrapper[4802]: I1206 03:45:26.186141 4802 scope.go:117] "RemoveContainer" containerID="40851f37ccafbac8d005d368d0060f82adad5099f0b949faf7642d3964a74ae8" Dec 06 03:45:26 crc kubenswrapper[4802]: I1206 03:45:26.191143 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-d5d5747-vgjmr" podStartSLOduration=4.191126732 podStartE2EDuration="4.191126732s" podCreationTimestamp="2025-12-06 03:45:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:45:26.189048715 +0000 UTC m=+319.060957877" watchObservedRunningTime="2025-12-06 03:45:26.191126732 +0000 UTC m=+319.063035884" Dec 06 03:45:26 crc kubenswrapper[4802]: I1206 03:45:26.203228 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-575ccfcfc8-ct5bx"] Dec 06 03:45:26 crc kubenswrapper[4802]: I1206 03:45:26.210493 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-575ccfcfc8-ct5bx"] Dec 06 03:45:26 crc kubenswrapper[4802]: I1206 03:45:26.223546 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-ff6dfd758-fh9rf"] Dec 06 03:45:26 crc kubenswrapper[4802]: I1206 03:45:26.230389 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-ff6dfd758-fh9rf"] Dec 06 03:45:27 crc kubenswrapper[4802]: I1206 03:45:27.402003 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 06 03:45:27 crc kubenswrapper[4802]: I1206 03:45:27.456285 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8b92a83-a81a-4b3c-875a-65bd70b86a12" path="/var/lib/kubelet/pods/b8b92a83-a81a-4b3c-875a-65bd70b86a12/volumes" Dec 06 03:45:27 crc kubenswrapper[4802]: I1206 03:45:27.457493 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f053daeb-b5e4-444e-a580-89a049e64beb" path="/var/lib/kubelet/pods/f053daeb-b5e4-444e-a580-89a049e64beb/volumes" Dec 06 03:45:27 crc kubenswrapper[4802]: I1206 03:45:27.888974 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-55985dff9-cfd84"] Dec 06 03:45:27 crc kubenswrapper[4802]: E1206 03:45:27.889180 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8b92a83-a81a-4b3c-875a-65bd70b86a12" containerName="route-controller-manager" Dec 06 03:45:27 crc kubenswrapper[4802]: I1206 03:45:27.889192 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8b92a83-a81a-4b3c-875a-65bd70b86a12" containerName="route-controller-manager" Dec 06 03:45:27 crc kubenswrapper[4802]: I1206 03:45:27.889275 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8b92a83-a81a-4b3c-875a-65bd70b86a12" containerName="route-controller-manager" Dec 06 03:45:27 crc kubenswrapper[4802]: I1206 03:45:27.889598 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-55985dff9-cfd84" Dec 06 03:45:27 crc kubenswrapper[4802]: I1206 03:45:27.891595 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 06 03:45:27 crc kubenswrapper[4802]: I1206 03:45:27.892942 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 06 03:45:27 crc kubenswrapper[4802]: I1206 03:45:27.893129 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 06 03:45:27 crc kubenswrapper[4802]: I1206 03:45:27.893805 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 06 03:45:27 crc kubenswrapper[4802]: I1206 03:45:27.894037 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 06 03:45:27 crc kubenswrapper[4802]: I1206 03:45:27.898336 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 06 03:45:27 crc kubenswrapper[4802]: I1206 03:45:27.902839 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-55985dff9-cfd84"] Dec 06 03:45:27 crc kubenswrapper[4802]: I1206 03:45:27.988800 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2bc45ee3-fd5f-4bfa-a06f-2786b12652d4-client-ca\") pod \"route-controller-manager-55985dff9-cfd84\" (UID: \"2bc45ee3-fd5f-4bfa-a06f-2786b12652d4\") " pod="openshift-route-controller-manager/route-controller-manager-55985dff9-cfd84" Dec 06 03:45:27 crc kubenswrapper[4802]: I1206 03:45:27.988878 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qww87\" (UniqueName: \"kubernetes.io/projected/2bc45ee3-fd5f-4bfa-a06f-2786b12652d4-kube-api-access-qww87\") pod \"route-controller-manager-55985dff9-cfd84\" (UID: \"2bc45ee3-fd5f-4bfa-a06f-2786b12652d4\") " pod="openshift-route-controller-manager/route-controller-manager-55985dff9-cfd84" Dec 06 03:45:27 crc kubenswrapper[4802]: I1206 03:45:27.988909 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2bc45ee3-fd5f-4bfa-a06f-2786b12652d4-config\") pod \"route-controller-manager-55985dff9-cfd84\" (UID: \"2bc45ee3-fd5f-4bfa-a06f-2786b12652d4\") " pod="openshift-route-controller-manager/route-controller-manager-55985dff9-cfd84" Dec 06 03:45:27 crc kubenswrapper[4802]: I1206 03:45:27.989056 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2bc45ee3-fd5f-4bfa-a06f-2786b12652d4-serving-cert\") pod \"route-controller-manager-55985dff9-cfd84\" (UID: \"2bc45ee3-fd5f-4bfa-a06f-2786b12652d4\") " pod="openshift-route-controller-manager/route-controller-manager-55985dff9-cfd84" Dec 06 03:45:28 crc kubenswrapper[4802]: I1206 03:45:28.023247 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 06 03:45:28 crc kubenswrapper[4802]: I1206 03:45:28.090716 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2bc45ee3-fd5f-4bfa-a06f-2786b12652d4-client-ca\") pod \"route-controller-manager-55985dff9-cfd84\" (UID: \"2bc45ee3-fd5f-4bfa-a06f-2786b12652d4\") " pod="openshift-route-controller-manager/route-controller-manager-55985dff9-cfd84" Dec 06 03:45:28 crc kubenswrapper[4802]: I1206 03:45:28.090781 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qww87\" (UniqueName: \"kubernetes.io/projected/2bc45ee3-fd5f-4bfa-a06f-2786b12652d4-kube-api-access-qww87\") pod \"route-controller-manager-55985dff9-cfd84\" (UID: \"2bc45ee3-fd5f-4bfa-a06f-2786b12652d4\") " pod="openshift-route-controller-manager/route-controller-manager-55985dff9-cfd84" Dec 06 03:45:28 crc kubenswrapper[4802]: I1206 03:45:28.090812 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2bc45ee3-fd5f-4bfa-a06f-2786b12652d4-config\") pod \"route-controller-manager-55985dff9-cfd84\" (UID: \"2bc45ee3-fd5f-4bfa-a06f-2786b12652d4\") " pod="openshift-route-controller-manager/route-controller-manager-55985dff9-cfd84" Dec 06 03:45:28 crc kubenswrapper[4802]: I1206 03:45:28.090853 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2bc45ee3-fd5f-4bfa-a06f-2786b12652d4-serving-cert\") pod \"route-controller-manager-55985dff9-cfd84\" (UID: \"2bc45ee3-fd5f-4bfa-a06f-2786b12652d4\") " pod="openshift-route-controller-manager/route-controller-manager-55985dff9-cfd84" Dec 06 03:45:28 crc kubenswrapper[4802]: I1206 03:45:28.091621 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2bc45ee3-fd5f-4bfa-a06f-2786b12652d4-client-ca\") pod \"route-controller-manager-55985dff9-cfd84\" (UID: \"2bc45ee3-fd5f-4bfa-a06f-2786b12652d4\") " pod="openshift-route-controller-manager/route-controller-manager-55985dff9-cfd84" Dec 06 03:45:28 crc kubenswrapper[4802]: I1206 03:45:28.092301 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2bc45ee3-fd5f-4bfa-a06f-2786b12652d4-config\") pod \"route-controller-manager-55985dff9-cfd84\" (UID: \"2bc45ee3-fd5f-4bfa-a06f-2786b12652d4\") " pod="openshift-route-controller-manager/route-controller-manager-55985dff9-cfd84" Dec 06 03:45:28 crc kubenswrapper[4802]: I1206 03:45:28.098407 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2bc45ee3-fd5f-4bfa-a06f-2786b12652d4-serving-cert\") pod \"route-controller-manager-55985dff9-cfd84\" (UID: \"2bc45ee3-fd5f-4bfa-a06f-2786b12652d4\") " pod="openshift-route-controller-manager/route-controller-manager-55985dff9-cfd84" Dec 06 03:45:28 crc kubenswrapper[4802]: I1206 03:45:28.108366 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qww87\" (UniqueName: \"kubernetes.io/projected/2bc45ee3-fd5f-4bfa-a06f-2786b12652d4-kube-api-access-qww87\") pod \"route-controller-manager-55985dff9-cfd84\" (UID: \"2bc45ee3-fd5f-4bfa-a06f-2786b12652d4\") " pod="openshift-route-controller-manager/route-controller-manager-55985dff9-cfd84" Dec 06 03:45:28 crc kubenswrapper[4802]: I1206 03:45:28.248545 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 06 03:45:28 crc kubenswrapper[4802]: I1206 03:45:28.274204 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-55985dff9-cfd84" Dec 06 03:45:28 crc kubenswrapper[4802]: I1206 03:45:28.679552 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-55985dff9-cfd84"] Dec 06 03:45:29 crc kubenswrapper[4802]: I1206 03:45:29.172460 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 06 03:45:29 crc kubenswrapper[4802]: I1206 03:45:29.195314 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-55985dff9-cfd84" event={"ID":"2bc45ee3-fd5f-4bfa-a06f-2786b12652d4","Type":"ContainerStarted","Data":"cf1982c0eba1cf56b150a068b9994a31d6e8df9b5f9c0f34fef5d98cb22c09b3"} Dec 06 03:45:29 crc kubenswrapper[4802]: I1206 03:45:29.195364 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-55985dff9-cfd84" event={"ID":"2bc45ee3-fd5f-4bfa-a06f-2786b12652d4","Type":"ContainerStarted","Data":"20ff2381a09078ff043830cf2e80150fc4da9e4aa701b24ac7d8b828b8920d12"} Dec 06 03:45:29 crc kubenswrapper[4802]: I1206 03:45:29.195598 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-55985dff9-cfd84" Dec 06 03:45:29 crc kubenswrapper[4802]: I1206 03:45:29.225831 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-55985dff9-cfd84" podStartSLOduration=7.225812734 podStartE2EDuration="7.225812734s" podCreationTimestamp="2025-12-06 03:45:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:45:29.222805102 +0000 UTC m=+322.094714264" watchObservedRunningTime="2025-12-06 03:45:29.225812734 +0000 UTC m=+322.097721896" Dec 06 03:45:29 crc kubenswrapper[4802]: I1206 03:45:29.554654 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-55985dff9-cfd84" Dec 06 03:45:31 crc kubenswrapper[4802]: I1206 03:45:31.449826 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 06 03:45:31 crc kubenswrapper[4802]: I1206 03:45:31.486108 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 06 03:45:31 crc kubenswrapper[4802]: I1206 03:45:31.567851 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 06 03:45:31 crc kubenswrapper[4802]: I1206 03:45:31.590146 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 06 03:45:33 crc kubenswrapper[4802]: I1206 03:45:33.664497 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 06 03:45:33 crc kubenswrapper[4802]: I1206 03:45:33.951327 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 06 03:45:38 crc kubenswrapper[4802]: I1206 03:45:38.079536 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 06 03:45:38 crc kubenswrapper[4802]: I1206 03:45:38.657164 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 06 03:45:58 crc kubenswrapper[4802]: I1206 03:45:58.624826 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-d5d5747-vgjmr"] Dec 06 03:45:58 crc kubenswrapper[4802]: I1206 03:45:58.625513 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-d5d5747-vgjmr" podUID="1eed2052-fc62-4552-9ca6-3c23175b784a" containerName="controller-manager" containerID="cri-o://de14875ffed3a903daba61d2fa8c74c65ee77330d42b3ca8eea6c3181a0770db" gracePeriod=30 Dec 06 03:45:58 crc kubenswrapper[4802]: I1206 03:45:58.632468 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-55985dff9-cfd84"] Dec 06 03:45:58 crc kubenswrapper[4802]: I1206 03:45:58.632692 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-55985dff9-cfd84" podUID="2bc45ee3-fd5f-4bfa-a06f-2786b12652d4" containerName="route-controller-manager" containerID="cri-o://cf1982c0eba1cf56b150a068b9994a31d6e8df9b5f9c0f34fef5d98cb22c09b3" gracePeriod=30 Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.205698 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-55985dff9-cfd84" Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.210421 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-d5d5747-vgjmr" Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.303969 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fv8nn\" (UniqueName: \"kubernetes.io/projected/1eed2052-fc62-4552-9ca6-3c23175b784a-kube-api-access-fv8nn\") pod \"1eed2052-fc62-4552-9ca6-3c23175b784a\" (UID: \"1eed2052-fc62-4552-9ca6-3c23175b784a\") " Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.304017 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1eed2052-fc62-4552-9ca6-3c23175b784a-proxy-ca-bundles\") pod \"1eed2052-fc62-4552-9ca6-3c23175b784a\" (UID: \"1eed2052-fc62-4552-9ca6-3c23175b784a\") " Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.304043 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qww87\" (UniqueName: \"kubernetes.io/projected/2bc45ee3-fd5f-4bfa-a06f-2786b12652d4-kube-api-access-qww87\") pod \"2bc45ee3-fd5f-4bfa-a06f-2786b12652d4\" (UID: \"2bc45ee3-fd5f-4bfa-a06f-2786b12652d4\") " Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.304072 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2bc45ee3-fd5f-4bfa-a06f-2786b12652d4-config\") pod \"2bc45ee3-fd5f-4bfa-a06f-2786b12652d4\" (UID: \"2bc45ee3-fd5f-4bfa-a06f-2786b12652d4\") " Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.304125 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1eed2052-fc62-4552-9ca6-3c23175b784a-client-ca\") pod \"1eed2052-fc62-4552-9ca6-3c23175b784a\" (UID: \"1eed2052-fc62-4552-9ca6-3c23175b784a\") " Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.304160 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2bc45ee3-fd5f-4bfa-a06f-2786b12652d4-serving-cert\") pod \"2bc45ee3-fd5f-4bfa-a06f-2786b12652d4\" (UID: \"2bc45ee3-fd5f-4bfa-a06f-2786b12652d4\") " Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.304200 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1eed2052-fc62-4552-9ca6-3c23175b784a-config\") pod \"1eed2052-fc62-4552-9ca6-3c23175b784a\" (UID: \"1eed2052-fc62-4552-9ca6-3c23175b784a\") " Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.304218 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1eed2052-fc62-4552-9ca6-3c23175b784a-serving-cert\") pod \"1eed2052-fc62-4552-9ca6-3c23175b784a\" (UID: \"1eed2052-fc62-4552-9ca6-3c23175b784a\") " Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.304869 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2bc45ee3-fd5f-4bfa-a06f-2786b12652d4-config" (OuterVolumeSpecName: "config") pod "2bc45ee3-fd5f-4bfa-a06f-2786b12652d4" (UID: "2bc45ee3-fd5f-4bfa-a06f-2786b12652d4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.304888 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1eed2052-fc62-4552-9ca6-3c23175b784a-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "1eed2052-fc62-4552-9ca6-3c23175b784a" (UID: "1eed2052-fc62-4552-9ca6-3c23175b784a"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.304952 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1eed2052-fc62-4552-9ca6-3c23175b784a-config" (OuterVolumeSpecName: "config") pod "1eed2052-fc62-4552-9ca6-3c23175b784a" (UID: "1eed2052-fc62-4552-9ca6-3c23175b784a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.305227 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2bc45ee3-fd5f-4bfa-a06f-2786b12652d4-client-ca\") pod \"2bc45ee3-fd5f-4bfa-a06f-2786b12652d4\" (UID: \"2bc45ee3-fd5f-4bfa-a06f-2786b12652d4\") " Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.305275 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1eed2052-fc62-4552-9ca6-3c23175b784a-client-ca" (OuterVolumeSpecName: "client-ca") pod "1eed2052-fc62-4552-9ca6-3c23175b784a" (UID: "1eed2052-fc62-4552-9ca6-3c23175b784a"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.305472 4802 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1eed2052-fc62-4552-9ca6-3c23175b784a-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.305489 4802 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1eed2052-fc62-4552-9ca6-3c23175b784a-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.305517 4802 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2bc45ee3-fd5f-4bfa-a06f-2786b12652d4-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.305519 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2bc45ee3-fd5f-4bfa-a06f-2786b12652d4-client-ca" (OuterVolumeSpecName: "client-ca") pod "2bc45ee3-fd5f-4bfa-a06f-2786b12652d4" (UID: "2bc45ee3-fd5f-4bfa-a06f-2786b12652d4"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.308727 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1eed2052-fc62-4552-9ca6-3c23175b784a-kube-api-access-fv8nn" (OuterVolumeSpecName: "kube-api-access-fv8nn") pod "1eed2052-fc62-4552-9ca6-3c23175b784a" (UID: "1eed2052-fc62-4552-9ca6-3c23175b784a"). InnerVolumeSpecName "kube-api-access-fv8nn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.308764 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bc45ee3-fd5f-4bfa-a06f-2786b12652d4-kube-api-access-qww87" (OuterVolumeSpecName: "kube-api-access-qww87") pod "2bc45ee3-fd5f-4bfa-a06f-2786b12652d4" (UID: "2bc45ee3-fd5f-4bfa-a06f-2786b12652d4"). InnerVolumeSpecName "kube-api-access-qww87". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.309011 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bc45ee3-fd5f-4bfa-a06f-2786b12652d4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "2bc45ee3-fd5f-4bfa-a06f-2786b12652d4" (UID: "2bc45ee3-fd5f-4bfa-a06f-2786b12652d4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.323311 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1eed2052-fc62-4552-9ca6-3c23175b784a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1eed2052-fc62-4552-9ca6-3c23175b784a" (UID: "1eed2052-fc62-4552-9ca6-3c23175b784a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.399391 4802 generic.go:334] "Generic (PLEG): container finished" podID="1eed2052-fc62-4552-9ca6-3c23175b784a" containerID="de14875ffed3a903daba61d2fa8c74c65ee77330d42b3ca8eea6c3181a0770db" exitCode=0 Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.399472 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-d5d5747-vgjmr" event={"ID":"1eed2052-fc62-4552-9ca6-3c23175b784a","Type":"ContainerDied","Data":"de14875ffed3a903daba61d2fa8c74c65ee77330d42b3ca8eea6c3181a0770db"} Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.399503 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-d5d5747-vgjmr" Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.399533 4802 scope.go:117] "RemoveContainer" containerID="de14875ffed3a903daba61d2fa8c74c65ee77330d42b3ca8eea6c3181a0770db" Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.399517 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-d5d5747-vgjmr" event={"ID":"1eed2052-fc62-4552-9ca6-3c23175b784a","Type":"ContainerDied","Data":"c406bc8b7652cc175943a27d0c3e34b0b56a2c9d16e8d08f6564d0dff66eae59"} Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.401268 4802 generic.go:334] "Generic (PLEG): container finished" podID="2bc45ee3-fd5f-4bfa-a06f-2786b12652d4" containerID="cf1982c0eba1cf56b150a068b9994a31d6e8df9b5f9c0f34fef5d98cb22c09b3" exitCode=0 Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.401298 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-55985dff9-cfd84" event={"ID":"2bc45ee3-fd5f-4bfa-a06f-2786b12652d4","Type":"ContainerDied","Data":"cf1982c0eba1cf56b150a068b9994a31d6e8df9b5f9c0f34fef5d98cb22c09b3"} Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.401315 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-55985dff9-cfd84" event={"ID":"2bc45ee3-fd5f-4bfa-a06f-2786b12652d4","Type":"ContainerDied","Data":"20ff2381a09078ff043830cf2e80150fc4da9e4aa701b24ac7d8b828b8920d12"} Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.401360 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-55985dff9-cfd84" Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.407107 4802 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1eed2052-fc62-4552-9ca6-3c23175b784a-client-ca\") on node \"crc\" DevicePath \"\"" Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.407154 4802 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2bc45ee3-fd5f-4bfa-a06f-2786b12652d4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.407164 4802 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1eed2052-fc62-4552-9ca6-3c23175b784a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.407172 4802 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2bc45ee3-fd5f-4bfa-a06f-2786b12652d4-client-ca\") on node \"crc\" DevicePath \"\"" Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.407180 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fv8nn\" (UniqueName: \"kubernetes.io/projected/1eed2052-fc62-4552-9ca6-3c23175b784a-kube-api-access-fv8nn\") on node \"crc\" DevicePath \"\"" Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.407192 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qww87\" (UniqueName: \"kubernetes.io/projected/2bc45ee3-fd5f-4bfa-a06f-2786b12652d4-kube-api-access-qww87\") on node \"crc\" DevicePath \"\"" Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.432004 4802 scope.go:117] "RemoveContainer" containerID="de14875ffed3a903daba61d2fa8c74c65ee77330d42b3ca8eea6c3181a0770db" Dec 06 03:45:59 crc kubenswrapper[4802]: E1206 03:45:59.432520 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de14875ffed3a903daba61d2fa8c74c65ee77330d42b3ca8eea6c3181a0770db\": container with ID starting with de14875ffed3a903daba61d2fa8c74c65ee77330d42b3ca8eea6c3181a0770db not found: ID does not exist" containerID="de14875ffed3a903daba61d2fa8c74c65ee77330d42b3ca8eea6c3181a0770db" Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.432567 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de14875ffed3a903daba61d2fa8c74c65ee77330d42b3ca8eea6c3181a0770db"} err="failed to get container status \"de14875ffed3a903daba61d2fa8c74c65ee77330d42b3ca8eea6c3181a0770db\": rpc error: code = NotFound desc = could not find container \"de14875ffed3a903daba61d2fa8c74c65ee77330d42b3ca8eea6c3181a0770db\": container with ID starting with de14875ffed3a903daba61d2fa8c74c65ee77330d42b3ca8eea6c3181a0770db not found: ID does not exist" Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.432596 4802 scope.go:117] "RemoveContainer" containerID="cf1982c0eba1cf56b150a068b9994a31d6e8df9b5f9c0f34fef5d98cb22c09b3" Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.434175 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-55985dff9-cfd84"] Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.442739 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-55985dff9-cfd84"] Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.447109 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-d5d5747-vgjmr"] Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.457228 4802 scope.go:117] "RemoveContainer" containerID="cf1982c0eba1cf56b150a068b9994a31d6e8df9b5f9c0f34fef5d98cb22c09b3" Dec 06 03:45:59 crc kubenswrapper[4802]: E1206 03:45:59.457610 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf1982c0eba1cf56b150a068b9994a31d6e8df9b5f9c0f34fef5d98cb22c09b3\": container with ID starting with cf1982c0eba1cf56b150a068b9994a31d6e8df9b5f9c0f34fef5d98cb22c09b3 not found: ID does not exist" containerID="cf1982c0eba1cf56b150a068b9994a31d6e8df9b5f9c0f34fef5d98cb22c09b3" Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.457651 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf1982c0eba1cf56b150a068b9994a31d6e8df9b5f9c0f34fef5d98cb22c09b3"} err="failed to get container status \"cf1982c0eba1cf56b150a068b9994a31d6e8df9b5f9c0f34fef5d98cb22c09b3\": rpc error: code = NotFound desc = could not find container \"cf1982c0eba1cf56b150a068b9994a31d6e8df9b5f9c0f34fef5d98cb22c09b3\": container with ID starting with cf1982c0eba1cf56b150a068b9994a31d6e8df9b5f9c0f34fef5d98cb22c09b3 not found: ID does not exist" Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.457713 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2bc45ee3-fd5f-4bfa-a06f-2786b12652d4" path="/var/lib/kubelet/pods/2bc45ee3-fd5f-4bfa-a06f-2786b12652d4/volumes" Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.458244 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-d5d5747-vgjmr"] Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.918550 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-585df4fbc7-rbtxj"] Dec 06 03:45:59 crc kubenswrapper[4802]: E1206 03:45:59.919073 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bc45ee3-fd5f-4bfa-a06f-2786b12652d4" containerName="route-controller-manager" Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.919089 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bc45ee3-fd5f-4bfa-a06f-2786b12652d4" containerName="route-controller-manager" Dec 06 03:45:59 crc kubenswrapper[4802]: E1206 03:45:59.919110 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1eed2052-fc62-4552-9ca6-3c23175b784a" containerName="controller-manager" Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.919118 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="1eed2052-fc62-4552-9ca6-3c23175b784a" containerName="controller-manager" Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.919247 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bc45ee3-fd5f-4bfa-a06f-2786b12652d4" containerName="route-controller-manager" Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.919274 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="1eed2052-fc62-4552-9ca6-3c23175b784a" containerName="controller-manager" Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.919728 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-585df4fbc7-rbtxj" Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.924983 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-67644cbbd7-6vq75"] Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.925676 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-67644cbbd7-6vq75" Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.926234 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.926387 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.927163 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.927261 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.927305 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.928485 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.928570 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.928853 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.930084 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.935033 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.935191 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.936387 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.937916 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 06 03:45:59 crc kubenswrapper[4802]: I1206 03:45:59.993375 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-585df4fbc7-rbtxj"] Dec 06 03:46:00 crc kubenswrapper[4802]: I1206 03:46:00.012096 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-67644cbbd7-6vq75"] Dec 06 03:46:00 crc kubenswrapper[4802]: I1206 03:46:00.013594 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99f3a0d5-71a0-476a-8c0b-8f287aed76a4-config\") pod \"controller-manager-585df4fbc7-rbtxj\" (UID: \"99f3a0d5-71a0-476a-8c0b-8f287aed76a4\") " pod="openshift-controller-manager/controller-manager-585df4fbc7-rbtxj" Dec 06 03:46:00 crc kubenswrapper[4802]: I1206 03:46:00.013652 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/99f3a0d5-71a0-476a-8c0b-8f287aed76a4-client-ca\") pod \"controller-manager-585df4fbc7-rbtxj\" (UID: \"99f3a0d5-71a0-476a-8c0b-8f287aed76a4\") " pod="openshift-controller-manager/controller-manager-585df4fbc7-rbtxj" Dec 06 03:46:00 crc kubenswrapper[4802]: I1206 03:46:00.013694 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d291f291-567c-4b2b-9641-cce89ef3636d-serving-cert\") pod \"route-controller-manager-67644cbbd7-6vq75\" (UID: \"d291f291-567c-4b2b-9641-cce89ef3636d\") " pod="openshift-route-controller-manager/route-controller-manager-67644cbbd7-6vq75" Dec 06 03:46:00 crc kubenswrapper[4802]: I1206 03:46:00.013723 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsjkv\" (UniqueName: \"kubernetes.io/projected/99f3a0d5-71a0-476a-8c0b-8f287aed76a4-kube-api-access-vsjkv\") pod \"controller-manager-585df4fbc7-rbtxj\" (UID: \"99f3a0d5-71a0-476a-8c0b-8f287aed76a4\") " pod="openshift-controller-manager/controller-manager-585df4fbc7-rbtxj" Dec 06 03:46:00 crc kubenswrapper[4802]: I1206 03:46:00.013768 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d291f291-567c-4b2b-9641-cce89ef3636d-config\") pod \"route-controller-manager-67644cbbd7-6vq75\" (UID: \"d291f291-567c-4b2b-9641-cce89ef3636d\") " pod="openshift-route-controller-manager/route-controller-manager-67644cbbd7-6vq75" Dec 06 03:46:00 crc kubenswrapper[4802]: I1206 03:46:00.013791 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgmg6\" (UniqueName: \"kubernetes.io/projected/d291f291-567c-4b2b-9641-cce89ef3636d-kube-api-access-cgmg6\") pod \"route-controller-manager-67644cbbd7-6vq75\" (UID: \"d291f291-567c-4b2b-9641-cce89ef3636d\") " pod="openshift-route-controller-manager/route-controller-manager-67644cbbd7-6vq75" Dec 06 03:46:00 crc kubenswrapper[4802]: I1206 03:46:00.013816 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/99f3a0d5-71a0-476a-8c0b-8f287aed76a4-proxy-ca-bundles\") pod \"controller-manager-585df4fbc7-rbtxj\" (UID: \"99f3a0d5-71a0-476a-8c0b-8f287aed76a4\") " pod="openshift-controller-manager/controller-manager-585df4fbc7-rbtxj" Dec 06 03:46:00 crc kubenswrapper[4802]: I1206 03:46:00.013843 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/99f3a0d5-71a0-476a-8c0b-8f287aed76a4-serving-cert\") pod \"controller-manager-585df4fbc7-rbtxj\" (UID: \"99f3a0d5-71a0-476a-8c0b-8f287aed76a4\") " pod="openshift-controller-manager/controller-manager-585df4fbc7-rbtxj" Dec 06 03:46:00 crc kubenswrapper[4802]: I1206 03:46:00.013881 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d291f291-567c-4b2b-9641-cce89ef3636d-client-ca\") pod \"route-controller-manager-67644cbbd7-6vq75\" (UID: \"d291f291-567c-4b2b-9641-cce89ef3636d\") " pod="openshift-route-controller-manager/route-controller-manager-67644cbbd7-6vq75" Dec 06 03:46:00 crc kubenswrapper[4802]: I1206 03:46:00.114985 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d291f291-567c-4b2b-9641-cce89ef3636d-client-ca\") pod \"route-controller-manager-67644cbbd7-6vq75\" (UID: \"d291f291-567c-4b2b-9641-cce89ef3636d\") " pod="openshift-route-controller-manager/route-controller-manager-67644cbbd7-6vq75" Dec 06 03:46:00 crc kubenswrapper[4802]: I1206 03:46:00.115302 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99f3a0d5-71a0-476a-8c0b-8f287aed76a4-config\") pod \"controller-manager-585df4fbc7-rbtxj\" (UID: \"99f3a0d5-71a0-476a-8c0b-8f287aed76a4\") " pod="openshift-controller-manager/controller-manager-585df4fbc7-rbtxj" Dec 06 03:46:00 crc kubenswrapper[4802]: I1206 03:46:00.115448 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d291f291-567c-4b2b-9641-cce89ef3636d-serving-cert\") pod \"route-controller-manager-67644cbbd7-6vq75\" (UID: \"d291f291-567c-4b2b-9641-cce89ef3636d\") " pod="openshift-route-controller-manager/route-controller-manager-67644cbbd7-6vq75" Dec 06 03:46:00 crc kubenswrapper[4802]: I1206 03:46:00.115568 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/99f3a0d5-71a0-476a-8c0b-8f287aed76a4-client-ca\") pod \"controller-manager-585df4fbc7-rbtxj\" (UID: \"99f3a0d5-71a0-476a-8c0b-8f287aed76a4\") " pod="openshift-controller-manager/controller-manager-585df4fbc7-rbtxj" Dec 06 03:46:00 crc kubenswrapper[4802]: I1206 03:46:00.115696 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsjkv\" (UniqueName: \"kubernetes.io/projected/99f3a0d5-71a0-476a-8c0b-8f287aed76a4-kube-api-access-vsjkv\") pod \"controller-manager-585df4fbc7-rbtxj\" (UID: \"99f3a0d5-71a0-476a-8c0b-8f287aed76a4\") " pod="openshift-controller-manager/controller-manager-585df4fbc7-rbtxj" Dec 06 03:46:00 crc kubenswrapper[4802]: I1206 03:46:00.115879 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d291f291-567c-4b2b-9641-cce89ef3636d-config\") pod \"route-controller-manager-67644cbbd7-6vq75\" (UID: \"d291f291-567c-4b2b-9641-cce89ef3636d\") " pod="openshift-route-controller-manager/route-controller-manager-67644cbbd7-6vq75" Dec 06 03:46:00 crc kubenswrapper[4802]: I1206 03:46:00.116013 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgmg6\" (UniqueName: \"kubernetes.io/projected/d291f291-567c-4b2b-9641-cce89ef3636d-kube-api-access-cgmg6\") pod \"route-controller-manager-67644cbbd7-6vq75\" (UID: \"d291f291-567c-4b2b-9641-cce89ef3636d\") " pod="openshift-route-controller-manager/route-controller-manager-67644cbbd7-6vq75" Dec 06 03:46:00 crc kubenswrapper[4802]: I1206 03:46:00.116137 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/99f3a0d5-71a0-476a-8c0b-8f287aed76a4-proxy-ca-bundles\") pod \"controller-manager-585df4fbc7-rbtxj\" (UID: \"99f3a0d5-71a0-476a-8c0b-8f287aed76a4\") " pod="openshift-controller-manager/controller-manager-585df4fbc7-rbtxj" Dec 06 03:46:00 crc kubenswrapper[4802]: I1206 03:46:00.116217 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d291f291-567c-4b2b-9641-cce89ef3636d-client-ca\") pod \"route-controller-manager-67644cbbd7-6vq75\" (UID: \"d291f291-567c-4b2b-9641-cce89ef3636d\") " pod="openshift-route-controller-manager/route-controller-manager-67644cbbd7-6vq75" Dec 06 03:46:00 crc kubenswrapper[4802]: I1206 03:46:00.116344 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/99f3a0d5-71a0-476a-8c0b-8f287aed76a4-serving-cert\") pod \"controller-manager-585df4fbc7-rbtxj\" (UID: \"99f3a0d5-71a0-476a-8c0b-8f287aed76a4\") " pod="openshift-controller-manager/controller-manager-585df4fbc7-rbtxj" Dec 06 03:46:00 crc kubenswrapper[4802]: I1206 03:46:00.116768 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99f3a0d5-71a0-476a-8c0b-8f287aed76a4-config\") pod \"controller-manager-585df4fbc7-rbtxj\" (UID: \"99f3a0d5-71a0-476a-8c0b-8f287aed76a4\") " pod="openshift-controller-manager/controller-manager-585df4fbc7-rbtxj" Dec 06 03:46:00 crc kubenswrapper[4802]: I1206 03:46:00.117383 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/99f3a0d5-71a0-476a-8c0b-8f287aed76a4-client-ca\") pod \"controller-manager-585df4fbc7-rbtxj\" (UID: \"99f3a0d5-71a0-476a-8c0b-8f287aed76a4\") " pod="openshift-controller-manager/controller-manager-585df4fbc7-rbtxj" Dec 06 03:46:00 crc kubenswrapper[4802]: I1206 03:46:00.117420 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d291f291-567c-4b2b-9641-cce89ef3636d-config\") pod \"route-controller-manager-67644cbbd7-6vq75\" (UID: \"d291f291-567c-4b2b-9641-cce89ef3636d\") " pod="openshift-route-controller-manager/route-controller-manager-67644cbbd7-6vq75" Dec 06 03:46:00 crc kubenswrapper[4802]: I1206 03:46:00.117648 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/99f3a0d5-71a0-476a-8c0b-8f287aed76a4-proxy-ca-bundles\") pod \"controller-manager-585df4fbc7-rbtxj\" (UID: \"99f3a0d5-71a0-476a-8c0b-8f287aed76a4\") " pod="openshift-controller-manager/controller-manager-585df4fbc7-rbtxj" Dec 06 03:46:00 crc kubenswrapper[4802]: I1206 03:46:00.122102 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d291f291-567c-4b2b-9641-cce89ef3636d-serving-cert\") pod \"route-controller-manager-67644cbbd7-6vq75\" (UID: \"d291f291-567c-4b2b-9641-cce89ef3636d\") " pod="openshift-route-controller-manager/route-controller-manager-67644cbbd7-6vq75" Dec 06 03:46:00 crc kubenswrapper[4802]: I1206 03:46:00.131588 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/99f3a0d5-71a0-476a-8c0b-8f287aed76a4-serving-cert\") pod \"controller-manager-585df4fbc7-rbtxj\" (UID: \"99f3a0d5-71a0-476a-8c0b-8f287aed76a4\") " pod="openshift-controller-manager/controller-manager-585df4fbc7-rbtxj" Dec 06 03:46:00 crc kubenswrapper[4802]: I1206 03:46:00.133644 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsjkv\" (UniqueName: \"kubernetes.io/projected/99f3a0d5-71a0-476a-8c0b-8f287aed76a4-kube-api-access-vsjkv\") pod \"controller-manager-585df4fbc7-rbtxj\" (UID: \"99f3a0d5-71a0-476a-8c0b-8f287aed76a4\") " pod="openshift-controller-manager/controller-manager-585df4fbc7-rbtxj" Dec 06 03:46:00 crc kubenswrapper[4802]: I1206 03:46:00.147553 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgmg6\" (UniqueName: \"kubernetes.io/projected/d291f291-567c-4b2b-9641-cce89ef3636d-kube-api-access-cgmg6\") pod \"route-controller-manager-67644cbbd7-6vq75\" (UID: \"d291f291-567c-4b2b-9641-cce89ef3636d\") " pod="openshift-route-controller-manager/route-controller-manager-67644cbbd7-6vq75" Dec 06 03:46:00 crc kubenswrapper[4802]: I1206 03:46:00.260986 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-585df4fbc7-rbtxj" Dec 06 03:46:00 crc kubenswrapper[4802]: I1206 03:46:00.274867 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-67644cbbd7-6vq75" Dec 06 03:46:00 crc kubenswrapper[4802]: I1206 03:46:00.713023 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-67644cbbd7-6vq75"] Dec 06 03:46:00 crc kubenswrapper[4802]: I1206 03:46:00.745308 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-585df4fbc7-rbtxj"] Dec 06 03:46:00 crc kubenswrapper[4802]: W1206 03:46:00.747535 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod99f3a0d5_71a0_476a_8c0b_8f287aed76a4.slice/crio-8505db1ef3f2c67c246bcd5441c2aab8c507dc9bd7e1a78cd2faa6655fc18938 WatchSource:0}: Error finding container 8505db1ef3f2c67c246bcd5441c2aab8c507dc9bd7e1a78cd2faa6655fc18938: Status 404 returned error can't find the container with id 8505db1ef3f2c67c246bcd5441c2aab8c507dc9bd7e1a78cd2faa6655fc18938 Dec 06 03:46:01 crc kubenswrapper[4802]: I1206 03:46:01.418205 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-585df4fbc7-rbtxj" event={"ID":"99f3a0d5-71a0-476a-8c0b-8f287aed76a4","Type":"ContainerStarted","Data":"31c20dd933fbe4b1ca6b72583e8dccd361556bdd81eb9bf5c57c590199b0daea"} Dec 06 03:46:01 crc kubenswrapper[4802]: I1206 03:46:01.418506 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-585df4fbc7-rbtxj" event={"ID":"99f3a0d5-71a0-476a-8c0b-8f287aed76a4","Type":"ContainerStarted","Data":"8505db1ef3f2c67c246bcd5441c2aab8c507dc9bd7e1a78cd2faa6655fc18938"} Dec 06 03:46:01 crc kubenswrapper[4802]: I1206 03:46:01.418528 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-585df4fbc7-rbtxj" Dec 06 03:46:01 crc kubenswrapper[4802]: I1206 03:46:01.420567 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-67644cbbd7-6vq75" event={"ID":"d291f291-567c-4b2b-9641-cce89ef3636d","Type":"ContainerStarted","Data":"28c8a3de096d6b326b586bb5534654d7f8c06c58be477b6c9bc2c6ce3ca530e3"} Dec 06 03:46:01 crc kubenswrapper[4802]: I1206 03:46:01.420607 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-67644cbbd7-6vq75" event={"ID":"d291f291-567c-4b2b-9641-cce89ef3636d","Type":"ContainerStarted","Data":"5e19376c567ca234f2dd29edce9541ddcaf54c9fa59d3080a56357c933e4e257"} Dec 06 03:46:01 crc kubenswrapper[4802]: I1206 03:46:01.420994 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-67644cbbd7-6vq75" Dec 06 03:46:01 crc kubenswrapper[4802]: I1206 03:46:01.423286 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-585df4fbc7-rbtxj" Dec 06 03:46:01 crc kubenswrapper[4802]: I1206 03:46:01.427956 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-67644cbbd7-6vq75" Dec 06 03:46:01 crc kubenswrapper[4802]: I1206 03:46:01.435728 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-585df4fbc7-rbtxj" podStartSLOduration=3.435710363 podStartE2EDuration="3.435710363s" podCreationTimestamp="2025-12-06 03:45:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:46:01.434301844 +0000 UTC m=+354.306211006" watchObservedRunningTime="2025-12-06 03:46:01.435710363 +0000 UTC m=+354.307619515" Dec 06 03:46:01 crc kubenswrapper[4802]: I1206 03:46:01.457671 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-67644cbbd7-6vq75" podStartSLOduration=3.457651039 podStartE2EDuration="3.457651039s" podCreationTimestamp="2025-12-06 03:45:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:46:01.454316536 +0000 UTC m=+354.326225698" watchObservedRunningTime="2025-12-06 03:46:01.457651039 +0000 UTC m=+354.329560191" Dec 06 03:46:01 crc kubenswrapper[4802]: I1206 03:46:01.460633 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1eed2052-fc62-4552-9ca6-3c23175b784a" path="/var/lib/kubelet/pods/1eed2052-fc62-4552-9ca6-3c23175b784a/volumes" Dec 06 03:46:13 crc kubenswrapper[4802]: I1206 03:46:13.283352 4802 patch_prober.go:28] interesting pod/machine-config-daemon-zpxxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 03:46:13 crc kubenswrapper[4802]: I1206 03:46:13.284021 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 03:46:28 crc kubenswrapper[4802]: I1206 03:46:28.412217 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-vnm59"] Dec 06 03:46:28 crc kubenswrapper[4802]: I1206 03:46:28.413546 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-vnm59" Dec 06 03:46:28 crc kubenswrapper[4802]: I1206 03:46:28.429929 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-vnm59"] Dec 06 03:46:28 crc kubenswrapper[4802]: I1206 03:46:28.593734 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/af28a8c0-252b-4178-98c3-2abcfbcbd167-trusted-ca\") pod \"image-registry-66df7c8f76-vnm59\" (UID: \"af28a8c0-252b-4178-98c3-2abcfbcbd167\") " pod="openshift-image-registry/image-registry-66df7c8f76-vnm59" Dec 06 03:46:28 crc kubenswrapper[4802]: I1206 03:46:28.593822 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/af28a8c0-252b-4178-98c3-2abcfbcbd167-ca-trust-extracted\") pod \"image-registry-66df7c8f76-vnm59\" (UID: \"af28a8c0-252b-4178-98c3-2abcfbcbd167\") " pod="openshift-image-registry/image-registry-66df7c8f76-vnm59" Dec 06 03:46:28 crc kubenswrapper[4802]: I1206 03:46:28.593851 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sr668\" (UniqueName: \"kubernetes.io/projected/af28a8c0-252b-4178-98c3-2abcfbcbd167-kube-api-access-sr668\") pod \"image-registry-66df7c8f76-vnm59\" (UID: \"af28a8c0-252b-4178-98c3-2abcfbcbd167\") " pod="openshift-image-registry/image-registry-66df7c8f76-vnm59" Dec 06 03:46:28 crc kubenswrapper[4802]: I1206 03:46:28.593904 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-vnm59\" (UID: \"af28a8c0-252b-4178-98c3-2abcfbcbd167\") " pod="openshift-image-registry/image-registry-66df7c8f76-vnm59" Dec 06 03:46:28 crc kubenswrapper[4802]: I1206 03:46:28.593931 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/af28a8c0-252b-4178-98c3-2abcfbcbd167-registry-certificates\") pod \"image-registry-66df7c8f76-vnm59\" (UID: \"af28a8c0-252b-4178-98c3-2abcfbcbd167\") " pod="openshift-image-registry/image-registry-66df7c8f76-vnm59" Dec 06 03:46:28 crc kubenswrapper[4802]: I1206 03:46:28.593952 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/af28a8c0-252b-4178-98c3-2abcfbcbd167-registry-tls\") pod \"image-registry-66df7c8f76-vnm59\" (UID: \"af28a8c0-252b-4178-98c3-2abcfbcbd167\") " pod="openshift-image-registry/image-registry-66df7c8f76-vnm59" Dec 06 03:46:28 crc kubenswrapper[4802]: I1206 03:46:28.594079 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/af28a8c0-252b-4178-98c3-2abcfbcbd167-bound-sa-token\") pod \"image-registry-66df7c8f76-vnm59\" (UID: \"af28a8c0-252b-4178-98c3-2abcfbcbd167\") " pod="openshift-image-registry/image-registry-66df7c8f76-vnm59" Dec 06 03:46:28 crc kubenswrapper[4802]: I1206 03:46:28.594380 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/af28a8c0-252b-4178-98c3-2abcfbcbd167-installation-pull-secrets\") pod \"image-registry-66df7c8f76-vnm59\" (UID: \"af28a8c0-252b-4178-98c3-2abcfbcbd167\") " pod="openshift-image-registry/image-registry-66df7c8f76-vnm59" Dec 06 03:46:28 crc kubenswrapper[4802]: I1206 03:46:28.628304 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-vnm59\" (UID: \"af28a8c0-252b-4178-98c3-2abcfbcbd167\") " pod="openshift-image-registry/image-registry-66df7c8f76-vnm59" Dec 06 03:46:28 crc kubenswrapper[4802]: I1206 03:46:28.695501 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/af28a8c0-252b-4178-98c3-2abcfbcbd167-trusted-ca\") pod \"image-registry-66df7c8f76-vnm59\" (UID: \"af28a8c0-252b-4178-98c3-2abcfbcbd167\") " pod="openshift-image-registry/image-registry-66df7c8f76-vnm59" Dec 06 03:46:28 crc kubenswrapper[4802]: I1206 03:46:28.695587 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/af28a8c0-252b-4178-98c3-2abcfbcbd167-ca-trust-extracted\") pod \"image-registry-66df7c8f76-vnm59\" (UID: \"af28a8c0-252b-4178-98c3-2abcfbcbd167\") " pod="openshift-image-registry/image-registry-66df7c8f76-vnm59" Dec 06 03:46:28 crc kubenswrapper[4802]: I1206 03:46:28.695637 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sr668\" (UniqueName: \"kubernetes.io/projected/af28a8c0-252b-4178-98c3-2abcfbcbd167-kube-api-access-sr668\") pod \"image-registry-66df7c8f76-vnm59\" (UID: \"af28a8c0-252b-4178-98c3-2abcfbcbd167\") " pod="openshift-image-registry/image-registry-66df7c8f76-vnm59" Dec 06 03:46:28 crc kubenswrapper[4802]: I1206 03:46:28.695785 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/af28a8c0-252b-4178-98c3-2abcfbcbd167-registry-certificates\") pod \"image-registry-66df7c8f76-vnm59\" (UID: \"af28a8c0-252b-4178-98c3-2abcfbcbd167\") " pod="openshift-image-registry/image-registry-66df7c8f76-vnm59" Dec 06 03:46:28 crc kubenswrapper[4802]: I1206 03:46:28.695828 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/af28a8c0-252b-4178-98c3-2abcfbcbd167-registry-tls\") pod \"image-registry-66df7c8f76-vnm59\" (UID: \"af28a8c0-252b-4178-98c3-2abcfbcbd167\") " pod="openshift-image-registry/image-registry-66df7c8f76-vnm59" Dec 06 03:46:28 crc kubenswrapper[4802]: I1206 03:46:28.695870 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/af28a8c0-252b-4178-98c3-2abcfbcbd167-bound-sa-token\") pod \"image-registry-66df7c8f76-vnm59\" (UID: \"af28a8c0-252b-4178-98c3-2abcfbcbd167\") " pod="openshift-image-registry/image-registry-66df7c8f76-vnm59" Dec 06 03:46:28 crc kubenswrapper[4802]: I1206 03:46:28.695916 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/af28a8c0-252b-4178-98c3-2abcfbcbd167-installation-pull-secrets\") pod \"image-registry-66df7c8f76-vnm59\" (UID: \"af28a8c0-252b-4178-98c3-2abcfbcbd167\") " pod="openshift-image-registry/image-registry-66df7c8f76-vnm59" Dec 06 03:46:28 crc kubenswrapper[4802]: I1206 03:46:28.696505 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/af28a8c0-252b-4178-98c3-2abcfbcbd167-ca-trust-extracted\") pod \"image-registry-66df7c8f76-vnm59\" (UID: \"af28a8c0-252b-4178-98c3-2abcfbcbd167\") " pod="openshift-image-registry/image-registry-66df7c8f76-vnm59" Dec 06 03:46:28 crc kubenswrapper[4802]: I1206 03:46:28.697538 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/af28a8c0-252b-4178-98c3-2abcfbcbd167-trusted-ca\") pod \"image-registry-66df7c8f76-vnm59\" (UID: \"af28a8c0-252b-4178-98c3-2abcfbcbd167\") " pod="openshift-image-registry/image-registry-66df7c8f76-vnm59" Dec 06 03:46:28 crc kubenswrapper[4802]: I1206 03:46:28.697713 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/af28a8c0-252b-4178-98c3-2abcfbcbd167-registry-certificates\") pod \"image-registry-66df7c8f76-vnm59\" (UID: \"af28a8c0-252b-4178-98c3-2abcfbcbd167\") " pod="openshift-image-registry/image-registry-66df7c8f76-vnm59" Dec 06 03:46:28 crc kubenswrapper[4802]: I1206 03:46:28.705263 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/af28a8c0-252b-4178-98c3-2abcfbcbd167-installation-pull-secrets\") pod \"image-registry-66df7c8f76-vnm59\" (UID: \"af28a8c0-252b-4178-98c3-2abcfbcbd167\") " pod="openshift-image-registry/image-registry-66df7c8f76-vnm59" Dec 06 03:46:28 crc kubenswrapper[4802]: I1206 03:46:28.709975 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/af28a8c0-252b-4178-98c3-2abcfbcbd167-registry-tls\") pod \"image-registry-66df7c8f76-vnm59\" (UID: \"af28a8c0-252b-4178-98c3-2abcfbcbd167\") " pod="openshift-image-registry/image-registry-66df7c8f76-vnm59" Dec 06 03:46:28 crc kubenswrapper[4802]: I1206 03:46:28.718866 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sr668\" (UniqueName: \"kubernetes.io/projected/af28a8c0-252b-4178-98c3-2abcfbcbd167-kube-api-access-sr668\") pod \"image-registry-66df7c8f76-vnm59\" (UID: \"af28a8c0-252b-4178-98c3-2abcfbcbd167\") " pod="openshift-image-registry/image-registry-66df7c8f76-vnm59" Dec 06 03:46:28 crc kubenswrapper[4802]: I1206 03:46:28.731637 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/af28a8c0-252b-4178-98c3-2abcfbcbd167-bound-sa-token\") pod \"image-registry-66df7c8f76-vnm59\" (UID: \"af28a8c0-252b-4178-98c3-2abcfbcbd167\") " pod="openshift-image-registry/image-registry-66df7c8f76-vnm59" Dec 06 03:46:29 crc kubenswrapper[4802]: I1206 03:46:29.032528 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-vnm59" Dec 06 03:46:29 crc kubenswrapper[4802]: I1206 03:46:29.498551 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-vnm59"] Dec 06 03:46:29 crc kubenswrapper[4802]: I1206 03:46:29.578546 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-vnm59" event={"ID":"af28a8c0-252b-4178-98c3-2abcfbcbd167","Type":"ContainerStarted","Data":"b8bb7e80c97d60573a17b6a3b0ed656dad956a31542f466d215d82e139ed4d27"} Dec 06 03:46:30 crc kubenswrapper[4802]: I1206 03:46:30.587806 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-vnm59" event={"ID":"af28a8c0-252b-4178-98c3-2abcfbcbd167","Type":"ContainerStarted","Data":"723c7388f4698892736851bc0be58f9faeb30aa5f059cda6088b9a4da563de87"} Dec 06 03:46:30 crc kubenswrapper[4802]: I1206 03:46:30.588081 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-vnm59" Dec 06 03:46:30 crc kubenswrapper[4802]: I1206 03:46:30.617390 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-vnm59" podStartSLOduration=2.617362387 podStartE2EDuration="2.617362387s" podCreationTimestamp="2025-12-06 03:46:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:46:30.606902193 +0000 UTC m=+383.478811365" watchObservedRunningTime="2025-12-06 03:46:30.617362387 +0000 UTC m=+383.489271579" Dec 06 03:46:43 crc kubenswrapper[4802]: I1206 03:46:43.283571 4802 patch_prober.go:28] interesting pod/machine-config-daemon-zpxxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 03:46:43 crc kubenswrapper[4802]: I1206 03:46:43.284277 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 03:46:49 crc kubenswrapper[4802]: I1206 03:46:49.040266 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-vnm59" Dec 06 03:46:49 crc kubenswrapper[4802]: I1206 03:46:49.118740 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-blwnp"] Dec 06 03:47:07 crc kubenswrapper[4802]: I1206 03:47:07.619142 4802 scope.go:117] "RemoveContainer" containerID="9a21a97385f856ec996105b597d0ae81c56ed28b625ef3f2905e1af6ec8f6a45" Dec 06 03:47:07 crc kubenswrapper[4802]: I1206 03:47:07.641636 4802 scope.go:117] "RemoveContainer" containerID="16a92d501c9e626fe91e4a6793e426d10eb292cfbab766db8b8cc72e5dae94ef" Dec 06 03:47:07 crc kubenswrapper[4802]: I1206 03:47:07.657873 4802 scope.go:117] "RemoveContainer" containerID="646e01a58908ba2e703aae5eef53e1b981b0f0c3d1d3202487f2a8b94ee3f55d" Dec 06 03:47:07 crc kubenswrapper[4802]: I1206 03:47:07.684087 4802 scope.go:117] "RemoveContainer" containerID="31752b84d86bda30792b29e4c6ed963753cc576c5de784395e8e6dcbebaced47" Dec 06 03:47:07 crc kubenswrapper[4802]: I1206 03:47:07.700898 4802 scope.go:117] "RemoveContainer" containerID="1e8a04026ff3baf22afbc3e9fd7e6207a0c022eb156e5b6a2e64603dc6fcc3ce" Dec 06 03:47:07 crc kubenswrapper[4802]: I1206 03:47:07.724986 4802 scope.go:117] "RemoveContainer" containerID="a895e0cc04880c3287ba725ac5484835235f1e066a94ce98fcf4b6133c5cbd2f" Dec 06 03:47:13 crc kubenswrapper[4802]: I1206 03:47:13.284031 4802 patch_prober.go:28] interesting pod/machine-config-daemon-zpxxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 03:47:13 crc kubenswrapper[4802]: I1206 03:47:13.284452 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 03:47:13 crc kubenswrapper[4802]: I1206 03:47:13.284528 4802 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" Dec 06 03:47:13 crc kubenswrapper[4802]: I1206 03:47:13.285464 4802 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2200194f7108affdd47802bf08946e25d3e937d094def5243f711369c5ae1305"} pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 03:47:13 crc kubenswrapper[4802]: I1206 03:47:13.285552 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" containerID="cri-o://2200194f7108affdd47802bf08946e25d3e937d094def5243f711369c5ae1305" gracePeriod=600 Dec 06 03:47:13 crc kubenswrapper[4802]: I1206 03:47:13.881119 4802 generic.go:334] "Generic (PLEG): container finished" podID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerID="2200194f7108affdd47802bf08946e25d3e937d094def5243f711369c5ae1305" exitCode=0 Dec 06 03:47:13 crc kubenswrapper[4802]: I1206 03:47:13.881315 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" event={"ID":"6df38316-e0d3-4018-8d27-3620eba3a68d","Type":"ContainerDied","Data":"2200194f7108affdd47802bf08946e25d3e937d094def5243f711369c5ae1305"} Dec 06 03:47:13 crc kubenswrapper[4802]: I1206 03:47:13.882222 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" event={"ID":"6df38316-e0d3-4018-8d27-3620eba3a68d","Type":"ContainerStarted","Data":"25f142fb563466ee1de9442e9c6cd18bf7c0126b2d7681419cdb7a7267f4a6e6"} Dec 06 03:47:13 crc kubenswrapper[4802]: I1206 03:47:13.882273 4802 scope.go:117] "RemoveContainer" containerID="3bf5e41701c7a2d45adc4740ceb4ac1785db17bc7cb15daed439f7556812d1e9" Dec 06 03:47:14 crc kubenswrapper[4802]: I1206 03:47:14.176150 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" podUID="d13a541e-f90d-4222-8486-2d47854ac8e5" containerName="registry" containerID="cri-o://ed30078e684092fe648afb023cfb75ac7a68756ac6e18783bdb323fe369bec1f" gracePeriod=30 Dec 06 03:47:14 crc kubenswrapper[4802]: I1206 03:47:14.647412 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:47:14 crc kubenswrapper[4802]: I1206 03:47:14.686622 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d13a541e-f90d-4222-8486-2d47854ac8e5-installation-pull-secrets\") pod \"d13a541e-f90d-4222-8486-2d47854ac8e5\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " Dec 06 03:47:14 crc kubenswrapper[4802]: I1206 03:47:14.687174 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"d13a541e-f90d-4222-8486-2d47854ac8e5\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " Dec 06 03:47:14 crc kubenswrapper[4802]: I1206 03:47:14.687251 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d13a541e-f90d-4222-8486-2d47854ac8e5-bound-sa-token\") pod \"d13a541e-f90d-4222-8486-2d47854ac8e5\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " Dec 06 03:47:14 crc kubenswrapper[4802]: I1206 03:47:14.687324 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d13a541e-f90d-4222-8486-2d47854ac8e5-registry-tls\") pod \"d13a541e-f90d-4222-8486-2d47854ac8e5\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " Dec 06 03:47:14 crc kubenswrapper[4802]: I1206 03:47:14.687400 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d13a541e-f90d-4222-8486-2d47854ac8e5-registry-certificates\") pod \"d13a541e-f90d-4222-8486-2d47854ac8e5\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " Dec 06 03:47:14 crc kubenswrapper[4802]: I1206 03:47:14.687421 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d13a541e-f90d-4222-8486-2d47854ac8e5-trusted-ca\") pod \"d13a541e-f90d-4222-8486-2d47854ac8e5\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " Dec 06 03:47:14 crc kubenswrapper[4802]: I1206 03:47:14.688187 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tzkg4\" (UniqueName: \"kubernetes.io/projected/d13a541e-f90d-4222-8486-2d47854ac8e5-kube-api-access-tzkg4\") pod \"d13a541e-f90d-4222-8486-2d47854ac8e5\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " Dec 06 03:47:14 crc kubenswrapper[4802]: I1206 03:47:14.688124 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d13a541e-f90d-4222-8486-2d47854ac8e5-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "d13a541e-f90d-4222-8486-2d47854ac8e5" (UID: "d13a541e-f90d-4222-8486-2d47854ac8e5"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:47:14 crc kubenswrapper[4802]: I1206 03:47:14.688234 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d13a541e-f90d-4222-8486-2d47854ac8e5-ca-trust-extracted\") pod \"d13a541e-f90d-4222-8486-2d47854ac8e5\" (UID: \"d13a541e-f90d-4222-8486-2d47854ac8e5\") " Dec 06 03:47:14 crc kubenswrapper[4802]: I1206 03:47:14.688138 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d13a541e-f90d-4222-8486-2d47854ac8e5-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "d13a541e-f90d-4222-8486-2d47854ac8e5" (UID: "d13a541e-f90d-4222-8486-2d47854ac8e5"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:47:14 crc kubenswrapper[4802]: I1206 03:47:14.692872 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d13a541e-f90d-4222-8486-2d47854ac8e5-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "d13a541e-f90d-4222-8486-2d47854ac8e5" (UID: "d13a541e-f90d-4222-8486-2d47854ac8e5"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:47:14 crc kubenswrapper[4802]: I1206 03:47:14.693186 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d13a541e-f90d-4222-8486-2d47854ac8e5-kube-api-access-tzkg4" (OuterVolumeSpecName: "kube-api-access-tzkg4") pod "d13a541e-f90d-4222-8486-2d47854ac8e5" (UID: "d13a541e-f90d-4222-8486-2d47854ac8e5"). InnerVolumeSpecName "kube-api-access-tzkg4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:47:14 crc kubenswrapper[4802]: I1206 03:47:14.694293 4802 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d13a541e-f90d-4222-8486-2d47854ac8e5-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 06 03:47:14 crc kubenswrapper[4802]: I1206 03:47:14.694336 4802 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d13a541e-f90d-4222-8486-2d47854ac8e5-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 03:47:14 crc kubenswrapper[4802]: I1206 03:47:14.694358 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tzkg4\" (UniqueName: \"kubernetes.io/projected/d13a541e-f90d-4222-8486-2d47854ac8e5-kube-api-access-tzkg4\") on node \"crc\" DevicePath \"\"" Dec 06 03:47:14 crc kubenswrapper[4802]: I1206 03:47:14.694378 4802 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d13a541e-f90d-4222-8486-2d47854ac8e5-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 06 03:47:14 crc kubenswrapper[4802]: I1206 03:47:14.697279 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d13a541e-f90d-4222-8486-2d47854ac8e5-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "d13a541e-f90d-4222-8486-2d47854ac8e5" (UID: "d13a541e-f90d-4222-8486-2d47854ac8e5"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:47:14 crc kubenswrapper[4802]: I1206 03:47:14.697524 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d13a541e-f90d-4222-8486-2d47854ac8e5-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "d13a541e-f90d-4222-8486-2d47854ac8e5" (UID: "d13a541e-f90d-4222-8486-2d47854ac8e5"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:47:14 crc kubenswrapper[4802]: I1206 03:47:14.701464 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "d13a541e-f90d-4222-8486-2d47854ac8e5" (UID: "d13a541e-f90d-4222-8486-2d47854ac8e5"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 06 03:47:14 crc kubenswrapper[4802]: I1206 03:47:14.703394 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d13a541e-f90d-4222-8486-2d47854ac8e5-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "d13a541e-f90d-4222-8486-2d47854ac8e5" (UID: "d13a541e-f90d-4222-8486-2d47854ac8e5"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:47:14 crc kubenswrapper[4802]: I1206 03:47:14.796204 4802 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d13a541e-f90d-4222-8486-2d47854ac8e5-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 06 03:47:14 crc kubenswrapper[4802]: I1206 03:47:14.796566 4802 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d13a541e-f90d-4222-8486-2d47854ac8e5-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 06 03:47:14 crc kubenswrapper[4802]: I1206 03:47:14.796694 4802 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d13a541e-f90d-4222-8486-2d47854ac8e5-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 06 03:47:14 crc kubenswrapper[4802]: I1206 03:47:14.900888 4802 generic.go:334] "Generic (PLEG): container finished" podID="d13a541e-f90d-4222-8486-2d47854ac8e5" containerID="ed30078e684092fe648afb023cfb75ac7a68756ac6e18783bdb323fe369bec1f" exitCode=0 Dec 06 03:47:14 crc kubenswrapper[4802]: I1206 03:47:14.900959 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" event={"ID":"d13a541e-f90d-4222-8486-2d47854ac8e5","Type":"ContainerDied","Data":"ed30078e684092fe648afb023cfb75ac7a68756ac6e18783bdb323fe369bec1f"} Dec 06 03:47:14 crc kubenswrapper[4802]: I1206 03:47:14.900999 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" event={"ID":"d13a541e-f90d-4222-8486-2d47854ac8e5","Type":"ContainerDied","Data":"b86b5a19d5941c6c13bce1346c809fb07700a39b5a1c7efabde6bc0a19a338ab"} Dec 06 03:47:14 crc kubenswrapper[4802]: I1206 03:47:14.901026 4802 scope.go:117] "RemoveContainer" containerID="ed30078e684092fe648afb023cfb75ac7a68756ac6e18783bdb323fe369bec1f" Dec 06 03:47:14 crc kubenswrapper[4802]: I1206 03:47:14.901259 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-blwnp" Dec 06 03:47:14 crc kubenswrapper[4802]: I1206 03:47:14.929998 4802 scope.go:117] "RemoveContainer" containerID="ed30078e684092fe648afb023cfb75ac7a68756ac6e18783bdb323fe369bec1f" Dec 06 03:47:14 crc kubenswrapper[4802]: E1206 03:47:14.930863 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed30078e684092fe648afb023cfb75ac7a68756ac6e18783bdb323fe369bec1f\": container with ID starting with ed30078e684092fe648afb023cfb75ac7a68756ac6e18783bdb323fe369bec1f not found: ID does not exist" containerID="ed30078e684092fe648afb023cfb75ac7a68756ac6e18783bdb323fe369bec1f" Dec 06 03:47:14 crc kubenswrapper[4802]: I1206 03:47:14.930942 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed30078e684092fe648afb023cfb75ac7a68756ac6e18783bdb323fe369bec1f"} err="failed to get container status \"ed30078e684092fe648afb023cfb75ac7a68756ac6e18783bdb323fe369bec1f\": rpc error: code = NotFound desc = could not find container \"ed30078e684092fe648afb023cfb75ac7a68756ac6e18783bdb323fe369bec1f\": container with ID starting with ed30078e684092fe648afb023cfb75ac7a68756ac6e18783bdb323fe369bec1f not found: ID does not exist" Dec 06 03:47:14 crc kubenswrapper[4802]: I1206 03:47:14.957305 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-blwnp"] Dec 06 03:47:14 crc kubenswrapper[4802]: I1206 03:47:14.976624 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-blwnp"] Dec 06 03:47:15 crc kubenswrapper[4802]: I1206 03:47:15.462225 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d13a541e-f90d-4222-8486-2d47854ac8e5" path="/var/lib/kubelet/pods/d13a541e-f90d-4222-8486-2d47854ac8e5/volumes" Dec 06 03:49:13 crc kubenswrapper[4802]: I1206 03:49:13.283275 4802 patch_prober.go:28] interesting pod/machine-config-daemon-zpxxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 03:49:13 crc kubenswrapper[4802]: I1206 03:49:13.284023 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 03:49:43 crc kubenswrapper[4802]: I1206 03:49:43.284027 4802 patch_prober.go:28] interesting pod/machine-config-daemon-zpxxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 03:49:43 crc kubenswrapper[4802]: I1206 03:49:43.284650 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 03:50:13 crc kubenswrapper[4802]: I1206 03:50:13.283855 4802 patch_prober.go:28] interesting pod/machine-config-daemon-zpxxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 03:50:13 crc kubenswrapper[4802]: I1206 03:50:13.285090 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 03:50:13 crc kubenswrapper[4802]: I1206 03:50:13.285179 4802 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" Dec 06 03:50:13 crc kubenswrapper[4802]: I1206 03:50:13.285784 4802 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"25f142fb563466ee1de9442e9c6cd18bf7c0126b2d7681419cdb7a7267f4a6e6"} pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 03:50:13 crc kubenswrapper[4802]: I1206 03:50:13.285831 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" containerID="cri-o://25f142fb563466ee1de9442e9c6cd18bf7c0126b2d7681419cdb7a7267f4a6e6" gracePeriod=600 Dec 06 03:50:14 crc kubenswrapper[4802]: I1206 03:50:14.109329 4802 generic.go:334] "Generic (PLEG): container finished" podID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerID="25f142fb563466ee1de9442e9c6cd18bf7c0126b2d7681419cdb7a7267f4a6e6" exitCode=0 Dec 06 03:50:14 crc kubenswrapper[4802]: I1206 03:50:14.109443 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" event={"ID":"6df38316-e0d3-4018-8d27-3620eba3a68d","Type":"ContainerDied","Data":"25f142fb563466ee1de9442e9c6cd18bf7c0126b2d7681419cdb7a7267f4a6e6"} Dec 06 03:50:14 crc kubenswrapper[4802]: I1206 03:50:14.110095 4802 scope.go:117] "RemoveContainer" containerID="2200194f7108affdd47802bf08946e25d3e937d094def5243f711369c5ae1305" Dec 06 03:50:14 crc kubenswrapper[4802]: I1206 03:50:14.109918 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" event={"ID":"6df38316-e0d3-4018-8d27-3620eba3a68d","Type":"ContainerStarted","Data":"57ad81d5a8f13aaa4dc17e3a8d71a6ae21755975dbb001e2d8346604e492cf3d"} Dec 06 03:50:21 crc kubenswrapper[4802]: I1206 03:50:21.348657 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-7r6gm"] Dec 06 03:50:21 crc kubenswrapper[4802]: E1206 03:50:21.349477 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d13a541e-f90d-4222-8486-2d47854ac8e5" containerName="registry" Dec 06 03:50:21 crc kubenswrapper[4802]: I1206 03:50:21.349495 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="d13a541e-f90d-4222-8486-2d47854ac8e5" containerName="registry" Dec 06 03:50:21 crc kubenswrapper[4802]: I1206 03:50:21.349639 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="d13a541e-f90d-4222-8486-2d47854ac8e5" containerName="registry" Dec 06 03:50:21 crc kubenswrapper[4802]: I1206 03:50:21.350085 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-7r6gm" Dec 06 03:50:21 crc kubenswrapper[4802]: I1206 03:50:21.351866 4802 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-9wcdt" Dec 06 03:50:21 crc kubenswrapper[4802]: I1206 03:50:21.352137 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 06 03:50:21 crc kubenswrapper[4802]: I1206 03:50:21.352281 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 06 03:50:21 crc kubenswrapper[4802]: I1206 03:50:21.357026 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-7r6gm"] Dec 06 03:50:21 crc kubenswrapper[4802]: I1206 03:50:21.368971 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-sdbpz"] Dec 06 03:50:21 crc kubenswrapper[4802]: I1206 03:50:21.370141 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-sdbpz" Dec 06 03:50:21 crc kubenswrapper[4802]: I1206 03:50:21.381080 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-54qqk"] Dec 06 03:50:21 crc kubenswrapper[4802]: I1206 03:50:21.382740 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-54qqk" Dec 06 03:50:21 crc kubenswrapper[4802]: I1206 03:50:21.386864 4802 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-x2pwk" Dec 06 03:50:21 crc kubenswrapper[4802]: I1206 03:50:21.387033 4802 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-qbt2d" Dec 06 03:50:21 crc kubenswrapper[4802]: I1206 03:50:21.401457 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-sdbpz"] Dec 06 03:50:21 crc kubenswrapper[4802]: I1206 03:50:21.404262 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-54qqk"] Dec 06 03:50:21 crc kubenswrapper[4802]: I1206 03:50:21.484482 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7ph8\" (UniqueName: \"kubernetes.io/projected/5aaa82f6-dd8a-4ca3-8e3d-b574617c240a-kube-api-access-f7ph8\") pod \"cert-manager-cainjector-7f985d654d-7r6gm\" (UID: \"5aaa82f6-dd8a-4ca3-8e3d-b574617c240a\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-7r6gm" Dec 06 03:50:21 crc kubenswrapper[4802]: I1206 03:50:21.484565 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzr6l\" (UniqueName: \"kubernetes.io/projected/3f3aa4aa-3196-4dd8-802f-73235899e452-kube-api-access-zzr6l\") pod \"cert-manager-5b446d88c5-sdbpz\" (UID: \"3f3aa4aa-3196-4dd8-802f-73235899e452\") " pod="cert-manager/cert-manager-5b446d88c5-sdbpz" Dec 06 03:50:21 crc kubenswrapper[4802]: I1206 03:50:21.484597 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tbbmj\" (UniqueName: \"kubernetes.io/projected/a50fe25c-7be9-42ea-a81b-e6ba817043e8-kube-api-access-tbbmj\") pod \"cert-manager-webhook-5655c58dd6-54qqk\" (UID: \"a50fe25c-7be9-42ea-a81b-e6ba817043e8\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-54qqk" Dec 06 03:50:21 crc kubenswrapper[4802]: I1206 03:50:21.585804 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7ph8\" (UniqueName: \"kubernetes.io/projected/5aaa82f6-dd8a-4ca3-8e3d-b574617c240a-kube-api-access-f7ph8\") pod \"cert-manager-cainjector-7f985d654d-7r6gm\" (UID: \"5aaa82f6-dd8a-4ca3-8e3d-b574617c240a\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-7r6gm" Dec 06 03:50:21 crc kubenswrapper[4802]: I1206 03:50:21.585937 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzr6l\" (UniqueName: \"kubernetes.io/projected/3f3aa4aa-3196-4dd8-802f-73235899e452-kube-api-access-zzr6l\") pod \"cert-manager-5b446d88c5-sdbpz\" (UID: \"3f3aa4aa-3196-4dd8-802f-73235899e452\") " pod="cert-manager/cert-manager-5b446d88c5-sdbpz" Dec 06 03:50:21 crc kubenswrapper[4802]: I1206 03:50:21.585990 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tbbmj\" (UniqueName: \"kubernetes.io/projected/a50fe25c-7be9-42ea-a81b-e6ba817043e8-kube-api-access-tbbmj\") pod \"cert-manager-webhook-5655c58dd6-54qqk\" (UID: \"a50fe25c-7be9-42ea-a81b-e6ba817043e8\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-54qqk" Dec 06 03:50:21 crc kubenswrapper[4802]: I1206 03:50:21.605447 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tbbmj\" (UniqueName: \"kubernetes.io/projected/a50fe25c-7be9-42ea-a81b-e6ba817043e8-kube-api-access-tbbmj\") pod \"cert-manager-webhook-5655c58dd6-54qqk\" (UID: \"a50fe25c-7be9-42ea-a81b-e6ba817043e8\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-54qqk" Dec 06 03:50:21 crc kubenswrapper[4802]: I1206 03:50:21.605769 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzr6l\" (UniqueName: \"kubernetes.io/projected/3f3aa4aa-3196-4dd8-802f-73235899e452-kube-api-access-zzr6l\") pod \"cert-manager-5b446d88c5-sdbpz\" (UID: \"3f3aa4aa-3196-4dd8-802f-73235899e452\") " pod="cert-manager/cert-manager-5b446d88c5-sdbpz" Dec 06 03:50:21 crc kubenswrapper[4802]: I1206 03:50:21.607673 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7ph8\" (UniqueName: \"kubernetes.io/projected/5aaa82f6-dd8a-4ca3-8e3d-b574617c240a-kube-api-access-f7ph8\") pod \"cert-manager-cainjector-7f985d654d-7r6gm\" (UID: \"5aaa82f6-dd8a-4ca3-8e3d-b574617c240a\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-7r6gm" Dec 06 03:50:21 crc kubenswrapper[4802]: I1206 03:50:21.677408 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-7r6gm" Dec 06 03:50:21 crc kubenswrapper[4802]: I1206 03:50:21.701980 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-sdbpz" Dec 06 03:50:21 crc kubenswrapper[4802]: I1206 03:50:21.713495 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-54qqk" Dec 06 03:50:22 crc kubenswrapper[4802]: I1206 03:50:22.004901 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-54qqk"] Dec 06 03:50:22 crc kubenswrapper[4802]: I1206 03:50:22.010773 4802 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 03:50:22 crc kubenswrapper[4802]: I1206 03:50:22.167472 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-54qqk" event={"ID":"a50fe25c-7be9-42ea-a81b-e6ba817043e8","Type":"ContainerStarted","Data":"fa712263d29c3704966cc80d92a92fa59d9006c926307011495e31888eef5c6c"} Dec 06 03:50:22 crc kubenswrapper[4802]: I1206 03:50:22.173047 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-7r6gm"] Dec 06 03:50:22 crc kubenswrapper[4802]: I1206 03:50:22.180334 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-sdbpz"] Dec 06 03:50:22 crc kubenswrapper[4802]: W1206 03:50:22.182512 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5aaa82f6_dd8a_4ca3_8e3d_b574617c240a.slice/crio-20a2e56b0983a1e1283c3163bf3860f3506215cb3400b9d4c2854bf91ce1b593 WatchSource:0}: Error finding container 20a2e56b0983a1e1283c3163bf3860f3506215cb3400b9d4c2854bf91ce1b593: Status 404 returned error can't find the container with id 20a2e56b0983a1e1283c3163bf3860f3506215cb3400b9d4c2854bf91ce1b593 Dec 06 03:50:22 crc kubenswrapper[4802]: W1206 03:50:22.189025 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3f3aa4aa_3196_4dd8_802f_73235899e452.slice/crio-8cb8ea74905c34921dd1b0b5f6385f34e47c678e488f757ae2023cf7e6d634fa WatchSource:0}: Error finding container 8cb8ea74905c34921dd1b0b5f6385f34e47c678e488f757ae2023cf7e6d634fa: Status 404 returned error can't find the container with id 8cb8ea74905c34921dd1b0b5f6385f34e47c678e488f757ae2023cf7e6d634fa Dec 06 03:50:23 crc kubenswrapper[4802]: I1206 03:50:23.174326 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-sdbpz" event={"ID":"3f3aa4aa-3196-4dd8-802f-73235899e452","Type":"ContainerStarted","Data":"8cb8ea74905c34921dd1b0b5f6385f34e47c678e488f757ae2023cf7e6d634fa"} Dec 06 03:50:23 crc kubenswrapper[4802]: I1206 03:50:23.176114 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-7r6gm" event={"ID":"5aaa82f6-dd8a-4ca3-8e3d-b574617c240a","Type":"ContainerStarted","Data":"20a2e56b0983a1e1283c3163bf3860f3506215cb3400b9d4c2854bf91ce1b593"} Dec 06 03:50:25 crc kubenswrapper[4802]: I1206 03:50:25.189719 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-7r6gm" event={"ID":"5aaa82f6-dd8a-4ca3-8e3d-b574617c240a","Type":"ContainerStarted","Data":"66e50023f499a0de7eb5abe2ed28a725b65fb73835584695b69b38b4732c2405"} Dec 06 03:50:25 crc kubenswrapper[4802]: I1206 03:50:25.193239 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-sdbpz" event={"ID":"3f3aa4aa-3196-4dd8-802f-73235899e452","Type":"ContainerStarted","Data":"bb6ee6ab43cc936378f96f2de70897f16546511e831fc141024ebd7394a817ca"} Dec 06 03:50:25 crc kubenswrapper[4802]: I1206 03:50:25.195513 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-54qqk" event={"ID":"a50fe25c-7be9-42ea-a81b-e6ba817043e8","Type":"ContainerStarted","Data":"2d9bf5d6d7cce6293ad4992cea06b126b43ef41b51ede3d6a170f82bdcf31198"} Dec 06 03:50:25 crc kubenswrapper[4802]: I1206 03:50:25.196106 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-54qqk" Dec 06 03:50:25 crc kubenswrapper[4802]: I1206 03:50:25.214472 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-7r6gm" podStartSLOduration=1.567202022 podStartE2EDuration="4.214456634s" podCreationTimestamp="2025-12-06 03:50:21 +0000 UTC" firstStartedPulling="2025-12-06 03:50:22.184245545 +0000 UTC m=+615.056154697" lastFinishedPulling="2025-12-06 03:50:24.831500147 +0000 UTC m=+617.703409309" observedRunningTime="2025-12-06 03:50:25.212554914 +0000 UTC m=+618.084464066" watchObservedRunningTime="2025-12-06 03:50:25.214456634 +0000 UTC m=+618.086365786" Dec 06 03:50:25 crc kubenswrapper[4802]: I1206 03:50:25.229925 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-sdbpz" podStartSLOduration=1.527167179 podStartE2EDuration="4.229907614s" podCreationTimestamp="2025-12-06 03:50:21 +0000 UTC" firstStartedPulling="2025-12-06 03:50:22.191262935 +0000 UTC m=+615.063172087" lastFinishedPulling="2025-12-06 03:50:24.89400337 +0000 UTC m=+617.765912522" observedRunningTime="2025-12-06 03:50:25.228383192 +0000 UTC m=+618.100292344" watchObservedRunningTime="2025-12-06 03:50:25.229907614 +0000 UTC m=+618.101816766" Dec 06 03:50:25 crc kubenswrapper[4802]: I1206 03:50:25.262561 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-54qqk" podStartSLOduration=1.443084139 podStartE2EDuration="4.262540687s" podCreationTimestamp="2025-12-06 03:50:21 +0000 UTC" firstStartedPulling="2025-12-06 03:50:22.010333422 +0000 UTC m=+614.882242584" lastFinishedPulling="2025-12-06 03:50:24.82978996 +0000 UTC m=+617.701699132" observedRunningTime="2025-12-06 03:50:25.257884842 +0000 UTC m=+618.129793994" watchObservedRunningTime="2025-12-06 03:50:25.262540687 +0000 UTC m=+618.134449839" Dec 06 03:50:30 crc kubenswrapper[4802]: I1206 03:50:30.986049 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-pvm6q"] Dec 06 03:50:30 crc kubenswrapper[4802]: I1206 03:50:30.987377 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" podUID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerName="ovn-controller" containerID="cri-o://5cf6796f65ee9c5bd0c18f205697135d17677c5daf3ba378cc992cba795e0188" gracePeriod=30 Dec 06 03:50:30 crc kubenswrapper[4802]: I1206 03:50:30.987440 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" podUID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerName="northd" containerID="cri-o://8941bd25107333b264d3f081a811fdfeea23ad92193fc61367665fd2cd1209c5" gracePeriod=30 Dec 06 03:50:30 crc kubenswrapper[4802]: I1206 03:50:30.987466 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" podUID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerName="sbdb" containerID="cri-o://b3cb210bfc97a439a7ff98c02c59f8ad16bc400dc214a616a44b43a8b01e2782" gracePeriod=30 Dec 06 03:50:30 crc kubenswrapper[4802]: I1206 03:50:30.987536 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" podUID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://a158aebd701ef46225d3e37109ae28ab58d107087bf06650c4b6d0055a3bbba5" gracePeriod=30 Dec 06 03:50:30 crc kubenswrapper[4802]: I1206 03:50:30.987601 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" podUID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerName="kube-rbac-proxy-node" containerID="cri-o://7730de4f033561ad77046e028cd1a3f0bfe2775387cbdfa35f47ea79909ec4cd" gracePeriod=30 Dec 06 03:50:30 crc kubenswrapper[4802]: I1206 03:50:30.987653 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" podUID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerName="nbdb" containerID="cri-o://0a2357f37497fde64937eca7b38693893af745ab278653ebc79406e122a6d4ff" gracePeriod=30 Dec 06 03:50:30 crc kubenswrapper[4802]: I1206 03:50:30.987663 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" podUID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerName="ovn-acl-logging" containerID="cri-o://38f9cd4dc75e63f09f60b48b6da284643910bb32ba51f2a808cb4d9803f30042" gracePeriod=30 Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.024570 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" podUID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerName="ovnkube-controller" containerID="cri-o://4180d357d73e8243e980d1bcaef03f37d0eb15e93c2e8a73d9c204909ff5ac54" gracePeriod=30 Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.248199 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-22rrq_03112a5d-5882-4fac-9ae5-13525ff82fe0/kube-multus/2.log" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.248601 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-22rrq_03112a5d-5882-4fac-9ae5-13525ff82fe0/kube-multus/1.log" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.248631 4802 generic.go:334] "Generic (PLEG): container finished" podID="03112a5d-5882-4fac-9ae5-13525ff82fe0" containerID="a85adeebc0758de20abbd8fcca0cea561a182955a07b43139ba2e29f72134676" exitCode=2 Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.248675 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-22rrq" event={"ID":"03112a5d-5882-4fac-9ae5-13525ff82fe0","Type":"ContainerDied","Data":"a85adeebc0758de20abbd8fcca0cea561a182955a07b43139ba2e29f72134676"} Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.248706 4802 scope.go:117] "RemoveContainer" containerID="831dab65557627cbce0d055459052dae5b594f33b66858367ccd034ff6811924" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.249108 4802 scope.go:117] "RemoveContainer" containerID="a85adeebc0758de20abbd8fcca0cea561a182955a07b43139ba2e29f72134676" Dec 06 03:50:31 crc kubenswrapper[4802]: E1206 03:50:31.249295 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-22rrq_openshift-multus(03112a5d-5882-4fac-9ae5-13525ff82fe0)\"" pod="openshift-multus/multus-22rrq" podUID="03112a5d-5882-4fac-9ae5-13525ff82fe0" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.257646 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pvm6q_eafce4e0-e7fb-4877-b0ab-3283829ba1ef/ovnkube-controller/3.log" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.259801 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pvm6q_eafce4e0-e7fb-4877-b0ab-3283829ba1ef/ovn-acl-logging/0.log" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.260221 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pvm6q_eafce4e0-e7fb-4877-b0ab-3283829ba1ef/ovn-controller/0.log" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.260618 4802 generic.go:334] "Generic (PLEG): container finished" podID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerID="4180d357d73e8243e980d1bcaef03f37d0eb15e93c2e8a73d9c204909ff5ac54" exitCode=0 Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.260642 4802 generic.go:334] "Generic (PLEG): container finished" podID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerID="b3cb210bfc97a439a7ff98c02c59f8ad16bc400dc214a616a44b43a8b01e2782" exitCode=0 Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.260651 4802 generic.go:334] "Generic (PLEG): container finished" podID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerID="0a2357f37497fde64937eca7b38693893af745ab278653ebc79406e122a6d4ff" exitCode=0 Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.260658 4802 generic.go:334] "Generic (PLEG): container finished" podID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerID="8941bd25107333b264d3f081a811fdfeea23ad92193fc61367665fd2cd1209c5" exitCode=0 Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.260664 4802 generic.go:334] "Generic (PLEG): container finished" podID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerID="a158aebd701ef46225d3e37109ae28ab58d107087bf06650c4b6d0055a3bbba5" exitCode=0 Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.260671 4802 generic.go:334] "Generic (PLEG): container finished" podID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerID="7730de4f033561ad77046e028cd1a3f0bfe2775387cbdfa35f47ea79909ec4cd" exitCode=0 Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.260679 4802 generic.go:334] "Generic (PLEG): container finished" podID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerID="38f9cd4dc75e63f09f60b48b6da284643910bb32ba51f2a808cb4d9803f30042" exitCode=143 Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.260687 4802 generic.go:334] "Generic (PLEG): container finished" podID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerID="5cf6796f65ee9c5bd0c18f205697135d17677c5daf3ba378cc992cba795e0188" exitCode=143 Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.260685 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" event={"ID":"eafce4e0-e7fb-4877-b0ab-3283829ba1ef","Type":"ContainerDied","Data":"4180d357d73e8243e980d1bcaef03f37d0eb15e93c2e8a73d9c204909ff5ac54"} Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.260730 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" event={"ID":"eafce4e0-e7fb-4877-b0ab-3283829ba1ef","Type":"ContainerDied","Data":"b3cb210bfc97a439a7ff98c02c59f8ad16bc400dc214a616a44b43a8b01e2782"} Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.260764 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" event={"ID":"eafce4e0-e7fb-4877-b0ab-3283829ba1ef","Type":"ContainerDied","Data":"0a2357f37497fde64937eca7b38693893af745ab278653ebc79406e122a6d4ff"} Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.260780 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" event={"ID":"eafce4e0-e7fb-4877-b0ab-3283829ba1ef","Type":"ContainerDied","Data":"8941bd25107333b264d3f081a811fdfeea23ad92193fc61367665fd2cd1209c5"} Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.260792 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" event={"ID":"eafce4e0-e7fb-4877-b0ab-3283829ba1ef","Type":"ContainerDied","Data":"a158aebd701ef46225d3e37109ae28ab58d107087bf06650c4b6d0055a3bbba5"} Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.260803 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" event={"ID":"eafce4e0-e7fb-4877-b0ab-3283829ba1ef","Type":"ContainerDied","Data":"7730de4f033561ad77046e028cd1a3f0bfe2775387cbdfa35f47ea79909ec4cd"} Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.260814 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" event={"ID":"eafce4e0-e7fb-4877-b0ab-3283829ba1ef","Type":"ContainerDied","Data":"38f9cd4dc75e63f09f60b48b6da284643910bb32ba51f2a808cb4d9803f30042"} Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.260825 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" event={"ID":"eafce4e0-e7fb-4877-b0ab-3283829ba1ef","Type":"ContainerDied","Data":"5cf6796f65ee9c5bd0c18f205697135d17677c5daf3ba378cc992cba795e0188"} Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.304831 4802 scope.go:117] "RemoveContainer" containerID="bf554aaba0e773f6a6c1f29bb1819b7ae25b2b2db255c4bd0a16c512f314e9cf" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.308832 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pvm6q_eafce4e0-e7fb-4877-b0ab-3283829ba1ef/ovn-acl-logging/0.log" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.309256 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pvm6q_eafce4e0-e7fb-4877-b0ab-3283829ba1ef/ovn-controller/0.log" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.309569 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.374474 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-7p9zw"] Dec 06 03:50:31 crc kubenswrapper[4802]: E1206 03:50:31.374716 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerName="ovnkube-controller" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.374729 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerName="ovnkube-controller" Dec 06 03:50:31 crc kubenswrapper[4802]: E1206 03:50:31.374741 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerName="ovnkube-controller" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.374772 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerName="ovnkube-controller" Dec 06 03:50:31 crc kubenswrapper[4802]: E1206 03:50:31.374788 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerName="sbdb" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.374798 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerName="sbdb" Dec 06 03:50:31 crc kubenswrapper[4802]: E1206 03:50:31.374815 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerName="kubecfg-setup" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.374825 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerName="kubecfg-setup" Dec 06 03:50:31 crc kubenswrapper[4802]: E1206 03:50:31.374841 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerName="ovnkube-controller" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.374850 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerName="ovnkube-controller" Dec 06 03:50:31 crc kubenswrapper[4802]: E1206 03:50:31.374860 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerName="ovnkube-controller" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.374870 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerName="ovnkube-controller" Dec 06 03:50:31 crc kubenswrapper[4802]: E1206 03:50:31.374885 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerName="nbdb" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.374895 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerName="nbdb" Dec 06 03:50:31 crc kubenswrapper[4802]: E1206 03:50:31.374906 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerName="ovn-acl-logging" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.374914 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerName="ovn-acl-logging" Dec 06 03:50:31 crc kubenswrapper[4802]: E1206 03:50:31.374928 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerName="northd" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.374936 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerName="northd" Dec 06 03:50:31 crc kubenswrapper[4802]: E1206 03:50:31.374947 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerName="ovn-controller" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.374955 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerName="ovn-controller" Dec 06 03:50:31 crc kubenswrapper[4802]: E1206 03:50:31.374972 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerName="kube-rbac-proxy-node" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.374980 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerName="kube-rbac-proxy-node" Dec 06 03:50:31 crc kubenswrapper[4802]: E1206 03:50:31.374990 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerName="kube-rbac-proxy-ovn-metrics" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.374998 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerName="kube-rbac-proxy-ovn-metrics" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.375130 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerName="ovnkube-controller" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.375141 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerName="ovn-controller" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.375152 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerName="nbdb" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.375164 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerName="ovnkube-controller" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.375172 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerName="ovn-acl-logging" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.375185 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerName="sbdb" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.375194 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerName="northd" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.375205 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerName="ovnkube-controller" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.375214 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerName="kube-rbac-proxy-node" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.375224 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerName="kube-rbac-proxy-ovn-metrics" Dec 06 03:50:31 crc kubenswrapper[4802]: E1206 03:50:31.375357 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerName="ovnkube-controller" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.375367 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerName="ovnkube-controller" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.375471 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerName="ovnkube-controller" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.375486 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" containerName="ovnkube-controller" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.377621 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.434060 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-env-overrides\") pod \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.434444 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-var-lib-openvswitch\") pod \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.434472 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-host-run-ovn-kubernetes\") pod \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.434494 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-node-log\") pod \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.434525 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-run-openvswitch\") pod \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.434523 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "eafce4e0-e7fb-4877-b0ab-3283829ba1ef" (UID: "eafce4e0-e7fb-4877-b0ab-3283829ba1ef"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.434562 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-log-socket\") pod \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.434572 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-node-log" (OuterVolumeSpecName: "node-log") pod "eafce4e0-e7fb-4877-b0ab-3283829ba1ef" (UID: "eafce4e0-e7fb-4877-b0ab-3283829ba1ef"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.434587 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-run-systemd\") pod \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.434599 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-log-socket" (OuterVolumeSpecName: "log-socket") pod "eafce4e0-e7fb-4877-b0ab-3283829ba1ef" (UID: "eafce4e0-e7fb-4877-b0ab-3283829ba1ef"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.434617 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-ovn-node-metrics-cert\") pod \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.434663 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-host-cni-bin\") pod \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.434612 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "eafce4e0-e7fb-4877-b0ab-3283829ba1ef" (UID: "eafce4e0-e7fb-4877-b0ab-3283829ba1ef"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.434658 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "eafce4e0-e7fb-4877-b0ab-3283829ba1ef" (UID: "eafce4e0-e7fb-4877-b0ab-3283829ba1ef"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.434696 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-ovnkube-config\") pod \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.434795 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-etc-openvswitch\") pod \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.434849 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-host-run-netns\") pod \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.434874 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-host-var-lib-cni-networks-ovn-kubernetes\") pod \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.434903 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-ovnkube-script-lib\") pod \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.435128 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-62rc4\" (UniqueName: \"kubernetes.io/projected/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-kube-api-access-62rc4\") pod \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.435108 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "eafce4e0-e7fb-4877-b0ab-3283829ba1ef" (UID: "eafce4e0-e7fb-4877-b0ab-3283829ba1ef"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.435237 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "eafce4e0-e7fb-4877-b0ab-3283829ba1ef" (UID: "eafce4e0-e7fb-4877-b0ab-3283829ba1ef"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.435274 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "eafce4e0-e7fb-4877-b0ab-3283829ba1ef" (UID: "eafce4e0-e7fb-4877-b0ab-3283829ba1ef"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.435299 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "eafce4e0-e7fb-4877-b0ab-3283829ba1ef" (UID: "eafce4e0-e7fb-4877-b0ab-3283829ba1ef"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.435313 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "eafce4e0-e7fb-4877-b0ab-3283829ba1ef" (UID: "eafce4e0-e7fb-4877-b0ab-3283829ba1ef"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.435324 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "eafce4e0-e7fb-4877-b0ab-3283829ba1ef" (UID: "eafce4e0-e7fb-4877-b0ab-3283829ba1ef"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.435330 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "eafce4e0-e7fb-4877-b0ab-3283829ba1ef" (UID: "eafce4e0-e7fb-4877-b0ab-3283829ba1ef"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.435399 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-host-kubelet\") pod \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.435419 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "eafce4e0-e7fb-4877-b0ab-3283829ba1ef" (UID: "eafce4e0-e7fb-4877-b0ab-3283829ba1ef"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.435563 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-systemd-units\") pod \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.435607 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-run-ovn\") pod \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.435622 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-host-slash\") pod \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.435645 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-host-cni-netd\") pod \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\" (UID: \"eafce4e0-e7fb-4877-b0ab-3283829ba1ef\") " Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.435685 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "eafce4e0-e7fb-4877-b0ab-3283829ba1ef" (UID: "eafce4e0-e7fb-4877-b0ab-3283829ba1ef"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.435745 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "eafce4e0-e7fb-4877-b0ab-3283829ba1ef" (UID: "eafce4e0-e7fb-4877-b0ab-3283829ba1ef"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.435804 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-host-slash" (OuterVolumeSpecName: "host-slash") pod "eafce4e0-e7fb-4877-b0ab-3283829ba1ef" (UID: "eafce4e0-e7fb-4877-b0ab-3283829ba1ef"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.435853 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-env-overrides\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.435876 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-host-cni-netd\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.435874 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "eafce4e0-e7fb-4877-b0ab-3283829ba1ef" (UID: "eafce4e0-e7fb-4877-b0ab-3283829ba1ef"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.435896 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-systemd-units\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.435917 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-log-socket\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.435939 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-etc-openvswitch\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.435957 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-host-run-ovn-kubernetes\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.436097 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-host-kubelet\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.436155 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-ovn-node-metrics-cert\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.436212 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-run-ovn\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.436254 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.436464 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-var-lib-openvswitch\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.436492 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-ovnkube-config\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.436524 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-host-run-netns\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.436555 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-run-systemd\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.436655 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-host-cni-bin\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.436696 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-ovnkube-script-lib\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.436742 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-host-slash\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.436784 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-node-log\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.436808 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-run-openvswitch\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.436833 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bhgbx\" (UniqueName: \"kubernetes.io/projected/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-kube-api-access-bhgbx\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.437051 4802 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.437222 4802 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.437252 4802 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.437272 4802 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.437283 4802 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.437294 4802 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.437304 4802 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.437312 4802 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-host-slash\") on node \"crc\" DevicePath \"\"" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.437322 4802 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.437334 4802 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.437347 4802 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.437361 4802 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.437372 4802 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-node-log\") on node \"crc\" DevicePath \"\"" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.437383 4802 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.437393 4802 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-log-socket\") on node \"crc\" DevicePath \"\"" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.437402 4802 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.437411 4802 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.441054 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-kube-api-access-62rc4" (OuterVolumeSpecName: "kube-api-access-62rc4") pod "eafce4e0-e7fb-4877-b0ab-3283829ba1ef" (UID: "eafce4e0-e7fb-4877-b0ab-3283829ba1ef"). InnerVolumeSpecName "kube-api-access-62rc4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.441304 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "eafce4e0-e7fb-4877-b0ab-3283829ba1ef" (UID: "eafce4e0-e7fb-4877-b0ab-3283829ba1ef"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.451358 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "eafce4e0-e7fb-4877-b0ab-3283829ba1ef" (UID: "eafce4e0-e7fb-4877-b0ab-3283829ba1ef"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.538863 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bhgbx\" (UniqueName: \"kubernetes.io/projected/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-kube-api-access-bhgbx\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.538968 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-env-overrides\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.539024 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-host-cni-netd\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.539057 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-log-socket\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.539090 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-systemd-units\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.539119 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-etc-openvswitch\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.539179 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-host-run-ovn-kubernetes\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.539236 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-host-kubelet\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.539272 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-ovn-node-metrics-cert\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.539340 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-run-ovn\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.539377 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.539413 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-var-lib-openvswitch\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.539445 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-ovnkube-config\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.539476 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-run-systemd\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.539497 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-host-kubelet\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.539582 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-host-run-netns\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.539668 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-systemd-units\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.539824 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-etc-openvswitch\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.540001 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-host-run-ovn-kubernetes\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.539512 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-host-run-netns\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.540101 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-host-cni-netd\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.540041 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-log-socket\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.540169 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-host-cni-bin\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.540220 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.540238 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-ovnkube-script-lib\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.540329 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-host-slash\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.540388 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-node-log\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.540442 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-run-openvswitch\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.540597 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-host-cni-bin\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.540607 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-host-slash\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.540635 4802 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.540668 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-env-overrides\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.540681 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-node-log\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.540732 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-run-ovn\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.540745 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-run-openvswitch\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.541614 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-var-lib-openvswitch\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.541818 4802 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.541860 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-run-systemd\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.541891 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-62rc4\" (UniqueName: \"kubernetes.io/projected/eafce4e0-e7fb-4877-b0ab-3283829ba1ef-kube-api-access-62rc4\") on node \"crc\" DevicePath \"\"" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.542045 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-ovnkube-script-lib\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.542152 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-ovnkube-config\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.543998 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-ovn-node-metrics-cert\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.571876 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bhgbx\" (UniqueName: \"kubernetes.io/projected/c33979fa-7c1d-4536-8efc-39e0d3ce1b12-kube-api-access-bhgbx\") pod \"ovnkube-node-7p9zw\" (UID: \"c33979fa-7c1d-4536-8efc-39e0d3ce1b12\") " pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.704867 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:31 crc kubenswrapper[4802]: I1206 03:50:31.716686 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-54qqk" Dec 06 03:50:32 crc kubenswrapper[4802]: I1206 03:50:32.269653 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-22rrq_03112a5d-5882-4fac-9ae5-13525ff82fe0/kube-multus/2.log" Dec 06 03:50:32 crc kubenswrapper[4802]: I1206 03:50:32.271939 4802 generic.go:334] "Generic (PLEG): container finished" podID="c33979fa-7c1d-4536-8efc-39e0d3ce1b12" containerID="5140be1cf11f2215000b6487da590473cf9f022eae269ff8afd5a13c0503b674" exitCode=0 Dec 06 03:50:32 crc kubenswrapper[4802]: I1206 03:50:32.271985 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" event={"ID":"c33979fa-7c1d-4536-8efc-39e0d3ce1b12","Type":"ContainerDied","Data":"5140be1cf11f2215000b6487da590473cf9f022eae269ff8afd5a13c0503b674"} Dec 06 03:50:32 crc kubenswrapper[4802]: I1206 03:50:32.272053 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" event={"ID":"c33979fa-7c1d-4536-8efc-39e0d3ce1b12","Type":"ContainerStarted","Data":"70eb658a6ca3f38590d1f1c91df47bb7676a16814e00629c04400a0f54b3114b"} Dec 06 03:50:32 crc kubenswrapper[4802]: I1206 03:50:32.277874 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pvm6q_eafce4e0-e7fb-4877-b0ab-3283829ba1ef/ovn-acl-logging/0.log" Dec 06 03:50:32 crc kubenswrapper[4802]: I1206 03:50:32.279347 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pvm6q_eafce4e0-e7fb-4877-b0ab-3283829ba1ef/ovn-controller/0.log" Dec 06 03:50:32 crc kubenswrapper[4802]: I1206 03:50:32.280450 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" event={"ID":"eafce4e0-e7fb-4877-b0ab-3283829ba1ef","Type":"ContainerDied","Data":"5d23efa094b6dc407ad7c00fe428e299a5f32feddee8d4369230e195dbcaeb2f"} Dec 06 03:50:32 crc kubenswrapper[4802]: I1206 03:50:32.280516 4802 scope.go:117] "RemoveContainer" containerID="4180d357d73e8243e980d1bcaef03f37d0eb15e93c2e8a73d9c204909ff5ac54" Dec 06 03:50:32 crc kubenswrapper[4802]: I1206 03:50:32.280583 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-pvm6q" Dec 06 03:50:32 crc kubenswrapper[4802]: I1206 03:50:32.320770 4802 scope.go:117] "RemoveContainer" containerID="b3cb210bfc97a439a7ff98c02c59f8ad16bc400dc214a616a44b43a8b01e2782" Dec 06 03:50:32 crc kubenswrapper[4802]: I1206 03:50:32.375803 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-pvm6q"] Dec 06 03:50:32 crc kubenswrapper[4802]: I1206 03:50:32.376159 4802 scope.go:117] "RemoveContainer" containerID="0a2357f37497fde64937eca7b38693893af745ab278653ebc79406e122a6d4ff" Dec 06 03:50:32 crc kubenswrapper[4802]: I1206 03:50:32.380445 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-pvm6q"] Dec 06 03:50:32 crc kubenswrapper[4802]: I1206 03:50:32.394821 4802 scope.go:117] "RemoveContainer" containerID="8941bd25107333b264d3f081a811fdfeea23ad92193fc61367665fd2cd1209c5" Dec 06 03:50:32 crc kubenswrapper[4802]: I1206 03:50:32.419505 4802 scope.go:117] "RemoveContainer" containerID="a158aebd701ef46225d3e37109ae28ab58d107087bf06650c4b6d0055a3bbba5" Dec 06 03:50:32 crc kubenswrapper[4802]: I1206 03:50:32.435504 4802 scope.go:117] "RemoveContainer" containerID="7730de4f033561ad77046e028cd1a3f0bfe2775387cbdfa35f47ea79909ec4cd" Dec 06 03:50:32 crc kubenswrapper[4802]: I1206 03:50:32.447602 4802 scope.go:117] "RemoveContainer" containerID="38f9cd4dc75e63f09f60b48b6da284643910bb32ba51f2a808cb4d9803f30042" Dec 06 03:50:32 crc kubenswrapper[4802]: I1206 03:50:32.461518 4802 scope.go:117] "RemoveContainer" containerID="5cf6796f65ee9c5bd0c18f205697135d17677c5daf3ba378cc992cba795e0188" Dec 06 03:50:32 crc kubenswrapper[4802]: I1206 03:50:32.485215 4802 scope.go:117] "RemoveContainer" containerID="a863962985012c4bd59372460ad9f1b47130bfd62897a1ecef46a2506f8f01e7" Dec 06 03:50:33 crc kubenswrapper[4802]: I1206 03:50:33.292580 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" event={"ID":"c33979fa-7c1d-4536-8efc-39e0d3ce1b12","Type":"ContainerStarted","Data":"e9b705070774d62823334551364d2e35342004b6671c0d206898fe483ff4a535"} Dec 06 03:50:33 crc kubenswrapper[4802]: I1206 03:50:33.292922 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" event={"ID":"c33979fa-7c1d-4536-8efc-39e0d3ce1b12","Type":"ContainerStarted","Data":"3a84517de1a9b5a65ddd1c488e0673cda7c9443ebc52d44b6a07e906638b8017"} Dec 06 03:50:33 crc kubenswrapper[4802]: I1206 03:50:33.292958 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" event={"ID":"c33979fa-7c1d-4536-8efc-39e0d3ce1b12","Type":"ContainerStarted","Data":"d066c001be11c06a375327242dca0211808edc9dc8b6c0979889c7835921d0af"} Dec 06 03:50:33 crc kubenswrapper[4802]: I1206 03:50:33.292971 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" event={"ID":"c33979fa-7c1d-4536-8efc-39e0d3ce1b12","Type":"ContainerStarted","Data":"e96350658c5be8a82fa04df5a0b46d575e852bf72ef9ff6dce3a14eb07e7fc8c"} Dec 06 03:50:33 crc kubenswrapper[4802]: I1206 03:50:33.292983 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" event={"ID":"c33979fa-7c1d-4536-8efc-39e0d3ce1b12","Type":"ContainerStarted","Data":"208b0d5c3de42cf56ef294a2774a653ba7545c170e31d5085815f6228d087bd0"} Dec 06 03:50:33 crc kubenswrapper[4802]: I1206 03:50:33.292994 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" event={"ID":"c33979fa-7c1d-4536-8efc-39e0d3ce1b12","Type":"ContainerStarted","Data":"4b92519a4add6f88d3109ab4392d74a84633a38287e5a6daf1d09713e58a949a"} Dec 06 03:50:33 crc kubenswrapper[4802]: I1206 03:50:33.463496 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eafce4e0-e7fb-4877-b0ab-3283829ba1ef" path="/var/lib/kubelet/pods/eafce4e0-e7fb-4877-b0ab-3283829ba1ef/volumes" Dec 06 03:50:36 crc kubenswrapper[4802]: I1206 03:50:36.321230 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" event={"ID":"c33979fa-7c1d-4536-8efc-39e0d3ce1b12","Type":"ContainerStarted","Data":"6ce81e5df8e291e9b7405e5be876542e12073760379efbab0bc45a105946b296"} Dec 06 03:50:38 crc kubenswrapper[4802]: I1206 03:50:38.345826 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" event={"ID":"c33979fa-7c1d-4536-8efc-39e0d3ce1b12","Type":"ContainerStarted","Data":"63b1a4f9ce880280ae0a1b91cc2925a2509a85350c18359ecfa74b267ecc1dfe"} Dec 06 03:50:38 crc kubenswrapper[4802]: I1206 03:50:38.346205 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:38 crc kubenswrapper[4802]: I1206 03:50:38.346229 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:38 crc kubenswrapper[4802]: I1206 03:50:38.377340 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:38 crc kubenswrapper[4802]: I1206 03:50:38.386141 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" podStartSLOduration=7.386124695 podStartE2EDuration="7.386124695s" podCreationTimestamp="2025-12-06 03:50:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:50:38.385015375 +0000 UTC m=+631.256924537" watchObservedRunningTime="2025-12-06 03:50:38.386124695 +0000 UTC m=+631.258033847" Dec 06 03:50:39 crc kubenswrapper[4802]: I1206 03:50:39.353072 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:39 crc kubenswrapper[4802]: I1206 03:50:39.392984 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:50:44 crc kubenswrapper[4802]: I1206 03:50:44.450892 4802 scope.go:117] "RemoveContainer" containerID="a85adeebc0758de20abbd8fcca0cea561a182955a07b43139ba2e29f72134676" Dec 06 03:50:44 crc kubenswrapper[4802]: E1206 03:50:44.451618 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-22rrq_openshift-multus(03112a5d-5882-4fac-9ae5-13525ff82fe0)\"" pod="openshift-multus/multus-22rrq" podUID="03112a5d-5882-4fac-9ae5-13525ff82fe0" Dec 06 03:50:59 crc kubenswrapper[4802]: I1206 03:50:59.449943 4802 scope.go:117] "RemoveContainer" containerID="a85adeebc0758de20abbd8fcca0cea561a182955a07b43139ba2e29f72134676" Dec 06 03:51:00 crc kubenswrapper[4802]: I1206 03:51:00.490848 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-22rrq_03112a5d-5882-4fac-9ae5-13525ff82fe0/kube-multus/2.log" Dec 06 03:51:00 crc kubenswrapper[4802]: I1206 03:51:00.491250 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-22rrq" event={"ID":"03112a5d-5882-4fac-9ae5-13525ff82fe0","Type":"ContainerStarted","Data":"857fdc9612ef55880ec5aceb8b26c1660a19edd753f1876b2784aad9d99ffe3e"} Dec 06 03:51:01 crc kubenswrapper[4802]: I1206 03:51:01.736234 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-7p9zw" Dec 06 03:51:13 crc kubenswrapper[4802]: I1206 03:51:13.186092 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkz86h"] Dec 06 03:51:13 crc kubenswrapper[4802]: I1206 03:51:13.188936 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkz86h" Dec 06 03:51:13 crc kubenswrapper[4802]: I1206 03:51:13.191946 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 06 03:51:13 crc kubenswrapper[4802]: I1206 03:51:13.205955 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkz86h"] Dec 06 03:51:13 crc kubenswrapper[4802]: I1206 03:51:13.274168 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7dfbb06c-f867-4d00-aef7-e731fa8579a7-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkz86h\" (UID: \"7dfbb06c-f867-4d00-aef7-e731fa8579a7\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkz86h" Dec 06 03:51:13 crc kubenswrapper[4802]: I1206 03:51:13.274644 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fc2zp\" (UniqueName: \"kubernetes.io/projected/7dfbb06c-f867-4d00-aef7-e731fa8579a7-kube-api-access-fc2zp\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkz86h\" (UID: \"7dfbb06c-f867-4d00-aef7-e731fa8579a7\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkz86h" Dec 06 03:51:13 crc kubenswrapper[4802]: I1206 03:51:13.274683 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7dfbb06c-f867-4d00-aef7-e731fa8579a7-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkz86h\" (UID: \"7dfbb06c-f867-4d00-aef7-e731fa8579a7\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkz86h" Dec 06 03:51:13 crc kubenswrapper[4802]: I1206 03:51:13.376368 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fc2zp\" (UniqueName: \"kubernetes.io/projected/7dfbb06c-f867-4d00-aef7-e731fa8579a7-kube-api-access-fc2zp\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkz86h\" (UID: \"7dfbb06c-f867-4d00-aef7-e731fa8579a7\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkz86h" Dec 06 03:51:13 crc kubenswrapper[4802]: I1206 03:51:13.376443 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7dfbb06c-f867-4d00-aef7-e731fa8579a7-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkz86h\" (UID: \"7dfbb06c-f867-4d00-aef7-e731fa8579a7\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkz86h" Dec 06 03:51:13 crc kubenswrapper[4802]: I1206 03:51:13.376495 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7dfbb06c-f867-4d00-aef7-e731fa8579a7-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkz86h\" (UID: \"7dfbb06c-f867-4d00-aef7-e731fa8579a7\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkz86h" Dec 06 03:51:13 crc kubenswrapper[4802]: I1206 03:51:13.377378 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7dfbb06c-f867-4d00-aef7-e731fa8579a7-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkz86h\" (UID: \"7dfbb06c-f867-4d00-aef7-e731fa8579a7\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkz86h" Dec 06 03:51:13 crc kubenswrapper[4802]: I1206 03:51:13.378117 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7dfbb06c-f867-4d00-aef7-e731fa8579a7-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkz86h\" (UID: \"7dfbb06c-f867-4d00-aef7-e731fa8579a7\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkz86h" Dec 06 03:51:13 crc kubenswrapper[4802]: I1206 03:51:13.401241 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fc2zp\" (UniqueName: \"kubernetes.io/projected/7dfbb06c-f867-4d00-aef7-e731fa8579a7-kube-api-access-fc2zp\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkz86h\" (UID: \"7dfbb06c-f867-4d00-aef7-e731fa8579a7\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkz86h" Dec 06 03:51:13 crc kubenswrapper[4802]: I1206 03:51:13.518493 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkz86h" Dec 06 03:51:13 crc kubenswrapper[4802]: I1206 03:51:13.771261 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkz86h"] Dec 06 03:51:13 crc kubenswrapper[4802]: W1206 03:51:13.785456 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7dfbb06c_f867_4d00_aef7_e731fa8579a7.slice/crio-3a6c9029d29b5405e78620a921a91605ee99892674d94926598796aeae06ea9e WatchSource:0}: Error finding container 3a6c9029d29b5405e78620a921a91605ee99892674d94926598796aeae06ea9e: Status 404 returned error can't find the container with id 3a6c9029d29b5405e78620a921a91605ee99892674d94926598796aeae06ea9e Dec 06 03:51:14 crc kubenswrapper[4802]: I1206 03:51:14.592152 4802 generic.go:334] "Generic (PLEG): container finished" podID="7dfbb06c-f867-4d00-aef7-e731fa8579a7" containerID="6d45f06314b4c306f4d3439c56225ec3f95bdb3d7e8d487b7edb028251c8834c" exitCode=0 Dec 06 03:51:14 crc kubenswrapper[4802]: I1206 03:51:14.592203 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkz86h" event={"ID":"7dfbb06c-f867-4d00-aef7-e731fa8579a7","Type":"ContainerDied","Data":"6d45f06314b4c306f4d3439c56225ec3f95bdb3d7e8d487b7edb028251c8834c"} Dec 06 03:51:14 crc kubenswrapper[4802]: I1206 03:51:14.592232 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkz86h" event={"ID":"7dfbb06c-f867-4d00-aef7-e731fa8579a7","Type":"ContainerStarted","Data":"3a6c9029d29b5405e78620a921a91605ee99892674d94926598796aeae06ea9e"} Dec 06 03:51:22 crc kubenswrapper[4802]: I1206 03:51:22.646358 4802 generic.go:334] "Generic (PLEG): container finished" podID="7dfbb06c-f867-4d00-aef7-e731fa8579a7" containerID="a31b597536fca514a586cbaf2cf121918b3e2114fed04183ef762409b91497ea" exitCode=0 Dec 06 03:51:22 crc kubenswrapper[4802]: I1206 03:51:22.646469 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkz86h" event={"ID":"7dfbb06c-f867-4d00-aef7-e731fa8579a7","Type":"ContainerDied","Data":"a31b597536fca514a586cbaf2cf121918b3e2114fed04183ef762409b91497ea"} Dec 06 03:51:23 crc kubenswrapper[4802]: I1206 03:51:23.653640 4802 generic.go:334] "Generic (PLEG): container finished" podID="7dfbb06c-f867-4d00-aef7-e731fa8579a7" containerID="ab10729fad01c3f8dd986b90ca8e4c722713c2f8a9a90af20266f67dd43f5eba" exitCode=0 Dec 06 03:51:23 crc kubenswrapper[4802]: I1206 03:51:23.653703 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkz86h" event={"ID":"7dfbb06c-f867-4d00-aef7-e731fa8579a7","Type":"ContainerDied","Data":"ab10729fad01c3f8dd986b90ca8e4c722713c2f8a9a90af20266f67dd43f5eba"} Dec 06 03:51:24 crc kubenswrapper[4802]: I1206 03:51:24.982637 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkz86h" Dec 06 03:51:25 crc kubenswrapper[4802]: I1206 03:51:25.036921 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7dfbb06c-f867-4d00-aef7-e731fa8579a7-bundle\") pod \"7dfbb06c-f867-4d00-aef7-e731fa8579a7\" (UID: \"7dfbb06c-f867-4d00-aef7-e731fa8579a7\") " Dec 06 03:51:25 crc kubenswrapper[4802]: I1206 03:51:25.037125 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fc2zp\" (UniqueName: \"kubernetes.io/projected/7dfbb06c-f867-4d00-aef7-e731fa8579a7-kube-api-access-fc2zp\") pod \"7dfbb06c-f867-4d00-aef7-e731fa8579a7\" (UID: \"7dfbb06c-f867-4d00-aef7-e731fa8579a7\") " Dec 06 03:51:25 crc kubenswrapper[4802]: I1206 03:51:25.037207 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7dfbb06c-f867-4d00-aef7-e731fa8579a7-util\") pod \"7dfbb06c-f867-4d00-aef7-e731fa8579a7\" (UID: \"7dfbb06c-f867-4d00-aef7-e731fa8579a7\") " Dec 06 03:51:25 crc kubenswrapper[4802]: I1206 03:51:25.038497 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7dfbb06c-f867-4d00-aef7-e731fa8579a7-bundle" (OuterVolumeSpecName: "bundle") pod "7dfbb06c-f867-4d00-aef7-e731fa8579a7" (UID: "7dfbb06c-f867-4d00-aef7-e731fa8579a7"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:51:25 crc kubenswrapper[4802]: I1206 03:51:25.048047 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7dfbb06c-f867-4d00-aef7-e731fa8579a7-kube-api-access-fc2zp" (OuterVolumeSpecName: "kube-api-access-fc2zp") pod "7dfbb06c-f867-4d00-aef7-e731fa8579a7" (UID: "7dfbb06c-f867-4d00-aef7-e731fa8579a7"). InnerVolumeSpecName "kube-api-access-fc2zp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:51:25 crc kubenswrapper[4802]: I1206 03:51:25.059971 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7dfbb06c-f867-4d00-aef7-e731fa8579a7-util" (OuterVolumeSpecName: "util") pod "7dfbb06c-f867-4d00-aef7-e731fa8579a7" (UID: "7dfbb06c-f867-4d00-aef7-e731fa8579a7"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:51:25 crc kubenswrapper[4802]: I1206 03:51:25.138927 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fc2zp\" (UniqueName: \"kubernetes.io/projected/7dfbb06c-f867-4d00-aef7-e731fa8579a7-kube-api-access-fc2zp\") on node \"crc\" DevicePath \"\"" Dec 06 03:51:25 crc kubenswrapper[4802]: I1206 03:51:25.138975 4802 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7dfbb06c-f867-4d00-aef7-e731fa8579a7-util\") on node \"crc\" DevicePath \"\"" Dec 06 03:51:25 crc kubenswrapper[4802]: I1206 03:51:25.138993 4802 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7dfbb06c-f867-4d00-aef7-e731fa8579a7-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 03:51:25 crc kubenswrapper[4802]: I1206 03:51:25.667902 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkz86h" event={"ID":"7dfbb06c-f867-4d00-aef7-e731fa8579a7","Type":"ContainerDied","Data":"3a6c9029d29b5405e78620a921a91605ee99892674d94926598796aeae06ea9e"} Dec 06 03:51:25 crc kubenswrapper[4802]: I1206 03:51:25.667947 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3a6c9029d29b5405e78620a921a91605ee99892674d94926598796aeae06ea9e" Dec 06 03:51:25 crc kubenswrapper[4802]: I1206 03:51:25.667985 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkz86h" Dec 06 03:51:29 crc kubenswrapper[4802]: I1206 03:51:29.825305 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-l8t5h"] Dec 06 03:51:29 crc kubenswrapper[4802]: E1206 03:51:29.825852 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dfbb06c-f867-4d00-aef7-e731fa8579a7" containerName="extract" Dec 06 03:51:29 crc kubenswrapper[4802]: I1206 03:51:29.825865 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dfbb06c-f867-4d00-aef7-e731fa8579a7" containerName="extract" Dec 06 03:51:29 crc kubenswrapper[4802]: E1206 03:51:29.825872 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dfbb06c-f867-4d00-aef7-e731fa8579a7" containerName="util" Dec 06 03:51:29 crc kubenswrapper[4802]: I1206 03:51:29.825878 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dfbb06c-f867-4d00-aef7-e731fa8579a7" containerName="util" Dec 06 03:51:29 crc kubenswrapper[4802]: E1206 03:51:29.825898 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dfbb06c-f867-4d00-aef7-e731fa8579a7" containerName="pull" Dec 06 03:51:29 crc kubenswrapper[4802]: I1206 03:51:29.825903 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dfbb06c-f867-4d00-aef7-e731fa8579a7" containerName="pull" Dec 06 03:51:29 crc kubenswrapper[4802]: I1206 03:51:29.825992 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="7dfbb06c-f867-4d00-aef7-e731fa8579a7" containerName="extract" Dec 06 03:51:29 crc kubenswrapper[4802]: I1206 03:51:29.826326 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-l8t5h" Dec 06 03:51:29 crc kubenswrapper[4802]: I1206 03:51:29.828070 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-b25r5" Dec 06 03:51:29 crc kubenswrapper[4802]: I1206 03:51:29.828344 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 06 03:51:29 crc kubenswrapper[4802]: I1206 03:51:29.828621 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 06 03:51:29 crc kubenswrapper[4802]: I1206 03:51:29.838926 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-l8t5h"] Dec 06 03:51:29 crc kubenswrapper[4802]: I1206 03:51:29.897645 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvmj6\" (UniqueName: \"kubernetes.io/projected/a1e67ef9-99a6-4d92-a5d4-6c00837b8993-kube-api-access-hvmj6\") pod \"nmstate-operator-5b5b58f5c8-l8t5h\" (UID: \"a1e67ef9-99a6-4d92-a5d4-6c00837b8993\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-l8t5h" Dec 06 03:51:29 crc kubenswrapper[4802]: I1206 03:51:29.999643 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvmj6\" (UniqueName: \"kubernetes.io/projected/a1e67ef9-99a6-4d92-a5d4-6c00837b8993-kube-api-access-hvmj6\") pod \"nmstate-operator-5b5b58f5c8-l8t5h\" (UID: \"a1e67ef9-99a6-4d92-a5d4-6c00837b8993\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-l8t5h" Dec 06 03:51:30 crc kubenswrapper[4802]: I1206 03:51:30.023688 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvmj6\" (UniqueName: \"kubernetes.io/projected/a1e67ef9-99a6-4d92-a5d4-6c00837b8993-kube-api-access-hvmj6\") pod \"nmstate-operator-5b5b58f5c8-l8t5h\" (UID: \"a1e67ef9-99a6-4d92-a5d4-6c00837b8993\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-l8t5h" Dec 06 03:51:30 crc kubenswrapper[4802]: I1206 03:51:30.143190 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-l8t5h" Dec 06 03:51:30 crc kubenswrapper[4802]: I1206 03:51:30.329558 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-l8t5h"] Dec 06 03:51:30 crc kubenswrapper[4802]: I1206 03:51:30.695150 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-l8t5h" event={"ID":"a1e67ef9-99a6-4d92-a5d4-6c00837b8993","Type":"ContainerStarted","Data":"dbad8c722c39f1dd65b08f1be6a3f85e471e4ef79f96db0f74b982314b4c924a"} Dec 06 03:51:32 crc kubenswrapper[4802]: I1206 03:51:32.708061 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-l8t5h" event={"ID":"a1e67ef9-99a6-4d92-a5d4-6c00837b8993","Type":"ContainerStarted","Data":"b8c7757de8a6440f2479b4a962ca42e90534c8f8e9877754e6b033dc99cee885"} Dec 06 03:51:32 crc kubenswrapper[4802]: I1206 03:51:32.726242 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-l8t5h" podStartSLOduration=1.862456104 podStartE2EDuration="3.726223272s" podCreationTimestamp="2025-12-06 03:51:29 +0000 UTC" firstStartedPulling="2025-12-06 03:51:30.340518104 +0000 UTC m=+683.212427296" lastFinishedPulling="2025-12-06 03:51:32.204285312 +0000 UTC m=+685.076194464" observedRunningTime="2025-12-06 03:51:32.724802322 +0000 UTC m=+685.596711474" watchObservedRunningTime="2025-12-06 03:51:32.726223272 +0000 UTC m=+685.598132424" Dec 06 03:51:33 crc kubenswrapper[4802]: I1206 03:51:33.658882 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-d8q2s"] Dec 06 03:51:33 crc kubenswrapper[4802]: I1206 03:51:33.659978 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-d8q2s" Dec 06 03:51:33 crc kubenswrapper[4802]: I1206 03:51:33.672374 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-2m9r7" Dec 06 03:51:33 crc kubenswrapper[4802]: I1206 03:51:33.683079 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-d8q2s"] Dec 06 03:51:33 crc kubenswrapper[4802]: I1206 03:51:33.691977 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2trjk"] Dec 06 03:51:33 crc kubenswrapper[4802]: I1206 03:51:33.693039 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2trjk" Dec 06 03:51:33 crc kubenswrapper[4802]: I1206 03:51:33.697781 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 06 03:51:33 crc kubenswrapper[4802]: I1206 03:51:33.699165 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-swx72"] Dec 06 03:51:33 crc kubenswrapper[4802]: I1206 03:51:33.699924 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-swx72" Dec 06 03:51:33 crc kubenswrapper[4802]: I1206 03:51:33.725229 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2trjk"] Dec 06 03:51:33 crc kubenswrapper[4802]: I1206 03:51:33.743538 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dp546\" (UniqueName: \"kubernetes.io/projected/75129a1d-9a3e-4efb-b2bf-bf353a4f6ea2-kube-api-access-dp546\") pod \"nmstate-metrics-7f946cbc9-d8q2s\" (UID: \"75129a1d-9a3e-4efb-b2bf-bf353a4f6ea2\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-d8q2s" Dec 06 03:51:33 crc kubenswrapper[4802]: I1206 03:51:33.743607 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/92122afa-11ef-42cf-9b97-86d3a41c6e13-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-2trjk\" (UID: \"92122afa-11ef-42cf-9b97-86d3a41c6e13\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2trjk" Dec 06 03:51:33 crc kubenswrapper[4802]: I1206 03:51:33.743656 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/485510d8-9971-4648-bb7c-120875a7c00d-nmstate-lock\") pod \"nmstate-handler-swx72\" (UID: \"485510d8-9971-4648-bb7c-120875a7c00d\") " pod="openshift-nmstate/nmstate-handler-swx72" Dec 06 03:51:33 crc kubenswrapper[4802]: I1206 03:51:33.743684 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwjr9\" (UniqueName: \"kubernetes.io/projected/92122afa-11ef-42cf-9b97-86d3a41c6e13-kube-api-access-xwjr9\") pod \"nmstate-webhook-5f6d4c5ccb-2trjk\" (UID: \"92122afa-11ef-42cf-9b97-86d3a41c6e13\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2trjk" Dec 06 03:51:33 crc kubenswrapper[4802]: I1206 03:51:33.743704 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/485510d8-9971-4648-bb7c-120875a7c00d-ovs-socket\") pod \"nmstate-handler-swx72\" (UID: \"485510d8-9971-4648-bb7c-120875a7c00d\") " pod="openshift-nmstate/nmstate-handler-swx72" Dec 06 03:51:33 crc kubenswrapper[4802]: I1206 03:51:33.743781 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gb6k\" (UniqueName: \"kubernetes.io/projected/485510d8-9971-4648-bb7c-120875a7c00d-kube-api-access-2gb6k\") pod \"nmstate-handler-swx72\" (UID: \"485510d8-9971-4648-bb7c-120875a7c00d\") " pod="openshift-nmstate/nmstate-handler-swx72" Dec 06 03:51:33 crc kubenswrapper[4802]: I1206 03:51:33.743807 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/485510d8-9971-4648-bb7c-120875a7c00d-dbus-socket\") pod \"nmstate-handler-swx72\" (UID: \"485510d8-9971-4648-bb7c-120875a7c00d\") " pod="openshift-nmstate/nmstate-handler-swx72" Dec 06 03:51:33 crc kubenswrapper[4802]: I1206 03:51:33.807461 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-8wwln"] Dec 06 03:51:33 crc kubenswrapper[4802]: I1206 03:51:33.810619 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-8wwln" Dec 06 03:51:33 crc kubenswrapper[4802]: I1206 03:51:33.812161 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 06 03:51:33 crc kubenswrapper[4802]: I1206 03:51:33.812345 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-njnnd" Dec 06 03:51:33 crc kubenswrapper[4802]: I1206 03:51:33.812572 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 06 03:51:33 crc kubenswrapper[4802]: I1206 03:51:33.820297 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-8wwln"] Dec 06 03:51:33 crc kubenswrapper[4802]: I1206 03:51:33.844853 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwjr9\" (UniqueName: \"kubernetes.io/projected/92122afa-11ef-42cf-9b97-86d3a41c6e13-kube-api-access-xwjr9\") pod \"nmstate-webhook-5f6d4c5ccb-2trjk\" (UID: \"92122afa-11ef-42cf-9b97-86d3a41c6e13\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2trjk" Dec 06 03:51:33 crc kubenswrapper[4802]: I1206 03:51:33.845214 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/485510d8-9971-4648-bb7c-120875a7c00d-ovs-socket\") pod \"nmstate-handler-swx72\" (UID: \"485510d8-9971-4648-bb7c-120875a7c00d\") " pod="openshift-nmstate/nmstate-handler-swx72" Dec 06 03:51:33 crc kubenswrapper[4802]: I1206 03:51:33.845283 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/485510d8-9971-4648-bb7c-120875a7c00d-dbus-socket\") pod \"nmstate-handler-swx72\" (UID: \"485510d8-9971-4648-bb7c-120875a7c00d\") " pod="openshift-nmstate/nmstate-handler-swx72" Dec 06 03:51:33 crc kubenswrapper[4802]: I1206 03:51:33.845305 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gb6k\" (UniqueName: \"kubernetes.io/projected/485510d8-9971-4648-bb7c-120875a7c00d-kube-api-access-2gb6k\") pod \"nmstate-handler-swx72\" (UID: \"485510d8-9971-4648-bb7c-120875a7c00d\") " pod="openshift-nmstate/nmstate-handler-swx72" Dec 06 03:51:33 crc kubenswrapper[4802]: I1206 03:51:33.845356 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/db89a256-eef7-476e-ab22-755b4d6bb780-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-8wwln\" (UID: \"db89a256-eef7-476e-ab22-755b4d6bb780\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-8wwln" Dec 06 03:51:33 crc kubenswrapper[4802]: I1206 03:51:33.845388 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/db89a256-eef7-476e-ab22-755b4d6bb780-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-8wwln\" (UID: \"db89a256-eef7-476e-ab22-755b4d6bb780\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-8wwln" Dec 06 03:51:33 crc kubenswrapper[4802]: I1206 03:51:33.845415 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dp546\" (UniqueName: \"kubernetes.io/projected/75129a1d-9a3e-4efb-b2bf-bf353a4f6ea2-kube-api-access-dp546\") pod \"nmstate-metrics-7f946cbc9-d8q2s\" (UID: \"75129a1d-9a3e-4efb-b2bf-bf353a4f6ea2\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-d8q2s" Dec 06 03:51:33 crc kubenswrapper[4802]: I1206 03:51:33.845445 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7w7v\" (UniqueName: \"kubernetes.io/projected/db89a256-eef7-476e-ab22-755b4d6bb780-kube-api-access-s7w7v\") pod \"nmstate-console-plugin-7fbb5f6569-8wwln\" (UID: \"db89a256-eef7-476e-ab22-755b4d6bb780\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-8wwln" Dec 06 03:51:33 crc kubenswrapper[4802]: I1206 03:51:33.845476 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/92122afa-11ef-42cf-9b97-86d3a41c6e13-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-2trjk\" (UID: \"92122afa-11ef-42cf-9b97-86d3a41c6e13\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2trjk" Dec 06 03:51:33 crc kubenswrapper[4802]: I1206 03:51:33.845506 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/485510d8-9971-4648-bb7c-120875a7c00d-nmstate-lock\") pod \"nmstate-handler-swx72\" (UID: \"485510d8-9971-4648-bb7c-120875a7c00d\") " pod="openshift-nmstate/nmstate-handler-swx72" Dec 06 03:51:33 crc kubenswrapper[4802]: I1206 03:51:33.845576 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/485510d8-9971-4648-bb7c-120875a7c00d-nmstate-lock\") pod \"nmstate-handler-swx72\" (UID: \"485510d8-9971-4648-bb7c-120875a7c00d\") " pod="openshift-nmstate/nmstate-handler-swx72" Dec 06 03:51:33 crc kubenswrapper[4802]: I1206 03:51:33.845983 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/485510d8-9971-4648-bb7c-120875a7c00d-ovs-socket\") pod \"nmstate-handler-swx72\" (UID: \"485510d8-9971-4648-bb7c-120875a7c00d\") " pod="openshift-nmstate/nmstate-handler-swx72" Dec 06 03:51:33 crc kubenswrapper[4802]: I1206 03:51:33.846220 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/485510d8-9971-4648-bb7c-120875a7c00d-dbus-socket\") pod \"nmstate-handler-swx72\" (UID: \"485510d8-9971-4648-bb7c-120875a7c00d\") " pod="openshift-nmstate/nmstate-handler-swx72" Dec 06 03:51:33 crc kubenswrapper[4802]: E1206 03:51:33.846604 4802 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Dec 06 03:51:33 crc kubenswrapper[4802]: E1206 03:51:33.846659 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/92122afa-11ef-42cf-9b97-86d3a41c6e13-tls-key-pair podName:92122afa-11ef-42cf-9b97-86d3a41c6e13 nodeName:}" failed. No retries permitted until 2025-12-06 03:51:34.346642295 +0000 UTC m=+687.218551447 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/92122afa-11ef-42cf-9b97-86d3a41c6e13-tls-key-pair") pod "nmstate-webhook-5f6d4c5ccb-2trjk" (UID: "92122afa-11ef-42cf-9b97-86d3a41c6e13") : secret "openshift-nmstate-webhook" not found Dec 06 03:51:33 crc kubenswrapper[4802]: I1206 03:51:33.864115 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gb6k\" (UniqueName: \"kubernetes.io/projected/485510d8-9971-4648-bb7c-120875a7c00d-kube-api-access-2gb6k\") pod \"nmstate-handler-swx72\" (UID: \"485510d8-9971-4648-bb7c-120875a7c00d\") " pod="openshift-nmstate/nmstate-handler-swx72" Dec 06 03:51:33 crc kubenswrapper[4802]: I1206 03:51:33.867473 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dp546\" (UniqueName: \"kubernetes.io/projected/75129a1d-9a3e-4efb-b2bf-bf353a4f6ea2-kube-api-access-dp546\") pod \"nmstate-metrics-7f946cbc9-d8q2s\" (UID: \"75129a1d-9a3e-4efb-b2bf-bf353a4f6ea2\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-d8q2s" Dec 06 03:51:33 crc kubenswrapper[4802]: I1206 03:51:33.885056 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwjr9\" (UniqueName: \"kubernetes.io/projected/92122afa-11ef-42cf-9b97-86d3a41c6e13-kube-api-access-xwjr9\") pod \"nmstate-webhook-5f6d4c5ccb-2trjk\" (UID: \"92122afa-11ef-42cf-9b97-86d3a41c6e13\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2trjk" Dec 06 03:51:33 crc kubenswrapper[4802]: I1206 03:51:33.947175 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/db89a256-eef7-476e-ab22-755b4d6bb780-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-8wwln\" (UID: \"db89a256-eef7-476e-ab22-755b4d6bb780\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-8wwln" Dec 06 03:51:33 crc kubenswrapper[4802]: I1206 03:51:33.947237 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/db89a256-eef7-476e-ab22-755b4d6bb780-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-8wwln\" (UID: \"db89a256-eef7-476e-ab22-755b4d6bb780\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-8wwln" Dec 06 03:51:33 crc kubenswrapper[4802]: I1206 03:51:33.947273 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7w7v\" (UniqueName: \"kubernetes.io/projected/db89a256-eef7-476e-ab22-755b4d6bb780-kube-api-access-s7w7v\") pod \"nmstate-console-plugin-7fbb5f6569-8wwln\" (UID: \"db89a256-eef7-476e-ab22-755b4d6bb780\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-8wwln" Dec 06 03:51:33 crc kubenswrapper[4802]: E1206 03:51:33.947381 4802 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Dec 06 03:51:33 crc kubenswrapper[4802]: E1206 03:51:33.947462 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/db89a256-eef7-476e-ab22-755b4d6bb780-plugin-serving-cert podName:db89a256-eef7-476e-ab22-755b4d6bb780 nodeName:}" failed. No retries permitted until 2025-12-06 03:51:34.44744073 +0000 UTC m=+687.319349922 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/db89a256-eef7-476e-ab22-755b4d6bb780-plugin-serving-cert") pod "nmstate-console-plugin-7fbb5f6569-8wwln" (UID: "db89a256-eef7-476e-ab22-755b4d6bb780") : secret "plugin-serving-cert" not found Dec 06 03:51:33 crc kubenswrapper[4802]: I1206 03:51:33.948351 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/db89a256-eef7-476e-ab22-755b4d6bb780-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-8wwln\" (UID: \"db89a256-eef7-476e-ab22-755b4d6bb780\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-8wwln" Dec 06 03:51:33 crc kubenswrapper[4802]: I1206 03:51:33.974612 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7w7v\" (UniqueName: \"kubernetes.io/projected/db89a256-eef7-476e-ab22-755b4d6bb780-kube-api-access-s7w7v\") pod \"nmstate-console-plugin-7fbb5f6569-8wwln\" (UID: \"db89a256-eef7-476e-ab22-755b4d6bb780\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-8wwln" Dec 06 03:51:33 crc kubenswrapper[4802]: I1206 03:51:33.983477 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-d8q2s" Dec 06 03:51:34 crc kubenswrapper[4802]: I1206 03:51:34.015150 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f59d9d7ff-dgk7f"] Dec 06 03:51:34 crc kubenswrapper[4802]: I1206 03:51:34.015842 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f59d9d7ff-dgk7f" Dec 06 03:51:34 crc kubenswrapper[4802]: I1206 03:51:34.036684 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f59d9d7ff-dgk7f"] Dec 06 03:51:34 crc kubenswrapper[4802]: I1206 03:51:34.039204 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-swx72" Dec 06 03:51:34 crc kubenswrapper[4802]: I1206 03:51:34.048448 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/bf145d3c-c18f-465a-ac7e-05dca038788f-console-config\") pod \"console-f59d9d7ff-dgk7f\" (UID: \"bf145d3c-c18f-465a-ac7e-05dca038788f\") " pod="openshift-console/console-f59d9d7ff-dgk7f" Dec 06 03:51:34 crc kubenswrapper[4802]: I1206 03:51:34.048484 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/bf145d3c-c18f-465a-ac7e-05dca038788f-console-oauth-config\") pod \"console-f59d9d7ff-dgk7f\" (UID: \"bf145d3c-c18f-465a-ac7e-05dca038788f\") " pod="openshift-console/console-f59d9d7ff-dgk7f" Dec 06 03:51:34 crc kubenswrapper[4802]: I1206 03:51:34.048534 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/bf145d3c-c18f-465a-ac7e-05dca038788f-console-serving-cert\") pod \"console-f59d9d7ff-dgk7f\" (UID: \"bf145d3c-c18f-465a-ac7e-05dca038788f\") " pod="openshift-console/console-f59d9d7ff-dgk7f" Dec 06 03:51:34 crc kubenswrapper[4802]: I1206 03:51:34.048586 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bf145d3c-c18f-465a-ac7e-05dca038788f-trusted-ca-bundle\") pod \"console-f59d9d7ff-dgk7f\" (UID: \"bf145d3c-c18f-465a-ac7e-05dca038788f\") " pod="openshift-console/console-f59d9d7ff-dgk7f" Dec 06 03:51:34 crc kubenswrapper[4802]: I1206 03:51:34.048606 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/bf145d3c-c18f-465a-ac7e-05dca038788f-service-ca\") pod \"console-f59d9d7ff-dgk7f\" (UID: \"bf145d3c-c18f-465a-ac7e-05dca038788f\") " pod="openshift-console/console-f59d9d7ff-dgk7f" Dec 06 03:51:34 crc kubenswrapper[4802]: I1206 03:51:34.048620 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/bf145d3c-c18f-465a-ac7e-05dca038788f-oauth-serving-cert\") pod \"console-f59d9d7ff-dgk7f\" (UID: \"bf145d3c-c18f-465a-ac7e-05dca038788f\") " pod="openshift-console/console-f59d9d7ff-dgk7f" Dec 06 03:51:34 crc kubenswrapper[4802]: I1206 03:51:34.048639 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2zzn7\" (UniqueName: \"kubernetes.io/projected/bf145d3c-c18f-465a-ac7e-05dca038788f-kube-api-access-2zzn7\") pod \"console-f59d9d7ff-dgk7f\" (UID: \"bf145d3c-c18f-465a-ac7e-05dca038788f\") " pod="openshift-console/console-f59d9d7ff-dgk7f" Dec 06 03:51:34 crc kubenswrapper[4802]: I1206 03:51:34.150208 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/bf145d3c-c18f-465a-ac7e-05dca038788f-console-config\") pod \"console-f59d9d7ff-dgk7f\" (UID: \"bf145d3c-c18f-465a-ac7e-05dca038788f\") " pod="openshift-console/console-f59d9d7ff-dgk7f" Dec 06 03:51:34 crc kubenswrapper[4802]: I1206 03:51:34.150526 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/bf145d3c-c18f-465a-ac7e-05dca038788f-console-oauth-config\") pod \"console-f59d9d7ff-dgk7f\" (UID: \"bf145d3c-c18f-465a-ac7e-05dca038788f\") " pod="openshift-console/console-f59d9d7ff-dgk7f" Dec 06 03:51:34 crc kubenswrapper[4802]: I1206 03:51:34.150578 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/bf145d3c-c18f-465a-ac7e-05dca038788f-console-serving-cert\") pod \"console-f59d9d7ff-dgk7f\" (UID: \"bf145d3c-c18f-465a-ac7e-05dca038788f\") " pod="openshift-console/console-f59d9d7ff-dgk7f" Dec 06 03:51:34 crc kubenswrapper[4802]: I1206 03:51:34.150634 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bf145d3c-c18f-465a-ac7e-05dca038788f-trusted-ca-bundle\") pod \"console-f59d9d7ff-dgk7f\" (UID: \"bf145d3c-c18f-465a-ac7e-05dca038788f\") " pod="openshift-console/console-f59d9d7ff-dgk7f" Dec 06 03:51:34 crc kubenswrapper[4802]: I1206 03:51:34.150655 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/bf145d3c-c18f-465a-ac7e-05dca038788f-service-ca\") pod \"console-f59d9d7ff-dgk7f\" (UID: \"bf145d3c-c18f-465a-ac7e-05dca038788f\") " pod="openshift-console/console-f59d9d7ff-dgk7f" Dec 06 03:51:34 crc kubenswrapper[4802]: I1206 03:51:34.150671 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/bf145d3c-c18f-465a-ac7e-05dca038788f-oauth-serving-cert\") pod \"console-f59d9d7ff-dgk7f\" (UID: \"bf145d3c-c18f-465a-ac7e-05dca038788f\") " pod="openshift-console/console-f59d9d7ff-dgk7f" Dec 06 03:51:34 crc kubenswrapper[4802]: I1206 03:51:34.150931 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2zzn7\" (UniqueName: \"kubernetes.io/projected/bf145d3c-c18f-465a-ac7e-05dca038788f-kube-api-access-2zzn7\") pod \"console-f59d9d7ff-dgk7f\" (UID: \"bf145d3c-c18f-465a-ac7e-05dca038788f\") " pod="openshift-console/console-f59d9d7ff-dgk7f" Dec 06 03:51:34 crc kubenswrapper[4802]: I1206 03:51:34.151536 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/bf145d3c-c18f-465a-ac7e-05dca038788f-console-config\") pod \"console-f59d9d7ff-dgk7f\" (UID: \"bf145d3c-c18f-465a-ac7e-05dca038788f\") " pod="openshift-console/console-f59d9d7ff-dgk7f" Dec 06 03:51:34 crc kubenswrapper[4802]: I1206 03:51:34.152279 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bf145d3c-c18f-465a-ac7e-05dca038788f-trusted-ca-bundle\") pod \"console-f59d9d7ff-dgk7f\" (UID: \"bf145d3c-c18f-465a-ac7e-05dca038788f\") " pod="openshift-console/console-f59d9d7ff-dgk7f" Dec 06 03:51:34 crc kubenswrapper[4802]: I1206 03:51:34.152725 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/bf145d3c-c18f-465a-ac7e-05dca038788f-service-ca\") pod \"console-f59d9d7ff-dgk7f\" (UID: \"bf145d3c-c18f-465a-ac7e-05dca038788f\") " pod="openshift-console/console-f59d9d7ff-dgk7f" Dec 06 03:51:34 crc kubenswrapper[4802]: I1206 03:51:34.152977 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/bf145d3c-c18f-465a-ac7e-05dca038788f-oauth-serving-cert\") pod \"console-f59d9d7ff-dgk7f\" (UID: \"bf145d3c-c18f-465a-ac7e-05dca038788f\") " pod="openshift-console/console-f59d9d7ff-dgk7f" Dec 06 03:51:34 crc kubenswrapper[4802]: I1206 03:51:34.156083 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/bf145d3c-c18f-465a-ac7e-05dca038788f-console-serving-cert\") pod \"console-f59d9d7ff-dgk7f\" (UID: \"bf145d3c-c18f-465a-ac7e-05dca038788f\") " pod="openshift-console/console-f59d9d7ff-dgk7f" Dec 06 03:51:34 crc kubenswrapper[4802]: I1206 03:51:34.156132 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/bf145d3c-c18f-465a-ac7e-05dca038788f-console-oauth-config\") pod \"console-f59d9d7ff-dgk7f\" (UID: \"bf145d3c-c18f-465a-ac7e-05dca038788f\") " pod="openshift-console/console-f59d9d7ff-dgk7f" Dec 06 03:51:34 crc kubenswrapper[4802]: I1206 03:51:34.169933 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2zzn7\" (UniqueName: \"kubernetes.io/projected/bf145d3c-c18f-465a-ac7e-05dca038788f-kube-api-access-2zzn7\") pod \"console-f59d9d7ff-dgk7f\" (UID: \"bf145d3c-c18f-465a-ac7e-05dca038788f\") " pod="openshift-console/console-f59d9d7ff-dgk7f" Dec 06 03:51:34 crc kubenswrapper[4802]: I1206 03:51:34.353336 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/92122afa-11ef-42cf-9b97-86d3a41c6e13-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-2trjk\" (UID: \"92122afa-11ef-42cf-9b97-86d3a41c6e13\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2trjk" Dec 06 03:51:34 crc kubenswrapper[4802]: I1206 03:51:34.356560 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/92122afa-11ef-42cf-9b97-86d3a41c6e13-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-2trjk\" (UID: \"92122afa-11ef-42cf-9b97-86d3a41c6e13\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2trjk" Dec 06 03:51:34 crc kubenswrapper[4802]: I1206 03:51:34.375314 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f59d9d7ff-dgk7f" Dec 06 03:51:34 crc kubenswrapper[4802]: I1206 03:51:34.455278 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-d8q2s"] Dec 06 03:51:34 crc kubenswrapper[4802]: I1206 03:51:34.455690 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/db89a256-eef7-476e-ab22-755b4d6bb780-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-8wwln\" (UID: \"db89a256-eef7-476e-ab22-755b4d6bb780\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-8wwln" Dec 06 03:51:34 crc kubenswrapper[4802]: I1206 03:51:34.461447 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/db89a256-eef7-476e-ab22-755b4d6bb780-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-8wwln\" (UID: \"db89a256-eef7-476e-ab22-755b4d6bb780\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-8wwln" Dec 06 03:51:34 crc kubenswrapper[4802]: I1206 03:51:34.593628 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f59d9d7ff-dgk7f"] Dec 06 03:51:34 crc kubenswrapper[4802]: W1206 03:51:34.601952 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbf145d3c_c18f_465a_ac7e_05dca038788f.slice/crio-d12e6b5a9e8c58d994ac948041bea4e78b97b9d08c551b225100f81de7064f86 WatchSource:0}: Error finding container d12e6b5a9e8c58d994ac948041bea4e78b97b9d08c551b225100f81de7064f86: Status 404 returned error can't find the container with id d12e6b5a9e8c58d994ac948041bea4e78b97b9d08c551b225100f81de7064f86 Dec 06 03:51:34 crc kubenswrapper[4802]: I1206 03:51:34.629798 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2trjk" Dec 06 03:51:34 crc kubenswrapper[4802]: I1206 03:51:34.727154 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-8wwln" Dec 06 03:51:34 crc kubenswrapper[4802]: I1206 03:51:34.730446 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f59d9d7ff-dgk7f" event={"ID":"bf145d3c-c18f-465a-ac7e-05dca038788f","Type":"ContainerStarted","Data":"d12e6b5a9e8c58d994ac948041bea4e78b97b9d08c551b225100f81de7064f86"} Dec 06 03:51:34 crc kubenswrapper[4802]: I1206 03:51:34.731399 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-swx72" event={"ID":"485510d8-9971-4648-bb7c-120875a7c00d","Type":"ContainerStarted","Data":"16b27b944429b4a7ac3d547818c4bbba45de90da90504671c7e8f319e5e1a6e0"} Dec 06 03:51:34 crc kubenswrapper[4802]: I1206 03:51:34.732252 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-d8q2s" event={"ID":"75129a1d-9a3e-4efb-b2bf-bf353a4f6ea2","Type":"ContainerStarted","Data":"4e7c9b28d9a82305dc6f9c88f5d288abe2f4ab033f95f73cd725da5523b7a9e1"} Dec 06 03:51:34 crc kubenswrapper[4802]: I1206 03:51:34.859621 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2trjk"] Dec 06 03:51:34 crc kubenswrapper[4802]: I1206 03:51:34.912807 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-8wwln"] Dec 06 03:51:34 crc kubenswrapper[4802]: W1206 03:51:34.920493 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddb89a256_eef7_476e_ab22_755b4d6bb780.slice/crio-9bbd2988749bd7dfa35c401cd18b0c5bd065cae13270529fee3f88ab7648d32c WatchSource:0}: Error finding container 9bbd2988749bd7dfa35c401cd18b0c5bd065cae13270529fee3f88ab7648d32c: Status 404 returned error can't find the container with id 9bbd2988749bd7dfa35c401cd18b0c5bd065cae13270529fee3f88ab7648d32c Dec 06 03:51:35 crc kubenswrapper[4802]: I1206 03:51:35.739726 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-8wwln" event={"ID":"db89a256-eef7-476e-ab22-755b4d6bb780","Type":"ContainerStarted","Data":"9bbd2988749bd7dfa35c401cd18b0c5bd065cae13270529fee3f88ab7648d32c"} Dec 06 03:51:35 crc kubenswrapper[4802]: I1206 03:51:35.742100 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f59d9d7ff-dgk7f" event={"ID":"bf145d3c-c18f-465a-ac7e-05dca038788f","Type":"ContainerStarted","Data":"8cd8686165a4f23152f45704368cdd38a09bb32d8efb3f514cbe35b8352afbdd"} Dec 06 03:51:35 crc kubenswrapper[4802]: I1206 03:51:35.743134 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2trjk" event={"ID":"92122afa-11ef-42cf-9b97-86d3a41c6e13","Type":"ContainerStarted","Data":"32ce3f2ca24067401bc3387ce32c3128b8aabb7d5bbcb23bba0d0ec799486027"} Dec 06 03:51:35 crc kubenswrapper[4802]: I1206 03:51:35.764058 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f59d9d7ff-dgk7f" podStartSLOduration=2.764035985 podStartE2EDuration="2.764035985s" podCreationTimestamp="2025-12-06 03:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:51:35.759053477 +0000 UTC m=+688.630962629" watchObservedRunningTime="2025-12-06 03:51:35.764035985 +0000 UTC m=+688.635945137" Dec 06 03:51:37 crc kubenswrapper[4802]: I1206 03:51:37.756634 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2trjk" event={"ID":"92122afa-11ef-42cf-9b97-86d3a41c6e13","Type":"ContainerStarted","Data":"4b1b55a8b69c624dba27ca659b1e10de3a899c228ba2f1fb7259dcbc7e4af07b"} Dec 06 03:51:37 crc kubenswrapper[4802]: I1206 03:51:37.757168 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2trjk" Dec 06 03:51:37 crc kubenswrapper[4802]: I1206 03:51:37.761190 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-swx72" event={"ID":"485510d8-9971-4648-bb7c-120875a7c00d","Type":"ContainerStarted","Data":"658f40f4b785a3ec6bac7bc99b5357c23de592a3813e8199c14ff80ebbf184f1"} Dec 06 03:51:37 crc kubenswrapper[4802]: I1206 03:51:37.761795 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-swx72" Dec 06 03:51:37 crc kubenswrapper[4802]: I1206 03:51:37.763848 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-d8q2s" event={"ID":"75129a1d-9a3e-4efb-b2bf-bf353a4f6ea2","Type":"ContainerStarted","Data":"57104df048fe51415348afd34068d83f2ed8bf2d597614fe5e9c1c03a75afe0c"} Dec 06 03:51:37 crc kubenswrapper[4802]: I1206 03:51:37.776862 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2trjk" podStartSLOduration=3.016227012 podStartE2EDuration="4.776843882s" podCreationTimestamp="2025-12-06 03:51:33 +0000 UTC" firstStartedPulling="2025-12-06 03:51:34.873311568 +0000 UTC m=+687.745220730" lastFinishedPulling="2025-12-06 03:51:36.633928458 +0000 UTC m=+689.505837600" observedRunningTime="2025-12-06 03:51:37.77242452 +0000 UTC m=+690.644333692" watchObservedRunningTime="2025-12-06 03:51:37.776843882 +0000 UTC m=+690.648753044" Dec 06 03:51:37 crc kubenswrapper[4802]: I1206 03:51:37.790867 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-swx72" podStartSLOduration=2.26465314 podStartE2EDuration="4.790847759s" podCreationTimestamp="2025-12-06 03:51:33 +0000 UTC" firstStartedPulling="2025-12-06 03:51:34.086594124 +0000 UTC m=+686.958503276" lastFinishedPulling="2025-12-06 03:51:36.612788743 +0000 UTC m=+689.484697895" observedRunningTime="2025-12-06 03:51:37.788413452 +0000 UTC m=+690.660322624" watchObservedRunningTime="2025-12-06 03:51:37.790847759 +0000 UTC m=+690.662756921" Dec 06 03:51:38 crc kubenswrapper[4802]: I1206 03:51:38.770327 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-8wwln" event={"ID":"db89a256-eef7-476e-ab22-755b4d6bb780","Type":"ContainerStarted","Data":"18db2fe2bbea82867b897f6f6eaa985b67b7aaaceb8e339d07c342e1dafe1fc8"} Dec 06 03:51:38 crc kubenswrapper[4802]: I1206 03:51:38.785052 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-8wwln" podStartSLOduration=2.903023366 podStartE2EDuration="5.785031285s" podCreationTimestamp="2025-12-06 03:51:33 +0000 UTC" firstStartedPulling="2025-12-06 03:51:34.92258518 +0000 UTC m=+687.794494332" lastFinishedPulling="2025-12-06 03:51:37.804593099 +0000 UTC m=+690.676502251" observedRunningTime="2025-12-06 03:51:38.784486759 +0000 UTC m=+691.656395921" watchObservedRunningTime="2025-12-06 03:51:38.785031285 +0000 UTC m=+691.656940437" Dec 06 03:51:39 crc kubenswrapper[4802]: I1206 03:51:39.778744 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-d8q2s" event={"ID":"75129a1d-9a3e-4efb-b2bf-bf353a4f6ea2","Type":"ContainerStarted","Data":"e5dbb70a4049816518462c9669730c8c7860c4a0cdc09f9d669dbef211ae2fdc"} Dec 06 03:51:39 crc kubenswrapper[4802]: I1206 03:51:39.799797 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-d8q2s" podStartSLOduration=2.310485275 podStartE2EDuration="6.799776768s" podCreationTimestamp="2025-12-06 03:51:33 +0000 UTC" firstStartedPulling="2025-12-06 03:51:34.489878045 +0000 UTC m=+687.361787197" lastFinishedPulling="2025-12-06 03:51:38.979169538 +0000 UTC m=+691.851078690" observedRunningTime="2025-12-06 03:51:39.794496432 +0000 UTC m=+692.666405584" watchObservedRunningTime="2025-12-06 03:51:39.799776768 +0000 UTC m=+692.671685940" Dec 06 03:51:44 crc kubenswrapper[4802]: I1206 03:51:44.073768 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-swx72" Dec 06 03:51:44 crc kubenswrapper[4802]: I1206 03:51:44.376294 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f59d9d7ff-dgk7f" Dec 06 03:51:44 crc kubenswrapper[4802]: I1206 03:51:44.376353 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f59d9d7ff-dgk7f" Dec 06 03:51:44 crc kubenswrapper[4802]: I1206 03:51:44.384463 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f59d9d7ff-dgk7f" Dec 06 03:51:45 crc kubenswrapper[4802]: I1206 03:51:45.035276 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f59d9d7ff-dgk7f" Dec 06 03:51:45 crc kubenswrapper[4802]: I1206 03:51:45.098222 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-7ckfj"] Dec 06 03:51:54 crc kubenswrapper[4802]: I1206 03:51:54.638064 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2trjk" Dec 06 03:52:10 crc kubenswrapper[4802]: I1206 03:52:10.157013 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-7ckfj" podUID="2362780e-93d1-4f97-bda0-138ed5180bb0" containerName="console" containerID="cri-o://db66933a746dd27be52693eeaeb393a614c2a8afafa38bc15082419ffb470d35" gracePeriod=15 Dec 06 03:52:10 crc kubenswrapper[4802]: I1206 03:52:10.395255 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838j9dk"] Dec 06 03:52:10 crc kubenswrapper[4802]: I1206 03:52:10.397078 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838j9dk" Dec 06 03:52:10 crc kubenswrapper[4802]: I1206 03:52:10.400361 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 06 03:52:10 crc kubenswrapper[4802]: I1206 03:52:10.413316 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838j9dk"] Dec 06 03:52:10 crc kubenswrapper[4802]: I1206 03:52:10.510908 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtt92\" (UniqueName: \"kubernetes.io/projected/207dc360-1c22-429b-b49e-f27bdd02bc03-kube-api-access-dtt92\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838j9dk\" (UID: \"207dc360-1c22-429b-b49e-f27bdd02bc03\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838j9dk" Dec 06 03:52:10 crc kubenswrapper[4802]: I1206 03:52:10.510967 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/207dc360-1c22-429b-b49e-f27bdd02bc03-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838j9dk\" (UID: \"207dc360-1c22-429b-b49e-f27bdd02bc03\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838j9dk" Dec 06 03:52:10 crc kubenswrapper[4802]: I1206 03:52:10.511056 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/207dc360-1c22-429b-b49e-f27bdd02bc03-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838j9dk\" (UID: \"207dc360-1c22-429b-b49e-f27bdd02bc03\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838j9dk" Dec 06 03:52:10 crc kubenswrapper[4802]: I1206 03:52:10.612736 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtt92\" (UniqueName: \"kubernetes.io/projected/207dc360-1c22-429b-b49e-f27bdd02bc03-kube-api-access-dtt92\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838j9dk\" (UID: \"207dc360-1c22-429b-b49e-f27bdd02bc03\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838j9dk" Dec 06 03:52:10 crc kubenswrapper[4802]: I1206 03:52:10.612901 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/207dc360-1c22-429b-b49e-f27bdd02bc03-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838j9dk\" (UID: \"207dc360-1c22-429b-b49e-f27bdd02bc03\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838j9dk" Dec 06 03:52:10 crc kubenswrapper[4802]: I1206 03:52:10.613090 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/207dc360-1c22-429b-b49e-f27bdd02bc03-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838j9dk\" (UID: \"207dc360-1c22-429b-b49e-f27bdd02bc03\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838j9dk" Dec 06 03:52:10 crc kubenswrapper[4802]: I1206 03:52:10.613722 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/207dc360-1c22-429b-b49e-f27bdd02bc03-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838j9dk\" (UID: \"207dc360-1c22-429b-b49e-f27bdd02bc03\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838j9dk" Dec 06 03:52:10 crc kubenswrapper[4802]: I1206 03:52:10.614000 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/207dc360-1c22-429b-b49e-f27bdd02bc03-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838j9dk\" (UID: \"207dc360-1c22-429b-b49e-f27bdd02bc03\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838j9dk" Dec 06 03:52:10 crc kubenswrapper[4802]: I1206 03:52:10.645493 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtt92\" (UniqueName: \"kubernetes.io/projected/207dc360-1c22-429b-b49e-f27bdd02bc03-kube-api-access-dtt92\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838j9dk\" (UID: \"207dc360-1c22-429b-b49e-f27bdd02bc03\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838j9dk" Dec 06 03:52:10 crc kubenswrapper[4802]: I1206 03:52:10.718117 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838j9dk" Dec 06 03:52:11 crc kubenswrapper[4802]: I1206 03:52:11.155321 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-7ckfj_2362780e-93d1-4f97-bda0-138ed5180bb0/console/0.log" Dec 06 03:52:11 crc kubenswrapper[4802]: I1206 03:52:11.155717 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-7ckfj" Dec 06 03:52:11 crc kubenswrapper[4802]: I1206 03:52:11.203331 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-7ckfj_2362780e-93d1-4f97-bda0-138ed5180bb0/console/0.log" Dec 06 03:52:11 crc kubenswrapper[4802]: I1206 03:52:11.203389 4802 generic.go:334] "Generic (PLEG): container finished" podID="2362780e-93d1-4f97-bda0-138ed5180bb0" containerID="db66933a746dd27be52693eeaeb393a614c2a8afafa38bc15082419ffb470d35" exitCode=2 Dec 06 03:52:11 crc kubenswrapper[4802]: I1206 03:52:11.203423 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-7ckfj" event={"ID":"2362780e-93d1-4f97-bda0-138ed5180bb0","Type":"ContainerDied","Data":"db66933a746dd27be52693eeaeb393a614c2a8afafa38bc15082419ffb470d35"} Dec 06 03:52:11 crc kubenswrapper[4802]: I1206 03:52:11.203460 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-7ckfj" event={"ID":"2362780e-93d1-4f97-bda0-138ed5180bb0","Type":"ContainerDied","Data":"d3153a3a6e26c8493c7e1c5ad9c40b8a38751e3962d1244ed255328bae75eedc"} Dec 06 03:52:11 crc kubenswrapper[4802]: I1206 03:52:11.203480 4802 scope.go:117] "RemoveContainer" containerID="db66933a746dd27be52693eeaeb393a614c2a8afafa38bc15082419ffb470d35" Dec 06 03:52:11 crc kubenswrapper[4802]: I1206 03:52:11.203510 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-7ckfj" Dec 06 03:52:11 crc kubenswrapper[4802]: I1206 03:52:11.213879 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838j9dk"] Dec 06 03:52:11 crc kubenswrapper[4802]: I1206 03:52:11.223140 4802 scope.go:117] "RemoveContainer" containerID="db66933a746dd27be52693eeaeb393a614c2a8afafa38bc15082419ffb470d35" Dec 06 03:52:11 crc kubenswrapper[4802]: E1206 03:52:11.223705 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db66933a746dd27be52693eeaeb393a614c2a8afafa38bc15082419ffb470d35\": container with ID starting with db66933a746dd27be52693eeaeb393a614c2a8afafa38bc15082419ffb470d35 not found: ID does not exist" containerID="db66933a746dd27be52693eeaeb393a614c2a8afafa38bc15082419ffb470d35" Dec 06 03:52:11 crc kubenswrapper[4802]: I1206 03:52:11.223738 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db66933a746dd27be52693eeaeb393a614c2a8afafa38bc15082419ffb470d35"} err="failed to get container status \"db66933a746dd27be52693eeaeb393a614c2a8afafa38bc15082419ffb470d35\": rpc error: code = NotFound desc = could not find container \"db66933a746dd27be52693eeaeb393a614c2a8afafa38bc15082419ffb470d35\": container with ID starting with db66933a746dd27be52693eeaeb393a614c2a8afafa38bc15082419ffb470d35 not found: ID does not exist" Dec 06 03:52:11 crc kubenswrapper[4802]: W1206 03:52:11.229449 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod207dc360_1c22_429b_b49e_f27bdd02bc03.slice/crio-1d3d04244861ee53b12fd654fa32b7aa9d69c16935b787f6346762ddbbfd96f7 WatchSource:0}: Error finding container 1d3d04244861ee53b12fd654fa32b7aa9d69c16935b787f6346762ddbbfd96f7: Status 404 returned error can't find the container with id 1d3d04244861ee53b12fd654fa32b7aa9d69c16935b787f6346762ddbbfd96f7 Dec 06 03:52:11 crc kubenswrapper[4802]: I1206 03:52:11.322156 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/2362780e-93d1-4f97-bda0-138ed5180bb0-console-oauth-config\") pod \"2362780e-93d1-4f97-bda0-138ed5180bb0\" (UID: \"2362780e-93d1-4f97-bda0-138ed5180bb0\") " Dec 06 03:52:11 crc kubenswrapper[4802]: I1206 03:52:11.322458 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2362780e-93d1-4f97-bda0-138ed5180bb0-trusted-ca-bundle\") pod \"2362780e-93d1-4f97-bda0-138ed5180bb0\" (UID: \"2362780e-93d1-4f97-bda0-138ed5180bb0\") " Dec 06 03:52:11 crc kubenswrapper[4802]: I1206 03:52:11.322488 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/2362780e-93d1-4f97-bda0-138ed5180bb0-console-config\") pod \"2362780e-93d1-4f97-bda0-138ed5180bb0\" (UID: \"2362780e-93d1-4f97-bda0-138ed5180bb0\") " Dec 06 03:52:11 crc kubenswrapper[4802]: I1206 03:52:11.322533 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2362780e-93d1-4f97-bda0-138ed5180bb0-service-ca\") pod \"2362780e-93d1-4f97-bda0-138ed5180bb0\" (UID: \"2362780e-93d1-4f97-bda0-138ed5180bb0\") " Dec 06 03:52:11 crc kubenswrapper[4802]: I1206 03:52:11.322553 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qbrbz\" (UniqueName: \"kubernetes.io/projected/2362780e-93d1-4f97-bda0-138ed5180bb0-kube-api-access-qbrbz\") pod \"2362780e-93d1-4f97-bda0-138ed5180bb0\" (UID: \"2362780e-93d1-4f97-bda0-138ed5180bb0\") " Dec 06 03:52:11 crc kubenswrapper[4802]: I1206 03:52:11.322610 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/2362780e-93d1-4f97-bda0-138ed5180bb0-oauth-serving-cert\") pod \"2362780e-93d1-4f97-bda0-138ed5180bb0\" (UID: \"2362780e-93d1-4f97-bda0-138ed5180bb0\") " Dec 06 03:52:11 crc kubenswrapper[4802]: I1206 03:52:11.322662 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/2362780e-93d1-4f97-bda0-138ed5180bb0-console-serving-cert\") pod \"2362780e-93d1-4f97-bda0-138ed5180bb0\" (UID: \"2362780e-93d1-4f97-bda0-138ed5180bb0\") " Dec 06 03:52:11 crc kubenswrapper[4802]: I1206 03:52:11.323381 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2362780e-93d1-4f97-bda0-138ed5180bb0-service-ca" (OuterVolumeSpecName: "service-ca") pod "2362780e-93d1-4f97-bda0-138ed5180bb0" (UID: "2362780e-93d1-4f97-bda0-138ed5180bb0"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:52:11 crc kubenswrapper[4802]: I1206 03:52:11.323687 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2362780e-93d1-4f97-bda0-138ed5180bb0-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "2362780e-93d1-4f97-bda0-138ed5180bb0" (UID: "2362780e-93d1-4f97-bda0-138ed5180bb0"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:52:11 crc kubenswrapper[4802]: I1206 03:52:11.324078 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2362780e-93d1-4f97-bda0-138ed5180bb0-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "2362780e-93d1-4f97-bda0-138ed5180bb0" (UID: "2362780e-93d1-4f97-bda0-138ed5180bb0"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:52:11 crc kubenswrapper[4802]: I1206 03:52:11.324183 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2362780e-93d1-4f97-bda0-138ed5180bb0-console-config" (OuterVolumeSpecName: "console-config") pod "2362780e-93d1-4f97-bda0-138ed5180bb0" (UID: "2362780e-93d1-4f97-bda0-138ed5180bb0"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:52:11 crc kubenswrapper[4802]: I1206 03:52:11.327963 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2362780e-93d1-4f97-bda0-138ed5180bb0-kube-api-access-qbrbz" (OuterVolumeSpecName: "kube-api-access-qbrbz") pod "2362780e-93d1-4f97-bda0-138ed5180bb0" (UID: "2362780e-93d1-4f97-bda0-138ed5180bb0"). InnerVolumeSpecName "kube-api-access-qbrbz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:52:11 crc kubenswrapper[4802]: I1206 03:52:11.328368 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2362780e-93d1-4f97-bda0-138ed5180bb0-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "2362780e-93d1-4f97-bda0-138ed5180bb0" (UID: "2362780e-93d1-4f97-bda0-138ed5180bb0"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:52:11 crc kubenswrapper[4802]: I1206 03:52:11.329061 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2362780e-93d1-4f97-bda0-138ed5180bb0-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "2362780e-93d1-4f97-bda0-138ed5180bb0" (UID: "2362780e-93d1-4f97-bda0-138ed5180bb0"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:52:11 crc kubenswrapper[4802]: I1206 03:52:11.424504 4802 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2362780e-93d1-4f97-bda0-138ed5180bb0-service-ca\") on node \"crc\" DevicePath \"\"" Dec 06 03:52:11 crc kubenswrapper[4802]: I1206 03:52:11.424547 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qbrbz\" (UniqueName: \"kubernetes.io/projected/2362780e-93d1-4f97-bda0-138ed5180bb0-kube-api-access-qbrbz\") on node \"crc\" DevicePath \"\"" Dec 06 03:52:11 crc kubenswrapper[4802]: I1206 03:52:11.424561 4802 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/2362780e-93d1-4f97-bda0-138ed5180bb0-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:52:11 crc kubenswrapper[4802]: I1206 03:52:11.424576 4802 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/2362780e-93d1-4f97-bda0-138ed5180bb0-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:52:11 crc kubenswrapper[4802]: I1206 03:52:11.424591 4802 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/2362780e-93d1-4f97-bda0-138ed5180bb0-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:52:11 crc kubenswrapper[4802]: I1206 03:52:11.424602 4802 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2362780e-93d1-4f97-bda0-138ed5180bb0-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 03:52:11 crc kubenswrapper[4802]: I1206 03:52:11.424614 4802 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/2362780e-93d1-4f97-bda0-138ed5180bb0-console-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:52:11 crc kubenswrapper[4802]: I1206 03:52:11.521054 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-7ckfj"] Dec 06 03:52:11 crc kubenswrapper[4802]: I1206 03:52:11.524633 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-7ckfj"] Dec 06 03:52:12 crc kubenswrapper[4802]: I1206 03:52:12.211102 4802 generic.go:334] "Generic (PLEG): container finished" podID="207dc360-1c22-429b-b49e-f27bdd02bc03" containerID="31bbd1c406c3b323589c44da808d151145e3de7904e8ae66e4186320aedb0a37" exitCode=0 Dec 06 03:52:12 crc kubenswrapper[4802]: I1206 03:52:12.211194 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838j9dk" event={"ID":"207dc360-1c22-429b-b49e-f27bdd02bc03","Type":"ContainerDied","Data":"31bbd1c406c3b323589c44da808d151145e3de7904e8ae66e4186320aedb0a37"} Dec 06 03:52:12 crc kubenswrapper[4802]: I1206 03:52:12.214020 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838j9dk" event={"ID":"207dc360-1c22-429b-b49e-f27bdd02bc03","Type":"ContainerStarted","Data":"1d3d04244861ee53b12fd654fa32b7aa9d69c16935b787f6346762ddbbfd96f7"} Dec 06 03:52:13 crc kubenswrapper[4802]: I1206 03:52:13.283456 4802 patch_prober.go:28] interesting pod/machine-config-daemon-zpxxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 03:52:13 crc kubenswrapper[4802]: I1206 03:52:13.283856 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 03:52:13 crc kubenswrapper[4802]: I1206 03:52:13.480572 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2362780e-93d1-4f97-bda0-138ed5180bb0" path="/var/lib/kubelet/pods/2362780e-93d1-4f97-bda0-138ed5180bb0/volumes" Dec 06 03:52:14 crc kubenswrapper[4802]: I1206 03:52:14.229405 4802 generic.go:334] "Generic (PLEG): container finished" podID="207dc360-1c22-429b-b49e-f27bdd02bc03" containerID="5abc5bd0546155b71915ae77cb9d1976f3ca3a45b18a0616fa64bd5dcb41c016" exitCode=0 Dec 06 03:52:14 crc kubenswrapper[4802]: I1206 03:52:14.229447 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838j9dk" event={"ID":"207dc360-1c22-429b-b49e-f27bdd02bc03","Type":"ContainerDied","Data":"5abc5bd0546155b71915ae77cb9d1976f3ca3a45b18a0616fa64bd5dcb41c016"} Dec 06 03:52:15 crc kubenswrapper[4802]: I1206 03:52:15.248100 4802 generic.go:334] "Generic (PLEG): container finished" podID="207dc360-1c22-429b-b49e-f27bdd02bc03" containerID="e4f1600e1ebf530e259ce71e48465f85a5c4ab3c601cd67d76cea5eb7d006dff" exitCode=0 Dec 06 03:52:15 crc kubenswrapper[4802]: I1206 03:52:15.248267 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838j9dk" event={"ID":"207dc360-1c22-429b-b49e-f27bdd02bc03","Type":"ContainerDied","Data":"e4f1600e1ebf530e259ce71e48465f85a5c4ab3c601cd67d76cea5eb7d006dff"} Dec 06 03:52:16 crc kubenswrapper[4802]: I1206 03:52:16.508908 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838j9dk" Dec 06 03:52:16 crc kubenswrapper[4802]: I1206 03:52:16.620669 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dtt92\" (UniqueName: \"kubernetes.io/projected/207dc360-1c22-429b-b49e-f27bdd02bc03-kube-api-access-dtt92\") pod \"207dc360-1c22-429b-b49e-f27bdd02bc03\" (UID: \"207dc360-1c22-429b-b49e-f27bdd02bc03\") " Dec 06 03:52:16 crc kubenswrapper[4802]: I1206 03:52:16.620790 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/207dc360-1c22-429b-b49e-f27bdd02bc03-bundle\") pod \"207dc360-1c22-429b-b49e-f27bdd02bc03\" (UID: \"207dc360-1c22-429b-b49e-f27bdd02bc03\") " Dec 06 03:52:16 crc kubenswrapper[4802]: I1206 03:52:16.620826 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/207dc360-1c22-429b-b49e-f27bdd02bc03-util\") pod \"207dc360-1c22-429b-b49e-f27bdd02bc03\" (UID: \"207dc360-1c22-429b-b49e-f27bdd02bc03\") " Dec 06 03:52:16 crc kubenswrapper[4802]: I1206 03:52:16.622192 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/207dc360-1c22-429b-b49e-f27bdd02bc03-bundle" (OuterVolumeSpecName: "bundle") pod "207dc360-1c22-429b-b49e-f27bdd02bc03" (UID: "207dc360-1c22-429b-b49e-f27bdd02bc03"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:52:16 crc kubenswrapper[4802]: I1206 03:52:16.627407 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/207dc360-1c22-429b-b49e-f27bdd02bc03-kube-api-access-dtt92" (OuterVolumeSpecName: "kube-api-access-dtt92") pod "207dc360-1c22-429b-b49e-f27bdd02bc03" (UID: "207dc360-1c22-429b-b49e-f27bdd02bc03"). InnerVolumeSpecName "kube-api-access-dtt92". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:52:16 crc kubenswrapper[4802]: I1206 03:52:16.717839 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/207dc360-1c22-429b-b49e-f27bdd02bc03-util" (OuterVolumeSpecName: "util") pod "207dc360-1c22-429b-b49e-f27bdd02bc03" (UID: "207dc360-1c22-429b-b49e-f27bdd02bc03"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:52:16 crc kubenswrapper[4802]: I1206 03:52:16.722439 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dtt92\" (UniqueName: \"kubernetes.io/projected/207dc360-1c22-429b-b49e-f27bdd02bc03-kube-api-access-dtt92\") on node \"crc\" DevicePath \"\"" Dec 06 03:52:16 crc kubenswrapper[4802]: I1206 03:52:16.722534 4802 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/207dc360-1c22-429b-b49e-f27bdd02bc03-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 03:52:16 crc kubenswrapper[4802]: I1206 03:52:16.722557 4802 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/207dc360-1c22-429b-b49e-f27bdd02bc03-util\") on node \"crc\" DevicePath \"\"" Dec 06 03:52:17 crc kubenswrapper[4802]: I1206 03:52:17.265134 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838j9dk" event={"ID":"207dc360-1c22-429b-b49e-f27bdd02bc03","Type":"ContainerDied","Data":"1d3d04244861ee53b12fd654fa32b7aa9d69c16935b787f6346762ddbbfd96f7"} Dec 06 03:52:17 crc kubenswrapper[4802]: I1206 03:52:17.265198 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1d3d04244861ee53b12fd654fa32b7aa9d69c16935b787f6346762ddbbfd96f7" Dec 06 03:52:17 crc kubenswrapper[4802]: I1206 03:52:17.265224 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838j9dk" Dec 06 03:52:31 crc kubenswrapper[4802]: I1206 03:52:31.717149 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-5c78fd4b89-5m46c"] Dec 06 03:52:31 crc kubenswrapper[4802]: E1206 03:52:31.717829 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2362780e-93d1-4f97-bda0-138ed5180bb0" containerName="console" Dec 06 03:52:31 crc kubenswrapper[4802]: I1206 03:52:31.717841 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="2362780e-93d1-4f97-bda0-138ed5180bb0" containerName="console" Dec 06 03:52:31 crc kubenswrapper[4802]: E1206 03:52:31.717854 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="207dc360-1c22-429b-b49e-f27bdd02bc03" containerName="pull" Dec 06 03:52:31 crc kubenswrapper[4802]: I1206 03:52:31.717860 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="207dc360-1c22-429b-b49e-f27bdd02bc03" containerName="pull" Dec 06 03:52:31 crc kubenswrapper[4802]: E1206 03:52:31.717870 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="207dc360-1c22-429b-b49e-f27bdd02bc03" containerName="extract" Dec 06 03:52:31 crc kubenswrapper[4802]: I1206 03:52:31.717877 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="207dc360-1c22-429b-b49e-f27bdd02bc03" containerName="extract" Dec 06 03:52:31 crc kubenswrapper[4802]: E1206 03:52:31.717887 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="207dc360-1c22-429b-b49e-f27bdd02bc03" containerName="util" Dec 06 03:52:31 crc kubenswrapper[4802]: I1206 03:52:31.717893 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="207dc360-1c22-429b-b49e-f27bdd02bc03" containerName="util" Dec 06 03:52:31 crc kubenswrapper[4802]: I1206 03:52:31.717996 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="2362780e-93d1-4f97-bda0-138ed5180bb0" containerName="console" Dec 06 03:52:31 crc kubenswrapper[4802]: I1206 03:52:31.718007 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="207dc360-1c22-429b-b49e-f27bdd02bc03" containerName="extract" Dec 06 03:52:31 crc kubenswrapper[4802]: I1206 03:52:31.718344 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5c78fd4b89-5m46c" Dec 06 03:52:31 crc kubenswrapper[4802]: I1206 03:52:31.724728 4802 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 06 03:52:31 crc kubenswrapper[4802]: I1206 03:52:31.728788 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 06 03:52:31 crc kubenswrapper[4802]: I1206 03:52:31.728978 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 06 03:52:31 crc kubenswrapper[4802]: I1206 03:52:31.729058 4802 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-4dt9j" Dec 06 03:52:31 crc kubenswrapper[4802]: I1206 03:52:31.729370 4802 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 06 03:52:31 crc kubenswrapper[4802]: I1206 03:52:31.732049 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5c78fd4b89-5m46c"] Dec 06 03:52:31 crc kubenswrapper[4802]: I1206 03:52:31.817004 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/392ad0d8-35bb-40df-abf2-7a98117f24f7-apiservice-cert\") pod \"metallb-operator-controller-manager-5c78fd4b89-5m46c\" (UID: \"392ad0d8-35bb-40df-abf2-7a98117f24f7\") " pod="metallb-system/metallb-operator-controller-manager-5c78fd4b89-5m46c" Dec 06 03:52:31 crc kubenswrapper[4802]: I1206 03:52:31.817072 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/392ad0d8-35bb-40df-abf2-7a98117f24f7-webhook-cert\") pod \"metallb-operator-controller-manager-5c78fd4b89-5m46c\" (UID: \"392ad0d8-35bb-40df-abf2-7a98117f24f7\") " pod="metallb-system/metallb-operator-controller-manager-5c78fd4b89-5m46c" Dec 06 03:52:31 crc kubenswrapper[4802]: I1206 03:52:31.817315 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxrtj\" (UniqueName: \"kubernetes.io/projected/392ad0d8-35bb-40df-abf2-7a98117f24f7-kube-api-access-mxrtj\") pod \"metallb-operator-controller-manager-5c78fd4b89-5m46c\" (UID: \"392ad0d8-35bb-40df-abf2-7a98117f24f7\") " pod="metallb-system/metallb-operator-controller-manager-5c78fd4b89-5m46c" Dec 06 03:52:31 crc kubenswrapper[4802]: I1206 03:52:31.918704 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/392ad0d8-35bb-40df-abf2-7a98117f24f7-webhook-cert\") pod \"metallb-operator-controller-manager-5c78fd4b89-5m46c\" (UID: \"392ad0d8-35bb-40df-abf2-7a98117f24f7\") " pod="metallb-system/metallb-operator-controller-manager-5c78fd4b89-5m46c" Dec 06 03:52:31 crc kubenswrapper[4802]: I1206 03:52:31.918800 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxrtj\" (UniqueName: \"kubernetes.io/projected/392ad0d8-35bb-40df-abf2-7a98117f24f7-kube-api-access-mxrtj\") pod \"metallb-operator-controller-manager-5c78fd4b89-5m46c\" (UID: \"392ad0d8-35bb-40df-abf2-7a98117f24f7\") " pod="metallb-system/metallb-operator-controller-manager-5c78fd4b89-5m46c" Dec 06 03:52:31 crc kubenswrapper[4802]: I1206 03:52:31.918869 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/392ad0d8-35bb-40df-abf2-7a98117f24f7-apiservice-cert\") pod \"metallb-operator-controller-manager-5c78fd4b89-5m46c\" (UID: \"392ad0d8-35bb-40df-abf2-7a98117f24f7\") " pod="metallb-system/metallb-operator-controller-manager-5c78fd4b89-5m46c" Dec 06 03:52:31 crc kubenswrapper[4802]: I1206 03:52:31.926517 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/392ad0d8-35bb-40df-abf2-7a98117f24f7-apiservice-cert\") pod \"metallb-operator-controller-manager-5c78fd4b89-5m46c\" (UID: \"392ad0d8-35bb-40df-abf2-7a98117f24f7\") " pod="metallb-system/metallb-operator-controller-manager-5c78fd4b89-5m46c" Dec 06 03:52:31 crc kubenswrapper[4802]: I1206 03:52:31.926545 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/392ad0d8-35bb-40df-abf2-7a98117f24f7-webhook-cert\") pod \"metallb-operator-controller-manager-5c78fd4b89-5m46c\" (UID: \"392ad0d8-35bb-40df-abf2-7a98117f24f7\") " pod="metallb-system/metallb-operator-controller-manager-5c78fd4b89-5m46c" Dec 06 03:52:31 crc kubenswrapper[4802]: I1206 03:52:31.941670 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxrtj\" (UniqueName: \"kubernetes.io/projected/392ad0d8-35bb-40df-abf2-7a98117f24f7-kube-api-access-mxrtj\") pod \"metallb-operator-controller-manager-5c78fd4b89-5m46c\" (UID: \"392ad0d8-35bb-40df-abf2-7a98117f24f7\") " pod="metallb-system/metallb-operator-controller-manager-5c78fd4b89-5m46c" Dec 06 03:52:31 crc kubenswrapper[4802]: I1206 03:52:31.960129 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-868c9d9dfb-lxrtk"] Dec 06 03:52:31 crc kubenswrapper[4802]: I1206 03:52:31.960939 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-868c9d9dfb-lxrtk" Dec 06 03:52:31 crc kubenswrapper[4802]: I1206 03:52:31.963012 4802 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 06 03:52:31 crc kubenswrapper[4802]: I1206 03:52:31.963143 4802 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-lnzhd" Dec 06 03:52:31 crc kubenswrapper[4802]: I1206 03:52:31.963230 4802 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 06 03:52:31 crc kubenswrapper[4802]: I1206 03:52:31.992108 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-868c9d9dfb-lxrtk"] Dec 06 03:52:32 crc kubenswrapper[4802]: I1206 03:52:32.020159 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9b4b22b1-625a-4bd0-9c82-a142a6f7f806-webhook-cert\") pod \"metallb-operator-webhook-server-868c9d9dfb-lxrtk\" (UID: \"9b4b22b1-625a-4bd0-9c82-a142a6f7f806\") " pod="metallb-system/metallb-operator-webhook-server-868c9d9dfb-lxrtk" Dec 06 03:52:32 crc kubenswrapper[4802]: I1206 03:52:32.020218 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jnth\" (UniqueName: \"kubernetes.io/projected/9b4b22b1-625a-4bd0-9c82-a142a6f7f806-kube-api-access-2jnth\") pod \"metallb-operator-webhook-server-868c9d9dfb-lxrtk\" (UID: \"9b4b22b1-625a-4bd0-9c82-a142a6f7f806\") " pod="metallb-system/metallb-operator-webhook-server-868c9d9dfb-lxrtk" Dec 06 03:52:32 crc kubenswrapper[4802]: I1206 03:52:32.020252 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9b4b22b1-625a-4bd0-9c82-a142a6f7f806-apiservice-cert\") pod \"metallb-operator-webhook-server-868c9d9dfb-lxrtk\" (UID: \"9b4b22b1-625a-4bd0-9c82-a142a6f7f806\") " pod="metallb-system/metallb-operator-webhook-server-868c9d9dfb-lxrtk" Dec 06 03:52:32 crc kubenswrapper[4802]: I1206 03:52:32.097005 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5c78fd4b89-5m46c" Dec 06 03:52:32 crc kubenswrapper[4802]: I1206 03:52:32.121149 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9b4b22b1-625a-4bd0-9c82-a142a6f7f806-apiservice-cert\") pod \"metallb-operator-webhook-server-868c9d9dfb-lxrtk\" (UID: \"9b4b22b1-625a-4bd0-9c82-a142a6f7f806\") " pod="metallb-system/metallb-operator-webhook-server-868c9d9dfb-lxrtk" Dec 06 03:52:32 crc kubenswrapper[4802]: I1206 03:52:32.121222 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9b4b22b1-625a-4bd0-9c82-a142a6f7f806-webhook-cert\") pod \"metallb-operator-webhook-server-868c9d9dfb-lxrtk\" (UID: \"9b4b22b1-625a-4bd0-9c82-a142a6f7f806\") " pod="metallb-system/metallb-operator-webhook-server-868c9d9dfb-lxrtk" Dec 06 03:52:32 crc kubenswrapper[4802]: I1206 03:52:32.121263 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jnth\" (UniqueName: \"kubernetes.io/projected/9b4b22b1-625a-4bd0-9c82-a142a6f7f806-kube-api-access-2jnth\") pod \"metallb-operator-webhook-server-868c9d9dfb-lxrtk\" (UID: \"9b4b22b1-625a-4bd0-9c82-a142a6f7f806\") " pod="metallb-system/metallb-operator-webhook-server-868c9d9dfb-lxrtk" Dec 06 03:52:32 crc kubenswrapper[4802]: I1206 03:52:32.126336 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9b4b22b1-625a-4bd0-9c82-a142a6f7f806-apiservice-cert\") pod \"metallb-operator-webhook-server-868c9d9dfb-lxrtk\" (UID: \"9b4b22b1-625a-4bd0-9c82-a142a6f7f806\") " pod="metallb-system/metallb-operator-webhook-server-868c9d9dfb-lxrtk" Dec 06 03:52:32 crc kubenswrapper[4802]: I1206 03:52:32.127240 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9b4b22b1-625a-4bd0-9c82-a142a6f7f806-webhook-cert\") pod \"metallb-operator-webhook-server-868c9d9dfb-lxrtk\" (UID: \"9b4b22b1-625a-4bd0-9c82-a142a6f7f806\") " pod="metallb-system/metallb-operator-webhook-server-868c9d9dfb-lxrtk" Dec 06 03:52:32 crc kubenswrapper[4802]: I1206 03:52:32.149378 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jnth\" (UniqueName: \"kubernetes.io/projected/9b4b22b1-625a-4bd0-9c82-a142a6f7f806-kube-api-access-2jnth\") pod \"metallb-operator-webhook-server-868c9d9dfb-lxrtk\" (UID: \"9b4b22b1-625a-4bd0-9c82-a142a6f7f806\") " pod="metallb-system/metallb-operator-webhook-server-868c9d9dfb-lxrtk" Dec 06 03:52:32 crc kubenswrapper[4802]: I1206 03:52:32.310960 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-868c9d9dfb-lxrtk" Dec 06 03:52:32 crc kubenswrapper[4802]: I1206 03:52:32.425794 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5c78fd4b89-5m46c"] Dec 06 03:52:32 crc kubenswrapper[4802]: I1206 03:52:32.552502 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-868c9d9dfb-lxrtk"] Dec 06 03:52:32 crc kubenswrapper[4802]: W1206 03:52:32.556323 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9b4b22b1_625a_4bd0_9c82_a142a6f7f806.slice/crio-111e37d635ebdbd70e62467ba811f6cd9d243da960c4af2cdf33213866ab7ae4 WatchSource:0}: Error finding container 111e37d635ebdbd70e62467ba811f6cd9d243da960c4af2cdf33213866ab7ae4: Status 404 returned error can't find the container with id 111e37d635ebdbd70e62467ba811f6cd9d243da960c4af2cdf33213866ab7ae4 Dec 06 03:52:33 crc kubenswrapper[4802]: I1206 03:52:33.383396 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5c78fd4b89-5m46c" event={"ID":"392ad0d8-35bb-40df-abf2-7a98117f24f7","Type":"ContainerStarted","Data":"eb3866e1896b4bfd115f2499da476380f2bd06c147274797422c2ae67a1c1b81"} Dec 06 03:52:33 crc kubenswrapper[4802]: I1206 03:52:33.394429 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-868c9d9dfb-lxrtk" event={"ID":"9b4b22b1-625a-4bd0-9c82-a142a6f7f806","Type":"ContainerStarted","Data":"111e37d635ebdbd70e62467ba811f6cd9d243da960c4af2cdf33213866ab7ae4"} Dec 06 03:52:38 crc kubenswrapper[4802]: I1206 03:52:38.424099 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-868c9d9dfb-lxrtk" event={"ID":"9b4b22b1-625a-4bd0-9c82-a142a6f7f806","Type":"ContainerStarted","Data":"bf5f42c1d75e595e2d25f526cf8712f7a4e8f6759acf5e5bb07e6e5c9bd486dd"} Dec 06 03:52:38 crc kubenswrapper[4802]: I1206 03:52:38.424618 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-868c9d9dfb-lxrtk" Dec 06 03:52:38 crc kubenswrapper[4802]: I1206 03:52:38.428232 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5c78fd4b89-5m46c" event={"ID":"392ad0d8-35bb-40df-abf2-7a98117f24f7","Type":"ContainerStarted","Data":"88ce05287b910c68011b67cdcecb1d0e1578a85a16ffb4c6eb4ae36ce83a2776"} Dec 06 03:52:38 crc kubenswrapper[4802]: I1206 03:52:38.428519 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-5c78fd4b89-5m46c" Dec 06 03:52:38 crc kubenswrapper[4802]: I1206 03:52:38.453667 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-868c9d9dfb-lxrtk" podStartSLOduration=2.49780291 podStartE2EDuration="7.453644194s" podCreationTimestamp="2025-12-06 03:52:31 +0000 UTC" firstStartedPulling="2025-12-06 03:52:32.559069423 +0000 UTC m=+745.430978575" lastFinishedPulling="2025-12-06 03:52:37.514910707 +0000 UTC m=+750.386819859" observedRunningTime="2025-12-06 03:52:38.450017356 +0000 UTC m=+751.321926528" watchObservedRunningTime="2025-12-06 03:52:38.453644194 +0000 UTC m=+751.325553356" Dec 06 03:52:38 crc kubenswrapper[4802]: I1206 03:52:38.476240 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-5c78fd4b89-5m46c" podStartSLOduration=2.405861557 podStartE2EDuration="7.476221501s" podCreationTimestamp="2025-12-06 03:52:31 +0000 UTC" firstStartedPulling="2025-12-06 03:52:32.436363669 +0000 UTC m=+745.308272821" lastFinishedPulling="2025-12-06 03:52:37.506723613 +0000 UTC m=+750.378632765" observedRunningTime="2025-12-06 03:52:38.473194679 +0000 UTC m=+751.345103861" watchObservedRunningTime="2025-12-06 03:52:38.476221501 +0000 UTC m=+751.348130663" Dec 06 03:52:43 crc kubenswrapper[4802]: I1206 03:52:43.283856 4802 patch_prober.go:28] interesting pod/machine-config-daemon-zpxxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 03:52:43 crc kubenswrapper[4802]: I1206 03:52:43.284644 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 03:52:47 crc kubenswrapper[4802]: I1206 03:52:47.019631 4802 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 06 03:52:52 crc kubenswrapper[4802]: I1206 03:52:52.316800 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-868c9d9dfb-lxrtk" Dec 06 03:53:12 crc kubenswrapper[4802]: I1206 03:53:12.100520 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-5c78fd4b89-5m46c" Dec 06 03:53:12 crc kubenswrapper[4802]: I1206 03:53:12.993365 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-nzpzt"] Dec 06 03:53:12 crc kubenswrapper[4802]: I1206 03:53:12.996138 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-nzpzt" Dec 06 03:53:12 crc kubenswrapper[4802]: I1206 03:53:12.998399 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-bjbtx"] Dec 06 03:53:12 crc kubenswrapper[4802]: I1206 03:53:12.998505 4802 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 06 03:53:12 crc kubenswrapper[4802]: I1206 03:53:12.998545 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 06 03:53:12 crc kubenswrapper[4802]: I1206 03:53:12.999250 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-bjbtx" Dec 06 03:53:12 crc kubenswrapper[4802]: I1206 03:53:12.999437 4802 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-wrbtl" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.001376 4802 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.012254 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-bjbtx"] Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.013351 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kpwg6\" (UniqueName: \"kubernetes.io/projected/9ff1fa79-a414-498e-9f64-659502eb6ae7-kube-api-access-kpwg6\") pod \"frr-k8s-nzpzt\" (UID: \"9ff1fa79-a414-498e-9f64-659502eb6ae7\") " pod="metallb-system/frr-k8s-nzpzt" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.013435 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9ff1fa79-a414-498e-9f64-659502eb6ae7-metrics-certs\") pod \"frr-k8s-nzpzt\" (UID: \"9ff1fa79-a414-498e-9f64-659502eb6ae7\") " pod="metallb-system/frr-k8s-nzpzt" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.013460 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/9ff1fa79-a414-498e-9f64-659502eb6ae7-frr-startup\") pod \"frr-k8s-nzpzt\" (UID: \"9ff1fa79-a414-498e-9f64-659502eb6ae7\") " pod="metallb-system/frr-k8s-nzpzt" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.013493 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/9ff1fa79-a414-498e-9f64-659502eb6ae7-reloader\") pod \"frr-k8s-nzpzt\" (UID: \"9ff1fa79-a414-498e-9f64-659502eb6ae7\") " pod="metallb-system/frr-k8s-nzpzt" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.013516 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/94e7ed3e-c72c-4130-81e6-ef954e2e62dd-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-bjbtx\" (UID: \"94e7ed3e-c72c-4130-81e6-ef954e2e62dd\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-bjbtx" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.013551 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/9ff1fa79-a414-498e-9f64-659502eb6ae7-frr-sockets\") pod \"frr-k8s-nzpzt\" (UID: \"9ff1fa79-a414-498e-9f64-659502eb6ae7\") " pod="metallb-system/frr-k8s-nzpzt" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.013576 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/9ff1fa79-a414-498e-9f64-659502eb6ae7-frr-conf\") pod \"frr-k8s-nzpzt\" (UID: \"9ff1fa79-a414-498e-9f64-659502eb6ae7\") " pod="metallb-system/frr-k8s-nzpzt" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.013598 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/9ff1fa79-a414-498e-9f64-659502eb6ae7-metrics\") pod \"frr-k8s-nzpzt\" (UID: \"9ff1fa79-a414-498e-9f64-659502eb6ae7\") " pod="metallb-system/frr-k8s-nzpzt" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.013646 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4584\" (UniqueName: \"kubernetes.io/projected/94e7ed3e-c72c-4130-81e6-ef954e2e62dd-kube-api-access-g4584\") pod \"frr-k8s-webhook-server-7fcb986d4-bjbtx\" (UID: \"94e7ed3e-c72c-4130-81e6-ef954e2e62dd\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-bjbtx" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.086324 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-rlwm5"] Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.087323 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-rlwm5" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.089242 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.090103 4802 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-bn92m" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.090332 4802 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.092723 4802 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.101382 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-46vm8"] Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.119792 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctkmp\" (UniqueName: \"kubernetes.io/projected/1e43b1ba-4d5f-47a4-8f9c-359875710a1e-kube-api-access-ctkmp\") pod \"speaker-rlwm5\" (UID: \"1e43b1ba-4d5f-47a4-8f9c-359875710a1e\") " pod="metallb-system/speaker-rlwm5" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.119846 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4584\" (UniqueName: \"kubernetes.io/projected/94e7ed3e-c72c-4130-81e6-ef954e2e62dd-kube-api-access-g4584\") pod \"frr-k8s-webhook-server-7fcb986d4-bjbtx\" (UID: \"94e7ed3e-c72c-4130-81e6-ef954e2e62dd\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-bjbtx" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.119885 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1e43b1ba-4d5f-47a4-8f9c-359875710a1e-metrics-certs\") pod \"speaker-rlwm5\" (UID: \"1e43b1ba-4d5f-47a4-8f9c-359875710a1e\") " pod="metallb-system/speaker-rlwm5" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.119912 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kpwg6\" (UniqueName: \"kubernetes.io/projected/9ff1fa79-a414-498e-9f64-659502eb6ae7-kube-api-access-kpwg6\") pod \"frr-k8s-nzpzt\" (UID: \"9ff1fa79-a414-498e-9f64-659502eb6ae7\") " pod="metallb-system/frr-k8s-nzpzt" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.119935 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9ff1fa79-a414-498e-9f64-659502eb6ae7-metrics-certs\") pod \"frr-k8s-nzpzt\" (UID: \"9ff1fa79-a414-498e-9f64-659502eb6ae7\") " pod="metallb-system/frr-k8s-nzpzt" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.119953 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/9ff1fa79-a414-498e-9f64-659502eb6ae7-frr-startup\") pod \"frr-k8s-nzpzt\" (UID: \"9ff1fa79-a414-498e-9f64-659502eb6ae7\") " pod="metallb-system/frr-k8s-nzpzt" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.119972 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/1e43b1ba-4d5f-47a4-8f9c-359875710a1e-metallb-excludel2\") pod \"speaker-rlwm5\" (UID: \"1e43b1ba-4d5f-47a4-8f9c-359875710a1e\") " pod="metallb-system/speaker-rlwm5" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.119993 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/1e43b1ba-4d5f-47a4-8f9c-359875710a1e-memberlist\") pod \"speaker-rlwm5\" (UID: \"1e43b1ba-4d5f-47a4-8f9c-359875710a1e\") " pod="metallb-system/speaker-rlwm5" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.120017 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/9ff1fa79-a414-498e-9f64-659502eb6ae7-reloader\") pod \"frr-k8s-nzpzt\" (UID: \"9ff1fa79-a414-498e-9f64-659502eb6ae7\") " pod="metallb-system/frr-k8s-nzpzt" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.120039 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/94e7ed3e-c72c-4130-81e6-ef954e2e62dd-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-bjbtx\" (UID: \"94e7ed3e-c72c-4130-81e6-ef954e2e62dd\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-bjbtx" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.120071 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/9ff1fa79-a414-498e-9f64-659502eb6ae7-frr-sockets\") pod \"frr-k8s-nzpzt\" (UID: \"9ff1fa79-a414-498e-9f64-659502eb6ae7\") " pod="metallb-system/frr-k8s-nzpzt" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.120090 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/9ff1fa79-a414-498e-9f64-659502eb6ae7-frr-conf\") pod \"frr-k8s-nzpzt\" (UID: \"9ff1fa79-a414-498e-9f64-659502eb6ae7\") " pod="metallb-system/frr-k8s-nzpzt" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.120108 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/9ff1fa79-a414-498e-9f64-659502eb6ae7-metrics\") pod \"frr-k8s-nzpzt\" (UID: \"9ff1fa79-a414-498e-9f64-659502eb6ae7\") " pod="metallb-system/frr-k8s-nzpzt" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.120466 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/9ff1fa79-a414-498e-9f64-659502eb6ae7-metrics\") pod \"frr-k8s-nzpzt\" (UID: \"9ff1fa79-a414-498e-9f64-659502eb6ae7\") " pod="metallb-system/frr-k8s-nzpzt" Dec 06 03:53:13 crc kubenswrapper[4802]: E1206 03:53:13.120927 4802 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Dec 06 03:53:13 crc kubenswrapper[4802]: E1206 03:53:13.120993 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9ff1fa79-a414-498e-9f64-659502eb6ae7-metrics-certs podName:9ff1fa79-a414-498e-9f64-659502eb6ae7 nodeName:}" failed. No retries permitted until 2025-12-06 03:53:13.620963742 +0000 UTC m=+786.492872894 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9ff1fa79-a414-498e-9f64-659502eb6ae7-metrics-certs") pod "frr-k8s-nzpzt" (UID: "9ff1fa79-a414-498e-9f64-659502eb6ae7") : secret "frr-k8s-certs-secret" not found Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.121622 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/9ff1fa79-a414-498e-9f64-659502eb6ae7-frr-startup\") pod \"frr-k8s-nzpzt\" (UID: \"9ff1fa79-a414-498e-9f64-659502eb6ae7\") " pod="metallb-system/frr-k8s-nzpzt" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.121841 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/9ff1fa79-a414-498e-9f64-659502eb6ae7-reloader\") pod \"frr-k8s-nzpzt\" (UID: \"9ff1fa79-a414-498e-9f64-659502eb6ae7\") " pod="metallb-system/frr-k8s-nzpzt" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.122051 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/9ff1fa79-a414-498e-9f64-659502eb6ae7-frr-conf\") pod \"frr-k8s-nzpzt\" (UID: \"9ff1fa79-a414-498e-9f64-659502eb6ae7\") " pod="metallb-system/frr-k8s-nzpzt" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.122179 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/9ff1fa79-a414-498e-9f64-659502eb6ae7-frr-sockets\") pod \"frr-k8s-nzpzt\" (UID: \"9ff1fa79-a414-498e-9f64-659502eb6ae7\") " pod="metallb-system/frr-k8s-nzpzt" Dec 06 03:53:13 crc kubenswrapper[4802]: E1206 03:53:13.122266 4802 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Dec 06 03:53:13 crc kubenswrapper[4802]: E1206 03:53:13.122298 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/94e7ed3e-c72c-4130-81e6-ef954e2e62dd-cert podName:94e7ed3e-c72c-4130-81e6-ef954e2e62dd nodeName:}" failed. No retries permitted until 2025-12-06 03:53:13.622289499 +0000 UTC m=+786.494198651 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/94e7ed3e-c72c-4130-81e6-ef954e2e62dd-cert") pod "frr-k8s-webhook-server-7fcb986d4-bjbtx" (UID: "94e7ed3e-c72c-4130-81e6-ef954e2e62dd") : secret "frr-k8s-webhook-server-cert" not found Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.122823 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-46vm8" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.127802 4802 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.141858 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-46vm8"] Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.162513 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kpwg6\" (UniqueName: \"kubernetes.io/projected/9ff1fa79-a414-498e-9f64-659502eb6ae7-kube-api-access-kpwg6\") pod \"frr-k8s-nzpzt\" (UID: \"9ff1fa79-a414-498e-9f64-659502eb6ae7\") " pod="metallb-system/frr-k8s-nzpzt" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.180741 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4584\" (UniqueName: \"kubernetes.io/projected/94e7ed3e-c72c-4130-81e6-ef954e2e62dd-kube-api-access-g4584\") pod \"frr-k8s-webhook-server-7fcb986d4-bjbtx\" (UID: \"94e7ed3e-c72c-4130-81e6-ef954e2e62dd\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-bjbtx" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.221033 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jk476\" (UniqueName: \"kubernetes.io/projected/3a078497-d2a1-4be5-87ee-ba151c992ecd-kube-api-access-jk476\") pod \"controller-f8648f98b-46vm8\" (UID: \"3a078497-d2a1-4be5-87ee-ba151c992ecd\") " pod="metallb-system/controller-f8648f98b-46vm8" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.221098 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctkmp\" (UniqueName: \"kubernetes.io/projected/1e43b1ba-4d5f-47a4-8f9c-359875710a1e-kube-api-access-ctkmp\") pod \"speaker-rlwm5\" (UID: \"1e43b1ba-4d5f-47a4-8f9c-359875710a1e\") " pod="metallb-system/speaker-rlwm5" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.221126 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3a078497-d2a1-4be5-87ee-ba151c992ecd-cert\") pod \"controller-f8648f98b-46vm8\" (UID: \"3a078497-d2a1-4be5-87ee-ba151c992ecd\") " pod="metallb-system/controller-f8648f98b-46vm8" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.221165 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1e43b1ba-4d5f-47a4-8f9c-359875710a1e-metrics-certs\") pod \"speaker-rlwm5\" (UID: \"1e43b1ba-4d5f-47a4-8f9c-359875710a1e\") " pod="metallb-system/speaker-rlwm5" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.221211 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/1e43b1ba-4d5f-47a4-8f9c-359875710a1e-metallb-excludel2\") pod \"speaker-rlwm5\" (UID: \"1e43b1ba-4d5f-47a4-8f9c-359875710a1e\") " pod="metallb-system/speaker-rlwm5" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.221236 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/1e43b1ba-4d5f-47a4-8f9c-359875710a1e-memberlist\") pod \"speaker-rlwm5\" (UID: \"1e43b1ba-4d5f-47a4-8f9c-359875710a1e\") " pod="metallb-system/speaker-rlwm5" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.221260 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3a078497-d2a1-4be5-87ee-ba151c992ecd-metrics-certs\") pod \"controller-f8648f98b-46vm8\" (UID: \"3a078497-d2a1-4be5-87ee-ba151c992ecd\") " pod="metallb-system/controller-f8648f98b-46vm8" Dec 06 03:53:13 crc kubenswrapper[4802]: E1206 03:53:13.221430 4802 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 06 03:53:13 crc kubenswrapper[4802]: E1206 03:53:13.221524 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1e43b1ba-4d5f-47a4-8f9c-359875710a1e-memberlist podName:1e43b1ba-4d5f-47a4-8f9c-359875710a1e nodeName:}" failed. No retries permitted until 2025-12-06 03:53:13.72150459 +0000 UTC m=+786.593413742 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/1e43b1ba-4d5f-47a4-8f9c-359875710a1e-memberlist") pod "speaker-rlwm5" (UID: "1e43b1ba-4d5f-47a4-8f9c-359875710a1e") : secret "metallb-memberlist" not found Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.221954 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/1e43b1ba-4d5f-47a4-8f9c-359875710a1e-metallb-excludel2\") pod \"speaker-rlwm5\" (UID: \"1e43b1ba-4d5f-47a4-8f9c-359875710a1e\") " pod="metallb-system/speaker-rlwm5" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.226248 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1e43b1ba-4d5f-47a4-8f9c-359875710a1e-metrics-certs\") pod \"speaker-rlwm5\" (UID: \"1e43b1ba-4d5f-47a4-8f9c-359875710a1e\") " pod="metallb-system/speaker-rlwm5" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.270289 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctkmp\" (UniqueName: \"kubernetes.io/projected/1e43b1ba-4d5f-47a4-8f9c-359875710a1e-kube-api-access-ctkmp\") pod \"speaker-rlwm5\" (UID: \"1e43b1ba-4d5f-47a4-8f9c-359875710a1e\") " pod="metallb-system/speaker-rlwm5" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.282994 4802 patch_prober.go:28] interesting pod/machine-config-daemon-zpxxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.283067 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.283131 4802 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.283821 4802 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"57ad81d5a8f13aaa4dc17e3a8d71a6ae21755975dbb001e2d8346604e492cf3d"} pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.283894 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" containerID="cri-o://57ad81d5a8f13aaa4dc17e3a8d71a6ae21755975dbb001e2d8346604e492cf3d" gracePeriod=600 Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.321772 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jk476\" (UniqueName: \"kubernetes.io/projected/3a078497-d2a1-4be5-87ee-ba151c992ecd-kube-api-access-jk476\") pod \"controller-f8648f98b-46vm8\" (UID: \"3a078497-d2a1-4be5-87ee-ba151c992ecd\") " pod="metallb-system/controller-f8648f98b-46vm8" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.321833 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3a078497-d2a1-4be5-87ee-ba151c992ecd-cert\") pod \"controller-f8648f98b-46vm8\" (UID: \"3a078497-d2a1-4be5-87ee-ba151c992ecd\") " pod="metallb-system/controller-f8648f98b-46vm8" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.321921 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3a078497-d2a1-4be5-87ee-ba151c992ecd-metrics-certs\") pod \"controller-f8648f98b-46vm8\" (UID: \"3a078497-d2a1-4be5-87ee-ba151c992ecd\") " pod="metallb-system/controller-f8648f98b-46vm8" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.323682 4802 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.329230 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3a078497-d2a1-4be5-87ee-ba151c992ecd-metrics-certs\") pod \"controller-f8648f98b-46vm8\" (UID: \"3a078497-d2a1-4be5-87ee-ba151c992ecd\") " pod="metallb-system/controller-f8648f98b-46vm8" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.336346 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3a078497-d2a1-4be5-87ee-ba151c992ecd-cert\") pod \"controller-f8648f98b-46vm8\" (UID: \"3a078497-d2a1-4be5-87ee-ba151c992ecd\") " pod="metallb-system/controller-f8648f98b-46vm8" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.343007 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jk476\" (UniqueName: \"kubernetes.io/projected/3a078497-d2a1-4be5-87ee-ba151c992ecd-kube-api-access-jk476\") pod \"controller-f8648f98b-46vm8\" (UID: \"3a078497-d2a1-4be5-87ee-ba151c992ecd\") " pod="metallb-system/controller-f8648f98b-46vm8" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.469688 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-46vm8" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.625605 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9ff1fa79-a414-498e-9f64-659502eb6ae7-metrics-certs\") pod \"frr-k8s-nzpzt\" (UID: \"9ff1fa79-a414-498e-9f64-659502eb6ae7\") " pod="metallb-system/frr-k8s-nzpzt" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.625865 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/94e7ed3e-c72c-4130-81e6-ef954e2e62dd-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-bjbtx\" (UID: \"94e7ed3e-c72c-4130-81e6-ef954e2e62dd\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-bjbtx" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.634839 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9ff1fa79-a414-498e-9f64-659502eb6ae7-metrics-certs\") pod \"frr-k8s-nzpzt\" (UID: \"9ff1fa79-a414-498e-9f64-659502eb6ae7\") " pod="metallb-system/frr-k8s-nzpzt" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.635250 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/94e7ed3e-c72c-4130-81e6-ef954e2e62dd-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-bjbtx\" (UID: \"94e7ed3e-c72c-4130-81e6-ef954e2e62dd\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-bjbtx" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.728612 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/1e43b1ba-4d5f-47a4-8f9c-359875710a1e-memberlist\") pod \"speaker-rlwm5\" (UID: \"1e43b1ba-4d5f-47a4-8f9c-359875710a1e\") " pod="metallb-system/speaker-rlwm5" Dec 06 03:53:13 crc kubenswrapper[4802]: E1206 03:53:13.729036 4802 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 06 03:53:13 crc kubenswrapper[4802]: E1206 03:53:13.729130 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1e43b1ba-4d5f-47a4-8f9c-359875710a1e-memberlist podName:1e43b1ba-4d5f-47a4-8f9c-359875710a1e nodeName:}" failed. No retries permitted until 2025-12-06 03:53:14.729107794 +0000 UTC m=+787.601016946 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/1e43b1ba-4d5f-47a4-8f9c-359875710a1e-memberlist") pod "speaker-rlwm5" (UID: "1e43b1ba-4d5f-47a4-8f9c-359875710a1e") : secret "metallb-memberlist" not found Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.734038 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-46vm8"] Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.916529 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-nzpzt" Dec 06 03:53:13 crc kubenswrapper[4802]: I1206 03:53:13.926458 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-bjbtx" Dec 06 03:53:14 crc kubenswrapper[4802]: I1206 03:53:14.247534 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-46vm8" event={"ID":"3a078497-d2a1-4be5-87ee-ba151c992ecd","Type":"ContainerStarted","Data":"3906dcc1f76887bf87e8b46bd04db3b8aa8a6a304a7b4248b4ab8f73a2a49b57"} Dec 06 03:53:14 crc kubenswrapper[4802]: I1206 03:53:14.247890 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-46vm8" event={"ID":"3a078497-d2a1-4be5-87ee-ba151c992ecd","Type":"ContainerStarted","Data":"5424880d3da9ec80ca6c23cfa134bd3b6c92ce94623130e522086c22d212ead7"} Dec 06 03:53:14 crc kubenswrapper[4802]: I1206 03:53:14.247918 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-46vm8" event={"ID":"3a078497-d2a1-4be5-87ee-ba151c992ecd","Type":"ContainerStarted","Data":"f0a6564a73fc7eb8b99b0353d98b2367d62b235069df15030630a4d0c6e99f58"} Dec 06 03:53:14 crc kubenswrapper[4802]: I1206 03:53:14.248171 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-46vm8" Dec 06 03:53:14 crc kubenswrapper[4802]: I1206 03:53:14.252278 4802 generic.go:334] "Generic (PLEG): container finished" podID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerID="57ad81d5a8f13aaa4dc17e3a8d71a6ae21755975dbb001e2d8346604e492cf3d" exitCode=0 Dec 06 03:53:14 crc kubenswrapper[4802]: I1206 03:53:14.252353 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" event={"ID":"6df38316-e0d3-4018-8d27-3620eba3a68d","Type":"ContainerDied","Data":"57ad81d5a8f13aaa4dc17e3a8d71a6ae21755975dbb001e2d8346604e492cf3d"} Dec 06 03:53:14 crc kubenswrapper[4802]: I1206 03:53:14.252392 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" event={"ID":"6df38316-e0d3-4018-8d27-3620eba3a68d","Type":"ContainerStarted","Data":"e13987c26ea7eba2b08231bba48c1e97a8d8b6df1eb4749d8e02e423ceb644c9"} Dec 06 03:53:14 crc kubenswrapper[4802]: I1206 03:53:14.252412 4802 scope.go:117] "RemoveContainer" containerID="25f142fb563466ee1de9442e9c6cd18bf7c0126b2d7681419cdb7a7267f4a6e6" Dec 06 03:53:14 crc kubenswrapper[4802]: I1206 03:53:14.253971 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-nzpzt" event={"ID":"9ff1fa79-a414-498e-9f64-659502eb6ae7","Type":"ContainerStarted","Data":"26fdcd66d1e040192b26faffab3f608de58d0bab9cd9adddff9fbbfeec851dba"} Dec 06 03:53:14 crc kubenswrapper[4802]: I1206 03:53:14.293958 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-46vm8" podStartSLOduration=1.293933662 podStartE2EDuration="1.293933662s" podCreationTimestamp="2025-12-06 03:53:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:53:14.265998029 +0000 UTC m=+787.137907211" watchObservedRunningTime="2025-12-06 03:53:14.293933662 +0000 UTC m=+787.165842834" Dec 06 03:53:14 crc kubenswrapper[4802]: I1206 03:53:14.348371 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-bjbtx"] Dec 06 03:53:14 crc kubenswrapper[4802]: W1206 03:53:14.356224 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod94e7ed3e_c72c_4130_81e6_ef954e2e62dd.slice/crio-17363b569d6c8c555086ecb7529b0bb3dc3590ae398ec0a108a2d94131726dd7 WatchSource:0}: Error finding container 17363b569d6c8c555086ecb7529b0bb3dc3590ae398ec0a108a2d94131726dd7: Status 404 returned error can't find the container with id 17363b569d6c8c555086ecb7529b0bb3dc3590ae398ec0a108a2d94131726dd7 Dec 06 03:53:14 crc kubenswrapper[4802]: I1206 03:53:14.745455 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/1e43b1ba-4d5f-47a4-8f9c-359875710a1e-memberlist\") pod \"speaker-rlwm5\" (UID: \"1e43b1ba-4d5f-47a4-8f9c-359875710a1e\") " pod="metallb-system/speaker-rlwm5" Dec 06 03:53:14 crc kubenswrapper[4802]: I1206 03:53:14.750379 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/1e43b1ba-4d5f-47a4-8f9c-359875710a1e-memberlist\") pod \"speaker-rlwm5\" (UID: \"1e43b1ba-4d5f-47a4-8f9c-359875710a1e\") " pod="metallb-system/speaker-rlwm5" Dec 06 03:53:14 crc kubenswrapper[4802]: I1206 03:53:14.902364 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-rlwm5" Dec 06 03:53:14 crc kubenswrapper[4802]: W1206 03:53:14.923853 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1e43b1ba_4d5f_47a4_8f9c_359875710a1e.slice/crio-eeb5f1a35dc25542e46833ed401e3e68478a1fdb02d7c031207311df6f05037b WatchSource:0}: Error finding container eeb5f1a35dc25542e46833ed401e3e68478a1fdb02d7c031207311df6f05037b: Status 404 returned error can't find the container with id eeb5f1a35dc25542e46833ed401e3e68478a1fdb02d7c031207311df6f05037b Dec 06 03:53:15 crc kubenswrapper[4802]: I1206 03:53:15.278046 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-rlwm5" event={"ID":"1e43b1ba-4d5f-47a4-8f9c-359875710a1e","Type":"ContainerStarted","Data":"afad15125e5d219a68ec614450e709a7396f2f20e9e13ea9e93819657627735b"} Dec 06 03:53:15 crc kubenswrapper[4802]: I1206 03:53:15.278314 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-rlwm5" event={"ID":"1e43b1ba-4d5f-47a4-8f9c-359875710a1e","Type":"ContainerStarted","Data":"eeb5f1a35dc25542e46833ed401e3e68478a1fdb02d7c031207311df6f05037b"} Dec 06 03:53:15 crc kubenswrapper[4802]: I1206 03:53:15.279943 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-bjbtx" event={"ID":"94e7ed3e-c72c-4130-81e6-ef954e2e62dd","Type":"ContainerStarted","Data":"17363b569d6c8c555086ecb7529b0bb3dc3590ae398ec0a108a2d94131726dd7"} Dec 06 03:53:16 crc kubenswrapper[4802]: I1206 03:53:16.288045 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-rlwm5" event={"ID":"1e43b1ba-4d5f-47a4-8f9c-359875710a1e","Type":"ContainerStarted","Data":"d2196397348a1bf4a90abdf2f3213a090089038513738b5e06d3a2361180b7b3"} Dec 06 03:53:16 crc kubenswrapper[4802]: I1206 03:53:16.288436 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-rlwm5" Dec 06 03:53:16 crc kubenswrapper[4802]: I1206 03:53:16.303826 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-rlwm5" podStartSLOduration=3.303804735 podStartE2EDuration="3.303804735s" podCreationTimestamp="2025-12-06 03:53:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:53:16.303531468 +0000 UTC m=+789.175440640" watchObservedRunningTime="2025-12-06 03:53:16.303804735 +0000 UTC m=+789.175713887" Dec 06 03:53:21 crc kubenswrapper[4802]: I1206 03:53:21.318306 4802 generic.go:334] "Generic (PLEG): container finished" podID="9ff1fa79-a414-498e-9f64-659502eb6ae7" containerID="870e41d47d7f8c32062681b51ff249eb9f05b20ce2872f0bbe2b5448d58f18e5" exitCode=0 Dec 06 03:53:21 crc kubenswrapper[4802]: I1206 03:53:21.318357 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-nzpzt" event={"ID":"9ff1fa79-a414-498e-9f64-659502eb6ae7","Type":"ContainerDied","Data":"870e41d47d7f8c32062681b51ff249eb9f05b20ce2872f0bbe2b5448d58f18e5"} Dec 06 03:53:21 crc kubenswrapper[4802]: I1206 03:53:21.321265 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-bjbtx" event={"ID":"94e7ed3e-c72c-4130-81e6-ef954e2e62dd","Type":"ContainerStarted","Data":"1783dbe723dfa192902586d81e926e588ee3d9694ec77261b067125a99ea5e67"} Dec 06 03:53:21 crc kubenswrapper[4802]: I1206 03:53:21.321467 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-bjbtx" Dec 06 03:53:21 crc kubenswrapper[4802]: I1206 03:53:21.366442 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-bjbtx" podStartSLOduration=3.480773912 podStartE2EDuration="9.366414618s" podCreationTimestamp="2025-12-06 03:53:12 +0000 UTC" firstStartedPulling="2025-12-06 03:53:14.358858487 +0000 UTC m=+787.230767659" lastFinishedPulling="2025-12-06 03:53:20.244499203 +0000 UTC m=+793.116408365" observedRunningTime="2025-12-06 03:53:21.35883676 +0000 UTC m=+794.230745942" watchObservedRunningTime="2025-12-06 03:53:21.366414618 +0000 UTC m=+794.238323810" Dec 06 03:53:22 crc kubenswrapper[4802]: I1206 03:53:22.331472 4802 generic.go:334] "Generic (PLEG): container finished" podID="9ff1fa79-a414-498e-9f64-659502eb6ae7" containerID="3e52b7e38c55a5c4941adf4e87eb1aeaece9c0bc59162eee67826ae7e0943806" exitCode=0 Dec 06 03:53:22 crc kubenswrapper[4802]: I1206 03:53:22.331539 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-nzpzt" event={"ID":"9ff1fa79-a414-498e-9f64-659502eb6ae7","Type":"ContainerDied","Data":"3e52b7e38c55a5c4941adf4e87eb1aeaece9c0bc59162eee67826ae7e0943806"} Dec 06 03:53:23 crc kubenswrapper[4802]: I1206 03:53:23.341990 4802 generic.go:334] "Generic (PLEG): container finished" podID="9ff1fa79-a414-498e-9f64-659502eb6ae7" containerID="34c77b221b1eb7cb9c6ef3df71cbeb0d973fda7667ee0815e0a2dc25ec778231" exitCode=0 Dec 06 03:53:23 crc kubenswrapper[4802]: I1206 03:53:23.342058 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-nzpzt" event={"ID":"9ff1fa79-a414-498e-9f64-659502eb6ae7","Type":"ContainerDied","Data":"34c77b221b1eb7cb9c6ef3df71cbeb0d973fda7667ee0815e0a2dc25ec778231"} Dec 06 03:53:23 crc kubenswrapper[4802]: I1206 03:53:23.476486 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-46vm8" Dec 06 03:53:24 crc kubenswrapper[4802]: I1206 03:53:24.351890 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-nzpzt" event={"ID":"9ff1fa79-a414-498e-9f64-659502eb6ae7","Type":"ContainerStarted","Data":"3958c2c03042d83b55c2d043b67070931548d1ecddf0a8ea6b2a3379c0d5b63b"} Dec 06 03:53:24 crc kubenswrapper[4802]: I1206 03:53:24.352200 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-nzpzt" event={"ID":"9ff1fa79-a414-498e-9f64-659502eb6ae7","Type":"ContainerStarted","Data":"824a5220c30d145c3a3e908025037df919380013e23c10320296d07d9c1be6f5"} Dec 06 03:53:24 crc kubenswrapper[4802]: I1206 03:53:24.352211 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-nzpzt" event={"ID":"9ff1fa79-a414-498e-9f64-659502eb6ae7","Type":"ContainerStarted","Data":"b92ed1e3b9ccd2667f21522780f161d15cb6d33d231b15f4dac5e3916dff3678"} Dec 06 03:53:24 crc kubenswrapper[4802]: I1206 03:53:24.352219 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-nzpzt" event={"ID":"9ff1fa79-a414-498e-9f64-659502eb6ae7","Type":"ContainerStarted","Data":"68de6a02eee1e021d771bd4941199a29262e40fe3da954658d5fe3b4d23678f6"} Dec 06 03:53:24 crc kubenswrapper[4802]: I1206 03:53:24.352228 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-nzpzt" event={"ID":"9ff1fa79-a414-498e-9f64-659502eb6ae7","Type":"ContainerStarted","Data":"3b0024e95e8746a5ce1f8809dacc803d276a97b0fd5193c2b8daba3323ca4691"} Dec 06 03:53:25 crc kubenswrapper[4802]: I1206 03:53:25.364069 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-nzpzt" event={"ID":"9ff1fa79-a414-498e-9f64-659502eb6ae7","Type":"ContainerStarted","Data":"0096b9a7277bb850dce3c0ac7e6071433ae0febffaf64687d0f8d448ba49be48"} Dec 06 03:53:25 crc kubenswrapper[4802]: I1206 03:53:25.364316 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-nzpzt" Dec 06 03:53:25 crc kubenswrapper[4802]: I1206 03:53:25.393634 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-nzpzt" podStartSLOduration=7.186155307 podStartE2EDuration="13.393598948s" podCreationTimestamp="2025-12-06 03:53:12 +0000 UTC" firstStartedPulling="2025-12-06 03:53:14.017174398 +0000 UTC m=+786.889083550" lastFinishedPulling="2025-12-06 03:53:20.224618039 +0000 UTC m=+793.096527191" observedRunningTime="2025-12-06 03:53:25.38923854 +0000 UTC m=+798.261147682" watchObservedRunningTime="2025-12-06 03:53:25.393598948 +0000 UTC m=+798.265508160" Dec 06 03:53:28 crc kubenswrapper[4802]: I1206 03:53:28.917558 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-nzpzt" Dec 06 03:53:28 crc kubenswrapper[4802]: I1206 03:53:28.991823 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-nzpzt" Dec 06 03:53:33 crc kubenswrapper[4802]: I1206 03:53:33.920956 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-nzpzt" Dec 06 03:53:33 crc kubenswrapper[4802]: I1206 03:53:33.936318 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-bjbtx" Dec 06 03:53:34 crc kubenswrapper[4802]: I1206 03:53:34.908581 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-rlwm5" Dec 06 03:53:38 crc kubenswrapper[4802]: I1206 03:53:38.417515 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-dzkk2"] Dec 06 03:53:38 crc kubenswrapper[4802]: I1206 03:53:38.418662 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-dzkk2" Dec 06 03:53:38 crc kubenswrapper[4802]: I1206 03:53:38.421162 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 06 03:53:38 crc kubenswrapper[4802]: I1206 03:53:38.424177 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-87fw6" Dec 06 03:53:38 crc kubenswrapper[4802]: I1206 03:53:38.432813 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-dzkk2"] Dec 06 03:53:38 crc kubenswrapper[4802]: I1206 03:53:38.433771 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 06 03:53:38 crc kubenswrapper[4802]: I1206 03:53:38.584589 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7g77\" (UniqueName: \"kubernetes.io/projected/e853c3cf-445f-4210-bedb-dfd77252770f-kube-api-access-r7g77\") pod \"openstack-operator-index-dzkk2\" (UID: \"e853c3cf-445f-4210-bedb-dfd77252770f\") " pod="openstack-operators/openstack-operator-index-dzkk2" Dec 06 03:53:38 crc kubenswrapper[4802]: I1206 03:53:38.691709 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7g77\" (UniqueName: \"kubernetes.io/projected/e853c3cf-445f-4210-bedb-dfd77252770f-kube-api-access-r7g77\") pod \"openstack-operator-index-dzkk2\" (UID: \"e853c3cf-445f-4210-bedb-dfd77252770f\") " pod="openstack-operators/openstack-operator-index-dzkk2" Dec 06 03:53:38 crc kubenswrapper[4802]: I1206 03:53:38.731386 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7g77\" (UniqueName: \"kubernetes.io/projected/e853c3cf-445f-4210-bedb-dfd77252770f-kube-api-access-r7g77\") pod \"openstack-operator-index-dzkk2\" (UID: \"e853c3cf-445f-4210-bedb-dfd77252770f\") " pod="openstack-operators/openstack-operator-index-dzkk2" Dec 06 03:53:38 crc kubenswrapper[4802]: I1206 03:53:38.743390 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-dzkk2" Dec 06 03:53:39 crc kubenswrapper[4802]: W1206 03:53:39.236858 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode853c3cf_445f_4210_bedb_dfd77252770f.slice/crio-ddfa1b5c369f1b57dcb09e7bf3bd8a301258d8d418212a4ea8081142e72152e5 WatchSource:0}: Error finding container ddfa1b5c369f1b57dcb09e7bf3bd8a301258d8d418212a4ea8081142e72152e5: Status 404 returned error can't find the container with id ddfa1b5c369f1b57dcb09e7bf3bd8a301258d8d418212a4ea8081142e72152e5 Dec 06 03:53:39 crc kubenswrapper[4802]: I1206 03:53:39.242942 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-dzkk2"] Dec 06 03:53:39 crc kubenswrapper[4802]: I1206 03:53:39.474058 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-dzkk2" event={"ID":"e853c3cf-445f-4210-bedb-dfd77252770f","Type":"ContainerStarted","Data":"ddfa1b5c369f1b57dcb09e7bf3bd8a301258d8d418212a4ea8081142e72152e5"} Dec 06 03:53:42 crc kubenswrapper[4802]: I1206 03:53:42.385143 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-dzkk2"] Dec 06 03:53:42 crc kubenswrapper[4802]: I1206 03:53:42.990343 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-xd8sc"] Dec 06 03:53:42 crc kubenswrapper[4802]: I1206 03:53:42.991135 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-xd8sc" Dec 06 03:53:43 crc kubenswrapper[4802]: I1206 03:53:43.006205 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-xd8sc"] Dec 06 03:53:43 crc kubenswrapper[4802]: I1206 03:53:43.145598 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x258f\" (UniqueName: \"kubernetes.io/projected/43e371ec-a36d-4082-827e-21518dc7e9c1-kube-api-access-x258f\") pod \"openstack-operator-index-xd8sc\" (UID: \"43e371ec-a36d-4082-827e-21518dc7e9c1\") " pod="openstack-operators/openstack-operator-index-xd8sc" Dec 06 03:53:43 crc kubenswrapper[4802]: I1206 03:53:43.273282 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x258f\" (UniqueName: \"kubernetes.io/projected/43e371ec-a36d-4082-827e-21518dc7e9c1-kube-api-access-x258f\") pod \"openstack-operator-index-xd8sc\" (UID: \"43e371ec-a36d-4082-827e-21518dc7e9c1\") " pod="openstack-operators/openstack-operator-index-xd8sc" Dec 06 03:53:43 crc kubenswrapper[4802]: I1206 03:53:43.302713 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x258f\" (UniqueName: \"kubernetes.io/projected/43e371ec-a36d-4082-827e-21518dc7e9c1-kube-api-access-x258f\") pod \"openstack-operator-index-xd8sc\" (UID: \"43e371ec-a36d-4082-827e-21518dc7e9c1\") " pod="openstack-operators/openstack-operator-index-xd8sc" Dec 06 03:53:43 crc kubenswrapper[4802]: I1206 03:53:43.316287 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-xd8sc" Dec 06 03:53:43 crc kubenswrapper[4802]: I1206 03:53:43.502977 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-dzkk2" event={"ID":"e853c3cf-445f-4210-bedb-dfd77252770f","Type":"ContainerStarted","Data":"7d498ed3bf9991f9ea5c7b95f57095e3c804abdfa991e066a12b94c9614d043c"} Dec 06 03:53:43 crc kubenswrapper[4802]: I1206 03:53:43.503095 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-dzkk2" podUID="e853c3cf-445f-4210-bedb-dfd77252770f" containerName="registry-server" containerID="cri-o://7d498ed3bf9991f9ea5c7b95f57095e3c804abdfa991e066a12b94c9614d043c" gracePeriod=2 Dec 06 03:53:43 crc kubenswrapper[4802]: I1206 03:53:43.521242 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-dzkk2" podStartSLOduration=1.4719533390000001 podStartE2EDuration="5.521229255s" podCreationTimestamp="2025-12-06 03:53:38 +0000 UTC" firstStartedPulling="2025-12-06 03:53:39.238948281 +0000 UTC m=+812.110857433" lastFinishedPulling="2025-12-06 03:53:43.288224197 +0000 UTC m=+816.160133349" observedRunningTime="2025-12-06 03:53:43.518891561 +0000 UTC m=+816.390800703" watchObservedRunningTime="2025-12-06 03:53:43.521229255 +0000 UTC m=+816.393138407" Dec 06 03:53:43 crc kubenswrapper[4802]: I1206 03:53:43.796115 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-xd8sc"] Dec 06 03:53:43 crc kubenswrapper[4802]: I1206 03:53:43.877183 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-dzkk2_e853c3cf-445f-4210-bedb-dfd77252770f/registry-server/0.log" Dec 06 03:53:43 crc kubenswrapper[4802]: I1206 03:53:43.877279 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-dzkk2" Dec 06 03:53:43 crc kubenswrapper[4802]: I1206 03:53:43.983140 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r7g77\" (UniqueName: \"kubernetes.io/projected/e853c3cf-445f-4210-bedb-dfd77252770f-kube-api-access-r7g77\") pod \"e853c3cf-445f-4210-bedb-dfd77252770f\" (UID: \"e853c3cf-445f-4210-bedb-dfd77252770f\") " Dec 06 03:53:43 crc kubenswrapper[4802]: I1206 03:53:43.988482 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e853c3cf-445f-4210-bedb-dfd77252770f-kube-api-access-r7g77" (OuterVolumeSpecName: "kube-api-access-r7g77") pod "e853c3cf-445f-4210-bedb-dfd77252770f" (UID: "e853c3cf-445f-4210-bedb-dfd77252770f"). InnerVolumeSpecName "kube-api-access-r7g77". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:53:44 crc kubenswrapper[4802]: I1206 03:53:44.084832 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r7g77\" (UniqueName: \"kubernetes.io/projected/e853c3cf-445f-4210-bedb-dfd77252770f-kube-api-access-r7g77\") on node \"crc\" DevicePath \"\"" Dec 06 03:53:44 crc kubenswrapper[4802]: I1206 03:53:44.510106 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-xd8sc" event={"ID":"43e371ec-a36d-4082-827e-21518dc7e9c1","Type":"ContainerStarted","Data":"868426a2f7950a3ab9920130a13be0f756fa3d958415bcb3d0dca999dfc46a41"} Dec 06 03:53:44 crc kubenswrapper[4802]: I1206 03:53:44.510180 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-xd8sc" event={"ID":"43e371ec-a36d-4082-827e-21518dc7e9c1","Type":"ContainerStarted","Data":"071d482ce968e3e336a75ac27c7762dbab581f752cad785956e5049a62c08088"} Dec 06 03:53:44 crc kubenswrapper[4802]: I1206 03:53:44.512123 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-dzkk2_e853c3cf-445f-4210-bedb-dfd77252770f/registry-server/0.log" Dec 06 03:53:44 crc kubenswrapper[4802]: I1206 03:53:44.512165 4802 generic.go:334] "Generic (PLEG): container finished" podID="e853c3cf-445f-4210-bedb-dfd77252770f" containerID="7d498ed3bf9991f9ea5c7b95f57095e3c804abdfa991e066a12b94c9614d043c" exitCode=2 Dec 06 03:53:44 crc kubenswrapper[4802]: I1206 03:53:44.512186 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-dzkk2" event={"ID":"e853c3cf-445f-4210-bedb-dfd77252770f","Type":"ContainerDied","Data":"7d498ed3bf9991f9ea5c7b95f57095e3c804abdfa991e066a12b94c9614d043c"} Dec 06 03:53:44 crc kubenswrapper[4802]: I1206 03:53:44.512201 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-dzkk2" event={"ID":"e853c3cf-445f-4210-bedb-dfd77252770f","Type":"ContainerDied","Data":"ddfa1b5c369f1b57dcb09e7bf3bd8a301258d8d418212a4ea8081142e72152e5"} Dec 06 03:53:44 crc kubenswrapper[4802]: I1206 03:53:44.512217 4802 scope.go:117] "RemoveContainer" containerID="7d498ed3bf9991f9ea5c7b95f57095e3c804abdfa991e066a12b94c9614d043c" Dec 06 03:53:44 crc kubenswrapper[4802]: I1206 03:53:44.512288 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-dzkk2" Dec 06 03:53:44 crc kubenswrapper[4802]: I1206 03:53:44.531533 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-xd8sc" podStartSLOduration=2.458663456 podStartE2EDuration="2.531517168s" podCreationTimestamp="2025-12-06 03:53:42 +0000 UTC" firstStartedPulling="2025-12-06 03:53:43.82144775 +0000 UTC m=+816.693356902" lastFinishedPulling="2025-12-06 03:53:43.894301442 +0000 UTC m=+816.766210614" observedRunningTime="2025-12-06 03:53:44.527744625 +0000 UTC m=+817.399669407" watchObservedRunningTime="2025-12-06 03:53:44.531517168 +0000 UTC m=+817.403426320" Dec 06 03:53:44 crc kubenswrapper[4802]: I1206 03:53:44.536325 4802 scope.go:117] "RemoveContainer" containerID="7d498ed3bf9991f9ea5c7b95f57095e3c804abdfa991e066a12b94c9614d043c" Dec 06 03:53:44 crc kubenswrapper[4802]: E1206 03:53:44.536724 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d498ed3bf9991f9ea5c7b95f57095e3c804abdfa991e066a12b94c9614d043c\": container with ID starting with 7d498ed3bf9991f9ea5c7b95f57095e3c804abdfa991e066a12b94c9614d043c not found: ID does not exist" containerID="7d498ed3bf9991f9ea5c7b95f57095e3c804abdfa991e066a12b94c9614d043c" Dec 06 03:53:44 crc kubenswrapper[4802]: I1206 03:53:44.536791 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d498ed3bf9991f9ea5c7b95f57095e3c804abdfa991e066a12b94c9614d043c"} err="failed to get container status \"7d498ed3bf9991f9ea5c7b95f57095e3c804abdfa991e066a12b94c9614d043c\": rpc error: code = NotFound desc = could not find container \"7d498ed3bf9991f9ea5c7b95f57095e3c804abdfa991e066a12b94c9614d043c\": container with ID starting with 7d498ed3bf9991f9ea5c7b95f57095e3c804abdfa991e066a12b94c9614d043c not found: ID does not exist" Dec 06 03:53:44 crc kubenswrapper[4802]: I1206 03:53:44.555892 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-dzkk2"] Dec 06 03:53:44 crc kubenswrapper[4802]: I1206 03:53:44.562939 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-dzkk2"] Dec 06 03:53:45 crc kubenswrapper[4802]: I1206 03:53:45.459910 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e853c3cf-445f-4210-bedb-dfd77252770f" path="/var/lib/kubelet/pods/e853c3cf-445f-4210-bedb-dfd77252770f/volumes" Dec 06 03:53:53 crc kubenswrapper[4802]: I1206 03:53:53.316809 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-xd8sc" Dec 06 03:53:53 crc kubenswrapper[4802]: I1206 03:53:53.317303 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-xd8sc" Dec 06 03:53:53 crc kubenswrapper[4802]: I1206 03:53:53.352089 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-xd8sc" Dec 06 03:53:53 crc kubenswrapper[4802]: I1206 03:53:53.598572 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-xd8sc" Dec 06 03:53:59 crc kubenswrapper[4802]: I1206 03:53:59.988106 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/11a37daaca2b5d044ad5e897ef78a9b78d5fd0aec8475f6e59b41a4af1nhddl"] Dec 06 03:53:59 crc kubenswrapper[4802]: E1206 03:53:59.988957 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e853c3cf-445f-4210-bedb-dfd77252770f" containerName="registry-server" Dec 06 03:53:59 crc kubenswrapper[4802]: I1206 03:53:59.988982 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="e853c3cf-445f-4210-bedb-dfd77252770f" containerName="registry-server" Dec 06 03:53:59 crc kubenswrapper[4802]: I1206 03:53:59.989211 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="e853c3cf-445f-4210-bedb-dfd77252770f" containerName="registry-server" Dec 06 03:53:59 crc kubenswrapper[4802]: I1206 03:53:59.990779 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/11a37daaca2b5d044ad5e897ef78a9b78d5fd0aec8475f6e59b41a4af1nhddl" Dec 06 03:53:59 crc kubenswrapper[4802]: I1206 03:53:59.993239 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-xqkj4" Dec 06 03:54:00 crc kubenswrapper[4802]: I1206 03:54:00.006722 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/11a37daaca2b5d044ad5e897ef78a9b78d5fd0aec8475f6e59b41a4af1nhddl"] Dec 06 03:54:00 crc kubenswrapper[4802]: I1206 03:54:00.008806 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgxnh\" (UniqueName: \"kubernetes.io/projected/e9a7628c-dba2-482e-b66b-e4046bb06ddc-kube-api-access-qgxnh\") pod \"11a37daaca2b5d044ad5e897ef78a9b78d5fd0aec8475f6e59b41a4af1nhddl\" (UID: \"e9a7628c-dba2-482e-b66b-e4046bb06ddc\") " pod="openstack-operators/11a37daaca2b5d044ad5e897ef78a9b78d5fd0aec8475f6e59b41a4af1nhddl" Dec 06 03:54:00 crc kubenswrapper[4802]: I1206 03:54:00.009027 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e9a7628c-dba2-482e-b66b-e4046bb06ddc-bundle\") pod \"11a37daaca2b5d044ad5e897ef78a9b78d5fd0aec8475f6e59b41a4af1nhddl\" (UID: \"e9a7628c-dba2-482e-b66b-e4046bb06ddc\") " pod="openstack-operators/11a37daaca2b5d044ad5e897ef78a9b78d5fd0aec8475f6e59b41a4af1nhddl" Dec 06 03:54:00 crc kubenswrapper[4802]: I1206 03:54:00.009083 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e9a7628c-dba2-482e-b66b-e4046bb06ddc-util\") pod \"11a37daaca2b5d044ad5e897ef78a9b78d5fd0aec8475f6e59b41a4af1nhddl\" (UID: \"e9a7628c-dba2-482e-b66b-e4046bb06ddc\") " pod="openstack-operators/11a37daaca2b5d044ad5e897ef78a9b78d5fd0aec8475f6e59b41a4af1nhddl" Dec 06 03:54:00 crc kubenswrapper[4802]: I1206 03:54:00.109608 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e9a7628c-dba2-482e-b66b-e4046bb06ddc-bundle\") pod \"11a37daaca2b5d044ad5e897ef78a9b78d5fd0aec8475f6e59b41a4af1nhddl\" (UID: \"e9a7628c-dba2-482e-b66b-e4046bb06ddc\") " pod="openstack-operators/11a37daaca2b5d044ad5e897ef78a9b78d5fd0aec8475f6e59b41a4af1nhddl" Dec 06 03:54:00 crc kubenswrapper[4802]: I1206 03:54:00.109685 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e9a7628c-dba2-482e-b66b-e4046bb06ddc-util\") pod \"11a37daaca2b5d044ad5e897ef78a9b78d5fd0aec8475f6e59b41a4af1nhddl\" (UID: \"e9a7628c-dba2-482e-b66b-e4046bb06ddc\") " pod="openstack-operators/11a37daaca2b5d044ad5e897ef78a9b78d5fd0aec8475f6e59b41a4af1nhddl" Dec 06 03:54:00 crc kubenswrapper[4802]: I1206 03:54:00.109791 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgxnh\" (UniqueName: \"kubernetes.io/projected/e9a7628c-dba2-482e-b66b-e4046bb06ddc-kube-api-access-qgxnh\") pod \"11a37daaca2b5d044ad5e897ef78a9b78d5fd0aec8475f6e59b41a4af1nhddl\" (UID: \"e9a7628c-dba2-482e-b66b-e4046bb06ddc\") " pod="openstack-operators/11a37daaca2b5d044ad5e897ef78a9b78d5fd0aec8475f6e59b41a4af1nhddl" Dec 06 03:54:00 crc kubenswrapper[4802]: I1206 03:54:00.110829 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e9a7628c-dba2-482e-b66b-e4046bb06ddc-bundle\") pod \"11a37daaca2b5d044ad5e897ef78a9b78d5fd0aec8475f6e59b41a4af1nhddl\" (UID: \"e9a7628c-dba2-482e-b66b-e4046bb06ddc\") " pod="openstack-operators/11a37daaca2b5d044ad5e897ef78a9b78d5fd0aec8475f6e59b41a4af1nhddl" Dec 06 03:54:00 crc kubenswrapper[4802]: I1206 03:54:00.110870 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e9a7628c-dba2-482e-b66b-e4046bb06ddc-util\") pod \"11a37daaca2b5d044ad5e897ef78a9b78d5fd0aec8475f6e59b41a4af1nhddl\" (UID: \"e9a7628c-dba2-482e-b66b-e4046bb06ddc\") " pod="openstack-operators/11a37daaca2b5d044ad5e897ef78a9b78d5fd0aec8475f6e59b41a4af1nhddl" Dec 06 03:54:00 crc kubenswrapper[4802]: I1206 03:54:00.133626 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgxnh\" (UniqueName: \"kubernetes.io/projected/e9a7628c-dba2-482e-b66b-e4046bb06ddc-kube-api-access-qgxnh\") pod \"11a37daaca2b5d044ad5e897ef78a9b78d5fd0aec8475f6e59b41a4af1nhddl\" (UID: \"e9a7628c-dba2-482e-b66b-e4046bb06ddc\") " pod="openstack-operators/11a37daaca2b5d044ad5e897ef78a9b78d5fd0aec8475f6e59b41a4af1nhddl" Dec 06 03:54:00 crc kubenswrapper[4802]: I1206 03:54:00.363282 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/11a37daaca2b5d044ad5e897ef78a9b78d5fd0aec8475f6e59b41a4af1nhddl" Dec 06 03:54:00 crc kubenswrapper[4802]: I1206 03:54:00.634878 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/11a37daaca2b5d044ad5e897ef78a9b78d5fd0aec8475f6e59b41a4af1nhddl"] Dec 06 03:54:00 crc kubenswrapper[4802]: W1206 03:54:00.642777 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode9a7628c_dba2_482e_b66b_e4046bb06ddc.slice/crio-03292b341d465df41eb4003a4fa480e6042e86bfa140dfd8e959b4bc27f2528f WatchSource:0}: Error finding container 03292b341d465df41eb4003a4fa480e6042e86bfa140dfd8e959b4bc27f2528f: Status 404 returned error can't find the container with id 03292b341d465df41eb4003a4fa480e6042e86bfa140dfd8e959b4bc27f2528f Dec 06 03:54:01 crc kubenswrapper[4802]: I1206 03:54:01.621464 4802 generic.go:334] "Generic (PLEG): container finished" podID="e9a7628c-dba2-482e-b66b-e4046bb06ddc" containerID="099d6d73f9bb556e18464da2383e39f59b960308b3a8aa97f2364e7318257509" exitCode=0 Dec 06 03:54:01 crc kubenswrapper[4802]: I1206 03:54:01.621526 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/11a37daaca2b5d044ad5e897ef78a9b78d5fd0aec8475f6e59b41a4af1nhddl" event={"ID":"e9a7628c-dba2-482e-b66b-e4046bb06ddc","Type":"ContainerDied","Data":"099d6d73f9bb556e18464da2383e39f59b960308b3a8aa97f2364e7318257509"} Dec 06 03:54:01 crc kubenswrapper[4802]: I1206 03:54:01.621564 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/11a37daaca2b5d044ad5e897ef78a9b78d5fd0aec8475f6e59b41a4af1nhddl" event={"ID":"e9a7628c-dba2-482e-b66b-e4046bb06ddc","Type":"ContainerStarted","Data":"03292b341d465df41eb4003a4fa480e6042e86bfa140dfd8e959b4bc27f2528f"} Dec 06 03:54:02 crc kubenswrapper[4802]: I1206 03:54:02.630237 4802 generic.go:334] "Generic (PLEG): container finished" podID="e9a7628c-dba2-482e-b66b-e4046bb06ddc" containerID="3ae39f54e17d482374fd9764c8f84751add02e3388decd92b6d006f38ea10031" exitCode=0 Dec 06 03:54:02 crc kubenswrapper[4802]: I1206 03:54:02.630362 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/11a37daaca2b5d044ad5e897ef78a9b78d5fd0aec8475f6e59b41a4af1nhddl" event={"ID":"e9a7628c-dba2-482e-b66b-e4046bb06ddc","Type":"ContainerDied","Data":"3ae39f54e17d482374fd9764c8f84751add02e3388decd92b6d006f38ea10031"} Dec 06 03:54:03 crc kubenswrapper[4802]: I1206 03:54:03.643592 4802 generic.go:334] "Generic (PLEG): container finished" podID="e9a7628c-dba2-482e-b66b-e4046bb06ddc" containerID="8f988d84f80b9d7587444fa776942d22b4ede39e4606ca32e8628d5f52002515" exitCode=0 Dec 06 03:54:03 crc kubenswrapper[4802]: I1206 03:54:03.643641 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/11a37daaca2b5d044ad5e897ef78a9b78d5fd0aec8475f6e59b41a4af1nhddl" event={"ID":"e9a7628c-dba2-482e-b66b-e4046bb06ddc","Type":"ContainerDied","Data":"8f988d84f80b9d7587444fa776942d22b4ede39e4606ca32e8628d5f52002515"} Dec 06 03:54:04 crc kubenswrapper[4802]: I1206 03:54:04.964658 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/11a37daaca2b5d044ad5e897ef78a9b78d5fd0aec8475f6e59b41a4af1nhddl" Dec 06 03:54:05 crc kubenswrapper[4802]: I1206 03:54:05.081168 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qgxnh\" (UniqueName: \"kubernetes.io/projected/e9a7628c-dba2-482e-b66b-e4046bb06ddc-kube-api-access-qgxnh\") pod \"e9a7628c-dba2-482e-b66b-e4046bb06ddc\" (UID: \"e9a7628c-dba2-482e-b66b-e4046bb06ddc\") " Dec 06 03:54:05 crc kubenswrapper[4802]: I1206 03:54:05.081262 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e9a7628c-dba2-482e-b66b-e4046bb06ddc-bundle\") pod \"e9a7628c-dba2-482e-b66b-e4046bb06ddc\" (UID: \"e9a7628c-dba2-482e-b66b-e4046bb06ddc\") " Dec 06 03:54:05 crc kubenswrapper[4802]: I1206 03:54:05.081338 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e9a7628c-dba2-482e-b66b-e4046bb06ddc-util\") pod \"e9a7628c-dba2-482e-b66b-e4046bb06ddc\" (UID: \"e9a7628c-dba2-482e-b66b-e4046bb06ddc\") " Dec 06 03:54:05 crc kubenswrapper[4802]: I1206 03:54:05.082447 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e9a7628c-dba2-482e-b66b-e4046bb06ddc-bundle" (OuterVolumeSpecName: "bundle") pod "e9a7628c-dba2-482e-b66b-e4046bb06ddc" (UID: "e9a7628c-dba2-482e-b66b-e4046bb06ddc"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:54:05 crc kubenswrapper[4802]: I1206 03:54:05.084314 4802 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e9a7628c-dba2-482e-b66b-e4046bb06ddc-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 03:54:05 crc kubenswrapper[4802]: I1206 03:54:05.087250 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9a7628c-dba2-482e-b66b-e4046bb06ddc-kube-api-access-qgxnh" (OuterVolumeSpecName: "kube-api-access-qgxnh") pod "e9a7628c-dba2-482e-b66b-e4046bb06ddc" (UID: "e9a7628c-dba2-482e-b66b-e4046bb06ddc"). InnerVolumeSpecName "kube-api-access-qgxnh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:54:05 crc kubenswrapper[4802]: I1206 03:54:05.113319 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e9a7628c-dba2-482e-b66b-e4046bb06ddc-util" (OuterVolumeSpecName: "util") pod "e9a7628c-dba2-482e-b66b-e4046bb06ddc" (UID: "e9a7628c-dba2-482e-b66b-e4046bb06ddc"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:54:05 crc kubenswrapper[4802]: I1206 03:54:05.185642 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qgxnh\" (UniqueName: \"kubernetes.io/projected/e9a7628c-dba2-482e-b66b-e4046bb06ddc-kube-api-access-qgxnh\") on node \"crc\" DevicePath \"\"" Dec 06 03:54:05 crc kubenswrapper[4802]: I1206 03:54:05.185682 4802 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e9a7628c-dba2-482e-b66b-e4046bb06ddc-util\") on node \"crc\" DevicePath \"\"" Dec 06 03:54:05 crc kubenswrapper[4802]: I1206 03:54:05.660031 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/11a37daaca2b5d044ad5e897ef78a9b78d5fd0aec8475f6e59b41a4af1nhddl" event={"ID":"e9a7628c-dba2-482e-b66b-e4046bb06ddc","Type":"ContainerDied","Data":"03292b341d465df41eb4003a4fa480e6042e86bfa140dfd8e959b4bc27f2528f"} Dec 06 03:54:05 crc kubenswrapper[4802]: I1206 03:54:05.660096 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="03292b341d465df41eb4003a4fa480e6042e86bfa140dfd8e959b4bc27f2528f" Dec 06 03:54:05 crc kubenswrapper[4802]: I1206 03:54:05.660140 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/11a37daaca2b5d044ad5e897ef78a9b78d5fd0aec8475f6e59b41a4af1nhddl" Dec 06 03:54:13 crc kubenswrapper[4802]: I1206 03:54:13.046254 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-57d4dc878b-qnwf9"] Dec 06 03:54:13 crc kubenswrapper[4802]: E1206 03:54:13.048073 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9a7628c-dba2-482e-b66b-e4046bb06ddc" containerName="pull" Dec 06 03:54:13 crc kubenswrapper[4802]: I1206 03:54:13.048186 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9a7628c-dba2-482e-b66b-e4046bb06ddc" containerName="pull" Dec 06 03:54:13 crc kubenswrapper[4802]: E1206 03:54:13.048276 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9a7628c-dba2-482e-b66b-e4046bb06ddc" containerName="util" Dec 06 03:54:13 crc kubenswrapper[4802]: I1206 03:54:13.048370 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9a7628c-dba2-482e-b66b-e4046bb06ddc" containerName="util" Dec 06 03:54:13 crc kubenswrapper[4802]: E1206 03:54:13.048481 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9a7628c-dba2-482e-b66b-e4046bb06ddc" containerName="extract" Dec 06 03:54:13 crc kubenswrapper[4802]: I1206 03:54:13.048602 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9a7628c-dba2-482e-b66b-e4046bb06ddc" containerName="extract" Dec 06 03:54:13 crc kubenswrapper[4802]: I1206 03:54:13.048874 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9a7628c-dba2-482e-b66b-e4046bb06ddc" containerName="extract" Dec 06 03:54:13 crc kubenswrapper[4802]: I1206 03:54:13.049611 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-57d4dc878b-qnwf9" Dec 06 03:54:13 crc kubenswrapper[4802]: I1206 03:54:13.054831 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-mq4lz" Dec 06 03:54:13 crc kubenswrapper[4802]: I1206 03:54:13.071554 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-57d4dc878b-qnwf9"] Dec 06 03:54:13 crc kubenswrapper[4802]: I1206 03:54:13.090451 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7wv2q\" (UniqueName: \"kubernetes.io/projected/31962197-d27b-41af-9469-27ae39fb5d4f-kube-api-access-7wv2q\") pod \"openstack-operator-controller-operator-57d4dc878b-qnwf9\" (UID: \"31962197-d27b-41af-9469-27ae39fb5d4f\") " pod="openstack-operators/openstack-operator-controller-operator-57d4dc878b-qnwf9" Dec 06 03:54:13 crc kubenswrapper[4802]: I1206 03:54:13.191447 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wv2q\" (UniqueName: \"kubernetes.io/projected/31962197-d27b-41af-9469-27ae39fb5d4f-kube-api-access-7wv2q\") pod \"openstack-operator-controller-operator-57d4dc878b-qnwf9\" (UID: \"31962197-d27b-41af-9469-27ae39fb5d4f\") " pod="openstack-operators/openstack-operator-controller-operator-57d4dc878b-qnwf9" Dec 06 03:54:13 crc kubenswrapper[4802]: I1206 03:54:13.209203 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wv2q\" (UniqueName: \"kubernetes.io/projected/31962197-d27b-41af-9469-27ae39fb5d4f-kube-api-access-7wv2q\") pod \"openstack-operator-controller-operator-57d4dc878b-qnwf9\" (UID: \"31962197-d27b-41af-9469-27ae39fb5d4f\") " pod="openstack-operators/openstack-operator-controller-operator-57d4dc878b-qnwf9" Dec 06 03:54:13 crc kubenswrapper[4802]: I1206 03:54:13.372357 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-57d4dc878b-qnwf9" Dec 06 03:54:13 crc kubenswrapper[4802]: I1206 03:54:13.775636 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-57d4dc878b-qnwf9"] Dec 06 03:54:14 crc kubenswrapper[4802]: I1206 03:54:14.715674 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-57d4dc878b-qnwf9" event={"ID":"31962197-d27b-41af-9469-27ae39fb5d4f","Type":"ContainerStarted","Data":"f4e416f2afb4cb49811ece86cbd8bee60bbc0d3c6a341d16b8e2eea3acd3c640"} Dec 06 03:54:22 crc kubenswrapper[4802]: I1206 03:54:22.793551 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-57d4dc878b-qnwf9" event={"ID":"31962197-d27b-41af-9469-27ae39fb5d4f","Type":"ContainerStarted","Data":"a6ad89f8e850906e26fbe467dd2aa076c18ccb4a3dca652f6cc7dc42cbb227df"} Dec 06 03:54:22 crc kubenswrapper[4802]: I1206 03:54:22.794017 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-57d4dc878b-qnwf9" Dec 06 03:54:22 crc kubenswrapper[4802]: I1206 03:54:22.828302 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-57d4dc878b-qnwf9" podStartSLOduration=1.828833075 podStartE2EDuration="9.828283504s" podCreationTimestamp="2025-12-06 03:54:13 +0000 UTC" firstStartedPulling="2025-12-06 03:54:13.788038415 +0000 UTC m=+846.659947567" lastFinishedPulling="2025-12-06 03:54:21.787488834 +0000 UTC m=+854.659397996" observedRunningTime="2025-12-06 03:54:22.823728601 +0000 UTC m=+855.695637783" watchObservedRunningTime="2025-12-06 03:54:22.828283504 +0000 UTC m=+855.700192656" Dec 06 03:54:33 crc kubenswrapper[4802]: I1206 03:54:33.376698 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-57d4dc878b-qnwf9" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.563835 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-p8kv6"] Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.566428 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-p8kv6" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.568955 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-kbdc4" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.570392 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-xk9cb"] Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.571671 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-xk9cb" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.574635 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-2nd4f" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.578611 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-xk9cb"] Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.585774 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-p8kv6"] Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.592001 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-wwgt6"] Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.592880 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-wwgt6" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.595973 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-skjlj" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.603450 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-wwgt6"] Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.612065 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-zkp5c"] Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.612989 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-zkp5c" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.615561 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-nk4hj" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.639128 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-zkp5c"] Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.651424 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-55cb9d4b9c-srntd"] Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.652960 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-55cb9d4b9c-srntd" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.658844 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2gdb\" (UniqueName: \"kubernetes.io/projected/e2bc662f-efe3-4e89-9b0e-e40eba87e41b-kube-api-access-v2gdb\") pod \"designate-operator-controller-manager-78b4bc895b-wwgt6\" (UID: \"e2bc662f-efe3-4e89-9b0e-e40eba87e41b\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-wwgt6" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.658920 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmfnp\" (UniqueName: \"kubernetes.io/projected/309f12a2-5b02-4d5f-9e28-63116ff8bb47-kube-api-access-dmfnp\") pod \"cinder-operator-controller-manager-859b6ccc6-xk9cb\" (UID: \"309f12a2-5b02-4d5f-9e28-63116ff8bb47\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-xk9cb" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.658960 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qxvg\" (UniqueName: \"kubernetes.io/projected/1c3d06b1-b999-4c82-85b5-11bdd996770c-kube-api-access-8qxvg\") pod \"barbican-operator-controller-manager-7d9dfd778-p8kv6\" (UID: \"1c3d06b1-b999-4c82-85b5-11bdd996770c\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-p8kv6" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.659377 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-nr756" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.665300 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-55cb9d4b9c-srntd"] Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.691169 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-gf27t"] Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.692105 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-gf27t" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.699938 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-bcpns" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.711823 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-gf27t"] Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.734180 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-tvqdm"] Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.736459 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tvqdm" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.746291 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-hpsfv"] Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.763040 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-hpsfv" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.764497 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2gdb\" (UniqueName: \"kubernetes.io/projected/e2bc662f-efe3-4e89-9b0e-e40eba87e41b-kube-api-access-v2gdb\") pod \"designate-operator-controller-manager-78b4bc895b-wwgt6\" (UID: \"e2bc662f-efe3-4e89-9b0e-e40eba87e41b\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-wwgt6" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.764542 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdk8j\" (UniqueName: \"kubernetes.io/projected/faf6c2c3-b028-431f-a302-3ad181dd93ae-kube-api-access-qdk8j\") pod \"heat-operator-controller-manager-55cb9d4b9c-srntd\" (UID: \"faf6c2c3-b028-431f-a302-3ad181dd93ae\") " pod="openstack-operators/heat-operator-controller-manager-55cb9d4b9c-srntd" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.764581 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmfnp\" (UniqueName: \"kubernetes.io/projected/309f12a2-5b02-4d5f-9e28-63116ff8bb47-kube-api-access-dmfnp\") pod \"cinder-operator-controller-manager-859b6ccc6-xk9cb\" (UID: \"309f12a2-5b02-4d5f-9e28-63116ff8bb47\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-xk9cb" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.764601 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8dzj\" (UniqueName: \"kubernetes.io/projected/08b2817d-5061-4375-82d4-726108586852-kube-api-access-k8dzj\") pod \"glance-operator-controller-manager-77987cd8cd-zkp5c\" (UID: \"08b2817d-5061-4375-82d4-726108586852\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-zkp5c" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.764632 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pssz8\" (UniqueName: \"kubernetes.io/projected/5f8e1071-aa69-4e21-94c9-f533d55f2cac-kube-api-access-pssz8\") pod \"horizon-operator-controller-manager-68c6d99b8f-gf27t\" (UID: \"5f8e1071-aa69-4e21-94c9-f533d55f2cac\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-gf27t" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.764653 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8qxvg\" (UniqueName: \"kubernetes.io/projected/1c3d06b1-b999-4c82-85b5-11bdd996770c-kube-api-access-8qxvg\") pod \"barbican-operator-controller-manager-7d9dfd778-p8kv6\" (UID: \"1c3d06b1-b999-4c82-85b5-11bdd996770c\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-p8kv6" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.778192 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-jrdns" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.778447 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.801027 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-c4sr8"] Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.802306 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-c4sr8" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.807146 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-wfx9x" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.816390 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qxvg\" (UniqueName: \"kubernetes.io/projected/1c3d06b1-b999-4c82-85b5-11bdd996770c-kube-api-access-8qxvg\") pod \"barbican-operator-controller-manager-7d9dfd778-p8kv6\" (UID: \"1c3d06b1-b999-4c82-85b5-11bdd996770c\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-p8kv6" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.817357 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2gdb\" (UniqueName: \"kubernetes.io/projected/e2bc662f-efe3-4e89-9b0e-e40eba87e41b-kube-api-access-v2gdb\") pod \"designate-operator-controller-manager-78b4bc895b-wwgt6\" (UID: \"e2bc662f-efe3-4e89-9b0e-e40eba87e41b\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-wwgt6" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.823196 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmfnp\" (UniqueName: \"kubernetes.io/projected/309f12a2-5b02-4d5f-9e28-63116ff8bb47-kube-api-access-dmfnp\") pod \"cinder-operator-controller-manager-859b6ccc6-xk9cb\" (UID: \"309f12a2-5b02-4d5f-9e28-63116ff8bb47\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-xk9cb" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.827874 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-hpsfv"] Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.856648 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-c4sr8"] Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.866173 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/534f227e-60b2-4149-b8d6-3762d39388e2-utilities\") pod \"redhat-marketplace-tvqdm\" (UID: \"534f227e-60b2-4149-b8d6-3762d39388e2\") " pod="openshift-marketplace/redhat-marketplace-tvqdm" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.866216 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/19f0c68c-32a3-41e2-a8ca-5ab0dc28d748-cert\") pod \"infra-operator-controller-manager-78d48bff9d-hpsfv\" (UID: \"19f0c68c-32a3-41e2-a8ca-5ab0dc28d748\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-hpsfv" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.866244 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8bsk\" (UniqueName: \"kubernetes.io/projected/19f0c68c-32a3-41e2-a8ca-5ab0dc28d748-kube-api-access-m8bsk\") pod \"infra-operator-controller-manager-78d48bff9d-hpsfv\" (UID: \"19f0c68c-32a3-41e2-a8ca-5ab0dc28d748\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-hpsfv" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.866262 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtrjk\" (UniqueName: \"kubernetes.io/projected/534f227e-60b2-4149-b8d6-3762d39388e2-kube-api-access-gtrjk\") pod \"redhat-marketplace-tvqdm\" (UID: \"534f227e-60b2-4149-b8d6-3762d39388e2\") " pod="openshift-marketplace/redhat-marketplace-tvqdm" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.866329 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdk8j\" (UniqueName: \"kubernetes.io/projected/faf6c2c3-b028-431f-a302-3ad181dd93ae-kube-api-access-qdk8j\") pod \"heat-operator-controller-manager-55cb9d4b9c-srntd\" (UID: \"faf6c2c3-b028-431f-a302-3ad181dd93ae\") " pod="openstack-operators/heat-operator-controller-manager-55cb9d4b9c-srntd" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.866361 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/534f227e-60b2-4149-b8d6-3762d39388e2-catalog-content\") pod \"redhat-marketplace-tvqdm\" (UID: \"534f227e-60b2-4149-b8d6-3762d39388e2\") " pod="openshift-marketplace/redhat-marketplace-tvqdm" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.866395 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8dzj\" (UniqueName: \"kubernetes.io/projected/08b2817d-5061-4375-82d4-726108586852-kube-api-access-k8dzj\") pod \"glance-operator-controller-manager-77987cd8cd-zkp5c\" (UID: \"08b2817d-5061-4375-82d4-726108586852\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-zkp5c" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.866419 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2kn7\" (UniqueName: \"kubernetes.io/projected/d71b7639-ee30-4323-8be0-a16d844d259f-kube-api-access-m2kn7\") pod \"ironic-operator-controller-manager-6c548fd776-c4sr8\" (UID: \"d71b7639-ee30-4323-8be0-a16d844d259f\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-c4sr8" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.866445 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pssz8\" (UniqueName: \"kubernetes.io/projected/5f8e1071-aa69-4e21-94c9-f533d55f2cac-kube-api-access-pssz8\") pod \"horizon-operator-controller-manager-68c6d99b8f-gf27t\" (UID: \"5f8e1071-aa69-4e21-94c9-f533d55f2cac\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-gf27t" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.867546 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-h7629"] Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.868551 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-h7629" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.874413 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-wq9jp" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.875389 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tvqdm"] Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.894176 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdk8j\" (UniqueName: \"kubernetes.io/projected/faf6c2c3-b028-431f-a302-3ad181dd93ae-kube-api-access-qdk8j\") pod \"heat-operator-controller-manager-55cb9d4b9c-srntd\" (UID: \"faf6c2c3-b028-431f-a302-3ad181dd93ae\") " pod="openstack-operators/heat-operator-controller-manager-55cb9d4b9c-srntd" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.894814 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-h7629"] Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.894995 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-p8kv6" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.902282 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-xk9cb" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.915609 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8dzj\" (UniqueName: \"kubernetes.io/projected/08b2817d-5061-4375-82d4-726108586852-kube-api-access-k8dzj\") pod \"glance-operator-controller-manager-77987cd8cd-zkp5c\" (UID: \"08b2817d-5061-4375-82d4-726108586852\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-zkp5c" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.915609 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pssz8\" (UniqueName: \"kubernetes.io/projected/5f8e1071-aa69-4e21-94c9-f533d55f2cac-kube-api-access-pssz8\") pod \"horizon-operator-controller-manager-68c6d99b8f-gf27t\" (UID: \"5f8e1071-aa69-4e21-94c9-f533d55f2cac\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-gf27t" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.922285 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-74c6z"] Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.923352 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-74c6z" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.927843 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-wmpwl" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.931761 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-wwgt6" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.937642 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-74c6z"] Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.944126 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-zkp5c" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.956689 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-bkfc6"] Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.958541 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-bkfc6" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.960043 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-slx8g" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.962669 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hlx24"] Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.963920 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hlx24" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.965039 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-p57t7" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.966354 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-bkfc6"] Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.967236 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/19f0c68c-32a3-41e2-a8ca-5ab0dc28d748-cert\") pod \"infra-operator-controller-manager-78d48bff9d-hpsfv\" (UID: \"19f0c68c-32a3-41e2-a8ca-5ab0dc28d748\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-hpsfv" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.967271 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtrjk\" (UniqueName: \"kubernetes.io/projected/534f227e-60b2-4149-b8d6-3762d39388e2-kube-api-access-gtrjk\") pod \"redhat-marketplace-tvqdm\" (UID: \"534f227e-60b2-4149-b8d6-3762d39388e2\") " pod="openshift-marketplace/redhat-marketplace-tvqdm" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.967292 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8bsk\" (UniqueName: \"kubernetes.io/projected/19f0c68c-32a3-41e2-a8ca-5ab0dc28d748-kube-api-access-m8bsk\") pod \"infra-operator-controller-manager-78d48bff9d-hpsfv\" (UID: \"19f0c68c-32a3-41e2-a8ca-5ab0dc28d748\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-hpsfv" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.967315 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2ljb\" (UniqueName: \"kubernetes.io/projected/c2e52d5b-b8cf-41f2-933e-af31359dc69d-kube-api-access-l2ljb\") pod \"keystone-operator-controller-manager-7765d96ddf-h7629\" (UID: \"c2e52d5b-b8cf-41f2-933e-af31359dc69d\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-h7629" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.967349 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/534f227e-60b2-4149-b8d6-3762d39388e2-catalog-content\") pod \"redhat-marketplace-tvqdm\" (UID: \"534f227e-60b2-4149-b8d6-3762d39388e2\") " pod="openshift-marketplace/redhat-marketplace-tvqdm" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.967392 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2kn7\" (UniqueName: \"kubernetes.io/projected/d71b7639-ee30-4323-8be0-a16d844d259f-kube-api-access-m2kn7\") pod \"ironic-operator-controller-manager-6c548fd776-c4sr8\" (UID: \"d71b7639-ee30-4323-8be0-a16d844d259f\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-c4sr8" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.967429 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/534f227e-60b2-4149-b8d6-3762d39388e2-utilities\") pod \"redhat-marketplace-tvqdm\" (UID: \"534f227e-60b2-4149-b8d6-3762d39388e2\") " pod="openshift-marketplace/redhat-marketplace-tvqdm" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.967846 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/534f227e-60b2-4149-b8d6-3762d39388e2-utilities\") pod \"redhat-marketplace-tvqdm\" (UID: \"534f227e-60b2-4149-b8d6-3762d39388e2\") " pod="openshift-marketplace/redhat-marketplace-tvqdm" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.968060 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/534f227e-60b2-4149-b8d6-3762d39388e2-catalog-content\") pod \"redhat-marketplace-tvqdm\" (UID: \"534f227e-60b2-4149-b8d6-3762d39388e2\") " pod="openshift-marketplace/redhat-marketplace-tvqdm" Dec 06 03:55:00 crc kubenswrapper[4802]: E1206 03:55:00.968266 4802 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 06 03:55:00 crc kubenswrapper[4802]: E1206 03:55:00.968303 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/19f0c68c-32a3-41e2-a8ca-5ab0dc28d748-cert podName:19f0c68c-32a3-41e2-a8ca-5ab0dc28d748 nodeName:}" failed. No retries permitted until 2025-12-06 03:55:01.468289822 +0000 UTC m=+894.340198974 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/19f0c68c-32a3-41e2-a8ca-5ab0dc28d748-cert") pod "infra-operator-controller-manager-78d48bff9d-hpsfv" (UID: "19f0c68c-32a3-41e2-a8ca-5ab0dc28d748") : secret "infra-operator-webhook-server-cert" not found Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.970130 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hlx24"] Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.980163 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-55cb9d4b9c-srntd" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.981799 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-4m2vm"] Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.982880 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-4m2vm" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.984378 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-tgsmm" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.985117 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8bsk\" (UniqueName: \"kubernetes.io/projected/19f0c68c-32a3-41e2-a8ca-5ab0dc28d748-kube-api-access-m8bsk\") pod \"infra-operator-controller-manager-78d48bff9d-hpsfv\" (UID: \"19f0c68c-32a3-41e2-a8ca-5ab0dc28d748\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-hpsfv" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.989775 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtrjk\" (UniqueName: \"kubernetes.io/projected/534f227e-60b2-4149-b8d6-3762d39388e2-kube-api-access-gtrjk\") pod \"redhat-marketplace-tvqdm\" (UID: \"534f227e-60b2-4149-b8d6-3762d39388e2\") " pod="openshift-marketplace/redhat-marketplace-tvqdm" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.991263 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2kn7\" (UniqueName: \"kubernetes.io/projected/d71b7639-ee30-4323-8be0-a16d844d259f-kube-api-access-m2kn7\") pod \"ironic-operator-controller-manager-6c548fd776-c4sr8\" (UID: \"d71b7639-ee30-4323-8be0-a16d844d259f\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-c4sr8" Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.993155 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-lk8bf"] Dec 06 03:55:00 crc kubenswrapper[4802]: I1206 03:55:00.994073 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-lk8bf" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.002135 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-grxmz" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.002552 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-8q2cg"] Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.003640 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-8q2cg" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.004995 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-fjj4s" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.006363 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-lk8bf"] Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.011137 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4pwn82"] Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.012194 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4pwn82" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.022027 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.022317 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-n55lf" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.024260 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-4m2vm"] Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.032108 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-gf27t" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.033316 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-8q2cg"] Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.068420 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lpxc\" (UniqueName: \"kubernetes.io/projected/8555a40e-f3d3-4627-8867-6db5ca5ed06a-kube-api-access-2lpxc\") pod \"octavia-operator-controller-manager-998648c74-lk8bf\" (UID: \"8555a40e-f3d3-4627-8867-6db5ca5ed06a\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-lk8bf" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.068467 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdgxg\" (UniqueName: \"kubernetes.io/projected/a076799b-6631-474a-b552-9cf2261c377a-kube-api-access-qdgxg\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-hlx24\" (UID: \"a076799b-6631-474a-b552-9cf2261c377a\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hlx24" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.068629 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9wb4\" (UniqueName: \"kubernetes.io/projected/119bce90-0fad-424d-89cf-2ca06d7126ae-kube-api-access-p9wb4\") pod \"ovn-operator-controller-manager-b6456fdb6-8q2cg\" (UID: \"119bce90-0fad-424d-89cf-2ca06d7126ae\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-8q2cg" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.068654 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2ljb\" (UniqueName: \"kubernetes.io/projected/c2e52d5b-b8cf-41f2-933e-af31359dc69d-kube-api-access-l2ljb\") pod \"keystone-operator-controller-manager-7765d96ddf-h7629\" (UID: \"c2e52d5b-b8cf-41f2-933e-af31359dc69d\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-h7629" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.068674 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwjn9\" (UniqueName: \"kubernetes.io/projected/ef258927-21c4-4147-8fb5-3db5d78658a3-kube-api-access-xwjn9\") pod \"nova-operator-controller-manager-697bc559fc-4m2vm\" (UID: \"ef258927-21c4-4147-8fb5-3db5d78658a3\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-4m2vm" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.068699 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fc9wr\" (UniqueName: \"kubernetes.io/projected/4460e15d-1d9a-4814-a7d2-042c6a84057f-kube-api-access-fc9wr\") pod \"mariadb-operator-controller-manager-56bbcc9d85-bkfc6\" (UID: \"4460e15d-1d9a-4814-a7d2-042c6a84057f\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-bkfc6" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.068737 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wgdv\" (UniqueName: \"kubernetes.io/projected/1d3c0479-80bf-46af-a261-42b3834807f1-kube-api-access-9wgdv\") pod \"manila-operator-controller-manager-7c79b5df47-74c6z\" (UID: \"1d3c0479-80bf-46af-a261-42b3834807f1\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-74c6z" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.080937 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4pwn82"] Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.085981 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tvqdm" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.089104 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2ljb\" (UniqueName: \"kubernetes.io/projected/c2e52d5b-b8cf-41f2-933e-af31359dc69d-kube-api-access-l2ljb\") pod \"keystone-operator-controller-manager-7765d96ddf-h7629\" (UID: \"c2e52d5b-b8cf-41f2-933e-af31359dc69d\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-h7629" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.095548 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-4x9rb"] Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.097077 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-4x9rb" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.099723 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-frpb6" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.104548 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-4x9rb"] Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.127783 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-mmh24"] Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.129636 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-mmh24" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.133160 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-p4q6f" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.172970 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wgdv\" (UniqueName: \"kubernetes.io/projected/1d3c0479-80bf-46af-a261-42b3834807f1-kube-api-access-9wgdv\") pod \"manila-operator-controller-manager-7c79b5df47-74c6z\" (UID: \"1d3c0479-80bf-46af-a261-42b3834807f1\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-74c6z" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.173064 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lpxc\" (UniqueName: \"kubernetes.io/projected/8555a40e-f3d3-4627-8867-6db5ca5ed06a-kube-api-access-2lpxc\") pod \"octavia-operator-controller-manager-998648c74-lk8bf\" (UID: \"8555a40e-f3d3-4627-8867-6db5ca5ed06a\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-lk8bf" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.173111 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdgxg\" (UniqueName: \"kubernetes.io/projected/a076799b-6631-474a-b552-9cf2261c377a-kube-api-access-qdgxg\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-hlx24\" (UID: \"a076799b-6631-474a-b552-9cf2261c377a\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hlx24" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.173156 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrbmv\" (UniqueName: \"kubernetes.io/projected/05736c59-c4db-4cfe-addd-1b29c552596e-kube-api-access-lrbmv\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4pwn82\" (UID: \"05736c59-c4db-4cfe-addd-1b29c552596e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4pwn82" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.173264 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9wb4\" (UniqueName: \"kubernetes.io/projected/119bce90-0fad-424d-89cf-2ca06d7126ae-kube-api-access-p9wb4\") pod \"ovn-operator-controller-manager-b6456fdb6-8q2cg\" (UID: \"119bce90-0fad-424d-89cf-2ca06d7126ae\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-8q2cg" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.173331 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwjn9\" (UniqueName: \"kubernetes.io/projected/ef258927-21c4-4147-8fb5-3db5d78658a3-kube-api-access-xwjn9\") pod \"nova-operator-controller-manager-697bc559fc-4m2vm\" (UID: \"ef258927-21c4-4147-8fb5-3db5d78658a3\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-4m2vm" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.173365 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvl2m\" (UniqueName: \"kubernetes.io/projected/79efe004-97e7-4586-b48e-08ce9285bb79-kube-api-access-wvl2m\") pod \"placement-operator-controller-manager-78f8948974-4x9rb\" (UID: \"79efe004-97e7-4586-b48e-08ce9285bb79\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-4x9rb" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.173393 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/05736c59-c4db-4cfe-addd-1b29c552596e-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4pwn82\" (UID: \"05736c59-c4db-4cfe-addd-1b29c552596e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4pwn82" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.173433 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fc9wr\" (UniqueName: \"kubernetes.io/projected/4460e15d-1d9a-4814-a7d2-042c6a84057f-kube-api-access-fc9wr\") pod \"mariadb-operator-controller-manager-56bbcc9d85-bkfc6\" (UID: \"4460e15d-1d9a-4814-a7d2-042c6a84057f\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-bkfc6" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.183392 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-mmh24"] Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.184943 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-c4sr8" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.239498 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9wb4\" (UniqueName: \"kubernetes.io/projected/119bce90-0fad-424d-89cf-2ca06d7126ae-kube-api-access-p9wb4\") pod \"ovn-operator-controller-manager-b6456fdb6-8q2cg\" (UID: \"119bce90-0fad-424d-89cf-2ca06d7126ae\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-8q2cg" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.243410 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwjn9\" (UniqueName: \"kubernetes.io/projected/ef258927-21c4-4147-8fb5-3db5d78658a3-kube-api-access-xwjn9\") pod \"nova-operator-controller-manager-697bc559fc-4m2vm\" (UID: \"ef258927-21c4-4147-8fb5-3db5d78658a3\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-4m2vm" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.245245 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-nd42v"] Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.246322 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lpxc\" (UniqueName: \"kubernetes.io/projected/8555a40e-f3d3-4627-8867-6db5ca5ed06a-kube-api-access-2lpxc\") pod \"octavia-operator-controller-manager-998648c74-lk8bf\" (UID: \"8555a40e-f3d3-4627-8867-6db5ca5ed06a\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-lk8bf" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.246855 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fc9wr\" (UniqueName: \"kubernetes.io/projected/4460e15d-1d9a-4814-a7d2-042c6a84057f-kube-api-access-fc9wr\") pod \"mariadb-operator-controller-manager-56bbcc9d85-bkfc6\" (UID: \"4460e15d-1d9a-4814-a7d2-042c6a84057f\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-bkfc6" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.248551 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdgxg\" (UniqueName: \"kubernetes.io/projected/a076799b-6631-474a-b552-9cf2261c377a-kube-api-access-qdgxg\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-hlx24\" (UID: \"a076799b-6631-474a-b552-9cf2261c377a\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hlx24" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.250114 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wgdv\" (UniqueName: \"kubernetes.io/projected/1d3c0479-80bf-46af-a261-42b3834807f1-kube-api-access-9wgdv\") pod \"manila-operator-controller-manager-7c79b5df47-74c6z\" (UID: \"1d3c0479-80bf-46af-a261-42b3834807f1\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-74c6z" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.267903 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-nd42v" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.268854 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-h7629" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.273825 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-7nkgq" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.274487 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tsfs2\" (UniqueName: \"kubernetes.io/projected/11885e74-7bc3-4207-b046-9e94ab51e499-kube-api-access-tsfs2\") pod \"swift-operator-controller-manager-5f8c65bbfc-mmh24\" (UID: \"11885e74-7bc3-4207-b046-9e94ab51e499\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-mmh24" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.274538 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrbmv\" (UniqueName: \"kubernetes.io/projected/05736c59-c4db-4cfe-addd-1b29c552596e-kube-api-access-lrbmv\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4pwn82\" (UID: \"05736c59-c4db-4cfe-addd-1b29c552596e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4pwn82" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.274584 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvl2m\" (UniqueName: \"kubernetes.io/projected/79efe004-97e7-4586-b48e-08ce9285bb79-kube-api-access-wvl2m\") pod \"placement-operator-controller-manager-78f8948974-4x9rb\" (UID: \"79efe004-97e7-4586-b48e-08ce9285bb79\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-4x9rb" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.274613 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/05736c59-c4db-4cfe-addd-1b29c552596e-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4pwn82\" (UID: \"05736c59-c4db-4cfe-addd-1b29c552596e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4pwn82" Dec 06 03:55:01 crc kubenswrapper[4802]: E1206 03:55:01.274713 4802 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 03:55:01 crc kubenswrapper[4802]: E1206 03:55:01.274762 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/05736c59-c4db-4cfe-addd-1b29c552596e-cert podName:05736c59-c4db-4cfe-addd-1b29c552596e nodeName:}" failed. No retries permitted until 2025-12-06 03:55:01.77473577 +0000 UTC m=+894.646644922 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/05736c59-c4db-4cfe-addd-1b29c552596e-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4pwn82" (UID: "05736c59-c4db-4cfe-addd-1b29c552596e") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.277713 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-nd42v"] Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.297008 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-74c6z" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.307853 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-bkfc6" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.312634 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrbmv\" (UniqueName: \"kubernetes.io/projected/05736c59-c4db-4cfe-addd-1b29c552596e-kube-api-access-lrbmv\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4pwn82\" (UID: \"05736c59-c4db-4cfe-addd-1b29c552596e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4pwn82" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.313275 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvl2m\" (UniqueName: \"kubernetes.io/projected/79efe004-97e7-4586-b48e-08ce9285bb79-kube-api-access-wvl2m\") pod \"placement-operator-controller-manager-78f8948974-4x9rb\" (UID: \"79efe004-97e7-4586-b48e-08ce9285bb79\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-4x9rb" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.338409 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hlx24" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.360216 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-4m2vm" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.381096 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tsfs2\" (UniqueName: \"kubernetes.io/projected/11885e74-7bc3-4207-b046-9e94ab51e499-kube-api-access-tsfs2\") pod \"swift-operator-controller-manager-5f8c65bbfc-mmh24\" (UID: \"11885e74-7bc3-4207-b046-9e94ab51e499\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-mmh24" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.381175 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwnnp\" (UniqueName: \"kubernetes.io/projected/18ad32ab-222c-41f9-8409-90fd51178f2d-kube-api-access-lwnnp\") pod \"telemetry-operator-controller-manager-76cc84c6bb-nd42v\" (UID: \"18ad32ab-222c-41f9-8409-90fd51178f2d\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-nd42v" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.386905 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-lk8bf" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.404881 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-q6k7m"] Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.406060 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-q6k7m" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.408444 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-8q2cg" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.412265 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-x9jxw" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.412324 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tsfs2\" (UniqueName: \"kubernetes.io/projected/11885e74-7bc3-4207-b046-9e94ab51e499-kube-api-access-tsfs2\") pod \"swift-operator-controller-manager-5f8c65bbfc-mmh24\" (UID: \"11885e74-7bc3-4207-b046-9e94ab51e499\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-mmh24" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.424041 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-q6k7m"] Dec 06 03:55:01 crc kubenswrapper[4802]: W1206 03:55:01.436818 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1c3d06b1_b999_4c82_85b5_11bdd996770c.slice/crio-097874769b1edd51da005a2686fde7aa07f6f801cdf04b59182e8d2f248a9506 WatchSource:0}: Error finding container 097874769b1edd51da005a2686fde7aa07f6f801cdf04b59182e8d2f248a9506: Status 404 returned error can't find the container with id 097874769b1edd51da005a2686fde7aa07f6f801cdf04b59182e8d2f248a9506 Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.445429 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-4x9rb" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.463502 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-mmh24" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.468910 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-w64dr"] Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.469880 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-w64dr" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.471740 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-s4n5k" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.474036 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-w64dr"] Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.478794 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-xk9cb"] Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.482967 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwnnp\" (UniqueName: \"kubernetes.io/projected/18ad32ab-222c-41f9-8409-90fd51178f2d-kube-api-access-lwnnp\") pod \"telemetry-operator-controller-manager-76cc84c6bb-nd42v\" (UID: \"18ad32ab-222c-41f9-8409-90fd51178f2d\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-nd42v" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.483009 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/19f0c68c-32a3-41e2-a8ca-5ab0dc28d748-cert\") pod \"infra-operator-controller-manager-78d48bff9d-hpsfv\" (UID: \"19f0c68c-32a3-41e2-a8ca-5ab0dc28d748\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-hpsfv" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.483137 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsnpc\" (UniqueName: \"kubernetes.io/projected/e9a14170-cafd-433c-83a4-22793b21d376-kube-api-access-jsnpc\") pod \"test-operator-controller-manager-5854674fcc-q6k7m\" (UID: \"e9a14170-cafd-433c-83a4-22793b21d376\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-q6k7m" Dec 06 03:55:01 crc kubenswrapper[4802]: E1206 03:55:01.483275 4802 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 06 03:55:01 crc kubenswrapper[4802]: E1206 03:55:01.483344 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/19f0c68c-32a3-41e2-a8ca-5ab0dc28d748-cert podName:19f0c68c-32a3-41e2-a8ca-5ab0dc28d748 nodeName:}" failed. No retries permitted until 2025-12-06 03:55:02.483326961 +0000 UTC m=+895.355236113 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/19f0c68c-32a3-41e2-a8ca-5ab0dc28d748-cert") pod "infra-operator-controller-manager-78d48bff9d-hpsfv" (UID: "19f0c68c-32a3-41e2-a8ca-5ab0dc28d748") : secret "infra-operator-webhook-server-cert" not found Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.487226 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-77f4d6cd-9s876"] Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.488384 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-77f4d6cd-9s876" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.495953 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-77f4d6cd-9s876"] Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.496927 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-krrt8" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.496966 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.497161 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.502233 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-p8kv6"] Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.507610 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwnnp\" (UniqueName: \"kubernetes.io/projected/18ad32ab-222c-41f9-8409-90fd51178f2d-kube-api-access-lwnnp\") pod \"telemetry-operator-controller-manager-76cc84c6bb-nd42v\" (UID: \"18ad32ab-222c-41f9-8409-90fd51178f2d\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-nd42v" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.508502 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-b5wv6"] Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.509966 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-b5wv6" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.511708 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-dmrc5" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.516160 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-b5wv6"] Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.584083 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5f6a5e2f-8db1-4d31-8a2d-5810c5e62083-webhook-certs\") pod \"openstack-operator-controller-manager-77f4d6cd-9s876\" (UID: \"5f6a5e2f-8db1-4d31-8a2d-5810c5e62083\") " pod="openstack-operators/openstack-operator-controller-manager-77f4d6cd-9s876" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.584121 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrtdq\" (UniqueName: \"kubernetes.io/projected/91b87773-2f3d-4e35-9a13-708f07cb2bbf-kube-api-access-rrtdq\") pod \"watcher-operator-controller-manager-769dc69bc-w64dr\" (UID: \"91b87773-2f3d-4e35-9a13-708f07cb2bbf\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-w64dr" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.584146 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swnsm\" (UniqueName: \"kubernetes.io/projected/5f6a5e2f-8db1-4d31-8a2d-5810c5e62083-kube-api-access-swnsm\") pod \"openstack-operator-controller-manager-77f4d6cd-9s876\" (UID: \"5f6a5e2f-8db1-4d31-8a2d-5810c5e62083\") " pod="openstack-operators/openstack-operator-controller-manager-77f4d6cd-9s876" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.584297 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4sbz\" (UniqueName: \"kubernetes.io/projected/a3bed886-a616-4733-947a-288ada88dc74-kube-api-access-x4sbz\") pod \"rabbitmq-cluster-operator-manager-668c99d594-b5wv6\" (UID: \"a3bed886-a616-4733-947a-288ada88dc74\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-b5wv6" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.584354 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5f6a5e2f-8db1-4d31-8a2d-5810c5e62083-metrics-certs\") pod \"openstack-operator-controller-manager-77f4d6cd-9s876\" (UID: \"5f6a5e2f-8db1-4d31-8a2d-5810c5e62083\") " pod="openstack-operators/openstack-operator-controller-manager-77f4d6cd-9s876" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.584484 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsnpc\" (UniqueName: \"kubernetes.io/projected/e9a14170-cafd-433c-83a4-22793b21d376-kube-api-access-jsnpc\") pod \"test-operator-controller-manager-5854674fcc-q6k7m\" (UID: \"e9a14170-cafd-433c-83a4-22793b21d376\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-q6k7m" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.604244 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsnpc\" (UniqueName: \"kubernetes.io/projected/e9a14170-cafd-433c-83a4-22793b21d376-kube-api-access-jsnpc\") pod \"test-operator-controller-manager-5854674fcc-q6k7m\" (UID: \"e9a14170-cafd-433c-83a4-22793b21d376\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-q6k7m" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.622824 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-nd42v" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.665803 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-wwgt6"] Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.686315 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5f6a5e2f-8db1-4d31-8a2d-5810c5e62083-webhook-certs\") pod \"openstack-operator-controller-manager-77f4d6cd-9s876\" (UID: \"5f6a5e2f-8db1-4d31-8a2d-5810c5e62083\") " pod="openstack-operators/openstack-operator-controller-manager-77f4d6cd-9s876" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.686355 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrtdq\" (UniqueName: \"kubernetes.io/projected/91b87773-2f3d-4e35-9a13-708f07cb2bbf-kube-api-access-rrtdq\") pod \"watcher-operator-controller-manager-769dc69bc-w64dr\" (UID: \"91b87773-2f3d-4e35-9a13-708f07cb2bbf\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-w64dr" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.686384 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swnsm\" (UniqueName: \"kubernetes.io/projected/5f6a5e2f-8db1-4d31-8a2d-5810c5e62083-kube-api-access-swnsm\") pod \"openstack-operator-controller-manager-77f4d6cd-9s876\" (UID: \"5f6a5e2f-8db1-4d31-8a2d-5810c5e62083\") " pod="openstack-operators/openstack-operator-controller-manager-77f4d6cd-9s876" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.686408 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4sbz\" (UniqueName: \"kubernetes.io/projected/a3bed886-a616-4733-947a-288ada88dc74-kube-api-access-x4sbz\") pod \"rabbitmq-cluster-operator-manager-668c99d594-b5wv6\" (UID: \"a3bed886-a616-4733-947a-288ada88dc74\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-b5wv6" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.686435 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5f6a5e2f-8db1-4d31-8a2d-5810c5e62083-metrics-certs\") pod \"openstack-operator-controller-manager-77f4d6cd-9s876\" (UID: \"5f6a5e2f-8db1-4d31-8a2d-5810c5e62083\") " pod="openstack-operators/openstack-operator-controller-manager-77f4d6cd-9s876" Dec 06 03:55:01 crc kubenswrapper[4802]: E1206 03:55:01.686581 4802 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 06 03:55:01 crc kubenswrapper[4802]: E1206 03:55:01.686623 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5f6a5e2f-8db1-4d31-8a2d-5810c5e62083-metrics-certs podName:5f6a5e2f-8db1-4d31-8a2d-5810c5e62083 nodeName:}" failed. No retries permitted until 2025-12-06 03:55:02.18660933 +0000 UTC m=+895.058518482 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5f6a5e2f-8db1-4d31-8a2d-5810c5e62083-metrics-certs") pod "openstack-operator-controller-manager-77f4d6cd-9s876" (UID: "5f6a5e2f-8db1-4d31-8a2d-5810c5e62083") : secret "metrics-server-cert" not found Dec 06 03:55:01 crc kubenswrapper[4802]: E1206 03:55:01.686884 4802 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 06 03:55:01 crc kubenswrapper[4802]: E1206 03:55:01.686908 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5f6a5e2f-8db1-4d31-8a2d-5810c5e62083-webhook-certs podName:5f6a5e2f-8db1-4d31-8a2d-5810c5e62083 nodeName:}" failed. No retries permitted until 2025-12-06 03:55:02.186901327 +0000 UTC m=+895.058810479 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/5f6a5e2f-8db1-4d31-8a2d-5810c5e62083-webhook-certs") pod "openstack-operator-controller-manager-77f4d6cd-9s876" (UID: "5f6a5e2f-8db1-4d31-8a2d-5810c5e62083") : secret "webhook-server-cert" not found Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.704347 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4sbz\" (UniqueName: \"kubernetes.io/projected/a3bed886-a616-4733-947a-288ada88dc74-kube-api-access-x4sbz\") pod \"rabbitmq-cluster-operator-manager-668c99d594-b5wv6\" (UID: \"a3bed886-a616-4733-947a-288ada88dc74\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-b5wv6" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.704816 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrtdq\" (UniqueName: \"kubernetes.io/projected/91b87773-2f3d-4e35-9a13-708f07cb2bbf-kube-api-access-rrtdq\") pod \"watcher-operator-controller-manager-769dc69bc-w64dr\" (UID: \"91b87773-2f3d-4e35-9a13-708f07cb2bbf\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-w64dr" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.728893 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swnsm\" (UniqueName: \"kubernetes.io/projected/5f6a5e2f-8db1-4d31-8a2d-5810c5e62083-kube-api-access-swnsm\") pod \"openstack-operator-controller-manager-77f4d6cd-9s876\" (UID: \"5f6a5e2f-8db1-4d31-8a2d-5810c5e62083\") " pod="openstack-operators/openstack-operator-controller-manager-77f4d6cd-9s876" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.787258 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/05736c59-c4db-4cfe-addd-1b29c552596e-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4pwn82\" (UID: \"05736c59-c4db-4cfe-addd-1b29c552596e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4pwn82" Dec 06 03:55:01 crc kubenswrapper[4802]: E1206 03:55:01.787391 4802 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 03:55:01 crc kubenswrapper[4802]: E1206 03:55:01.787432 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/05736c59-c4db-4cfe-addd-1b29c552596e-cert podName:05736c59-c4db-4cfe-addd-1b29c552596e nodeName:}" failed. No retries permitted until 2025-12-06 03:55:02.787418176 +0000 UTC m=+895.659327328 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/05736c59-c4db-4cfe-addd-1b29c552596e-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4pwn82" (UID: "05736c59-c4db-4cfe-addd-1b29c552596e") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.787663 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-q6k7m" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.874511 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-w64dr" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.887557 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-b5wv6" Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.926221 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tvqdm"] Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.947714 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-zkp5c"] Dec 06 03:55:01 crc kubenswrapper[4802]: I1206 03:55:01.951626 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-55cb9d4b9c-srntd"] Dec 06 03:55:02 crc kubenswrapper[4802]: W1206 03:55:02.037052 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod08b2817d_5061_4375_82d4_726108586852.slice/crio-9b4a648548550dc112bf2f1cf462cdd80908784b4f6a6516d5675352223f2d25 WatchSource:0}: Error finding container 9b4a648548550dc112bf2f1cf462cdd80908784b4f6a6516d5675352223f2d25: Status 404 returned error can't find the container with id 9b4a648548550dc112bf2f1cf462cdd80908784b4f6a6516d5675352223f2d25 Dec 06 03:55:02 crc kubenswrapper[4802]: W1206 03:55:02.038672 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfaf6c2c3_b028_431f_a302_3ad181dd93ae.slice/crio-d311450aaefd61d0f6868249c4fb088ecfdfc6f770e630741db449e8980f39b4 WatchSource:0}: Error finding container d311450aaefd61d0f6868249c4fb088ecfdfc6f770e630741db449e8980f39b4: Status 404 returned error can't find the container with id d311450aaefd61d0f6868249c4fb088ecfdfc6f770e630741db449e8980f39b4 Dec 06 03:55:02 crc kubenswrapper[4802]: I1206 03:55:02.080178 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tvqdm" event={"ID":"534f227e-60b2-4149-b8d6-3762d39388e2","Type":"ContainerStarted","Data":"ff92ee3f82287267027418e234cbd91d2d4b2e88e37246c067858483f86ebfac"} Dec 06 03:55:02 crc kubenswrapper[4802]: I1206 03:55:02.084913 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-xk9cb" event={"ID":"309f12a2-5b02-4d5f-9e28-63116ff8bb47","Type":"ContainerStarted","Data":"8198ed87447c5dc7b61befbfed443367e8eebf8f33f55c26acb8b8c299427925"} Dec 06 03:55:02 crc kubenswrapper[4802]: I1206 03:55:02.085800 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-55cb9d4b9c-srntd" event={"ID":"faf6c2c3-b028-431f-a302-3ad181dd93ae","Type":"ContainerStarted","Data":"d311450aaefd61d0f6868249c4fb088ecfdfc6f770e630741db449e8980f39b4"} Dec 06 03:55:02 crc kubenswrapper[4802]: I1206 03:55:02.087235 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-wwgt6" event={"ID":"e2bc662f-efe3-4e89-9b0e-e40eba87e41b","Type":"ContainerStarted","Data":"f6c1aa384c8d582888c4ac9c0f045d152838f3006bb757168a0289ec23e9afbc"} Dec 06 03:55:02 crc kubenswrapper[4802]: I1206 03:55:02.088670 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-p8kv6" event={"ID":"1c3d06b1-b999-4c82-85b5-11bdd996770c","Type":"ContainerStarted","Data":"097874769b1edd51da005a2686fde7aa07f6f801cdf04b59182e8d2f248a9506"} Dec 06 03:55:02 crc kubenswrapper[4802]: I1206 03:55:02.106922 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-zkp5c" event={"ID":"08b2817d-5061-4375-82d4-726108586852","Type":"ContainerStarted","Data":"9b4a648548550dc112bf2f1cf462cdd80908784b4f6a6516d5675352223f2d25"} Dec 06 03:55:02 crc kubenswrapper[4802]: I1206 03:55:02.207688 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-c4sr8"] Dec 06 03:55:02 crc kubenswrapper[4802]: I1206 03:55:02.209569 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5f6a5e2f-8db1-4d31-8a2d-5810c5e62083-webhook-certs\") pod \"openstack-operator-controller-manager-77f4d6cd-9s876\" (UID: \"5f6a5e2f-8db1-4d31-8a2d-5810c5e62083\") " pod="openstack-operators/openstack-operator-controller-manager-77f4d6cd-9s876" Dec 06 03:55:02 crc kubenswrapper[4802]: I1206 03:55:02.209628 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5f6a5e2f-8db1-4d31-8a2d-5810c5e62083-metrics-certs\") pod \"openstack-operator-controller-manager-77f4d6cd-9s876\" (UID: \"5f6a5e2f-8db1-4d31-8a2d-5810c5e62083\") " pod="openstack-operators/openstack-operator-controller-manager-77f4d6cd-9s876" Dec 06 03:55:02 crc kubenswrapper[4802]: E1206 03:55:02.209782 4802 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 06 03:55:02 crc kubenswrapper[4802]: E1206 03:55:02.209824 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5f6a5e2f-8db1-4d31-8a2d-5810c5e62083-metrics-certs podName:5f6a5e2f-8db1-4d31-8a2d-5810c5e62083 nodeName:}" failed. No retries permitted until 2025-12-06 03:55:03.20980937 +0000 UTC m=+896.081718522 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5f6a5e2f-8db1-4d31-8a2d-5810c5e62083-metrics-certs") pod "openstack-operator-controller-manager-77f4d6cd-9s876" (UID: "5f6a5e2f-8db1-4d31-8a2d-5810c5e62083") : secret "metrics-server-cert" not found Dec 06 03:55:02 crc kubenswrapper[4802]: E1206 03:55:02.210322 4802 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 06 03:55:02 crc kubenswrapper[4802]: E1206 03:55:02.210441 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5f6a5e2f-8db1-4d31-8a2d-5810c5e62083-webhook-certs podName:5f6a5e2f-8db1-4d31-8a2d-5810c5e62083 nodeName:}" failed. No retries permitted until 2025-12-06 03:55:03.210422777 +0000 UTC m=+896.082331929 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/5f6a5e2f-8db1-4d31-8a2d-5810c5e62083-webhook-certs") pod "openstack-operator-controller-manager-77f4d6cd-9s876" (UID: "5f6a5e2f-8db1-4d31-8a2d-5810c5e62083") : secret "webhook-server-cert" not found Dec 06 03:55:02 crc kubenswrapper[4802]: I1206 03:55:02.215134 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-gf27t"] Dec 06 03:55:02 crc kubenswrapper[4802]: I1206 03:55:02.437448 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-h7629"] Dec 06 03:55:02 crc kubenswrapper[4802]: W1206 03:55:02.451967 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc2e52d5b_b8cf_41f2_933e_af31359dc69d.slice/crio-6cee2b35fe0ae09b08ec7d7ab47071afcb5786071c64d01c8f20477592612f44 WatchSource:0}: Error finding container 6cee2b35fe0ae09b08ec7d7ab47071afcb5786071c64d01c8f20477592612f44: Status 404 returned error can't find the container with id 6cee2b35fe0ae09b08ec7d7ab47071afcb5786071c64d01c8f20477592612f44 Dec 06 03:55:02 crc kubenswrapper[4802]: I1206 03:55:02.461595 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-lk8bf"] Dec 06 03:55:02 crc kubenswrapper[4802]: W1206 03:55:02.467082 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda076799b_6631_474a_b552_9cf2261c377a.slice/crio-adace6ac7dd390dbcfa346c1c0670c94466e283509aa6925354cc92a3361f57c WatchSource:0}: Error finding container adace6ac7dd390dbcfa346c1c0670c94466e283509aa6925354cc92a3361f57c: Status 404 returned error can't find the container with id adace6ac7dd390dbcfa346c1c0670c94466e283509aa6925354cc92a3361f57c Dec 06 03:55:02 crc kubenswrapper[4802]: I1206 03:55:02.467128 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-4m2vm"] Dec 06 03:55:02 crc kubenswrapper[4802]: W1206 03:55:02.468824 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef258927_21c4_4147_8fb5_3db5d78658a3.slice/crio-932532fd81445277457ae5fb53432f9d70736a7b8d30b57bd455a122908355e9 WatchSource:0}: Error finding container 932532fd81445277457ae5fb53432f9d70736a7b8d30b57bd455a122908355e9: Status 404 returned error can't find the container with id 932532fd81445277457ae5fb53432f9d70736a7b8d30b57bd455a122908355e9 Dec 06 03:55:02 crc kubenswrapper[4802]: W1206 03:55:02.472229 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1d3c0479_80bf_46af_a261_42b3834807f1.slice/crio-f3d2319e19793e681185072d1029bc070ca9abf204a4f9bc829d46983327a656 WatchSource:0}: Error finding container f3d2319e19793e681185072d1029bc070ca9abf204a4f9bc829d46983327a656: Status 404 returned error can't find the container with id f3d2319e19793e681185072d1029bc070ca9abf204a4f9bc829d46983327a656 Dec 06 03:55:02 crc kubenswrapper[4802]: W1206 03:55:02.473765 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4460e15d_1d9a_4814_a7d2_042c6a84057f.slice/crio-2a3126b6297e5a341be861b96540b47b1cab8633b1d5d0965b46ad9ac8d817bc WatchSource:0}: Error finding container 2a3126b6297e5a341be861b96540b47b1cab8633b1d5d0965b46ad9ac8d817bc: Status 404 returned error can't find the container with id 2a3126b6297e5a341be861b96540b47b1cab8633b1d5d0965b46ad9ac8d817bc Dec 06 03:55:02 crc kubenswrapper[4802]: I1206 03:55:02.476768 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hlx24"] Dec 06 03:55:02 crc kubenswrapper[4802]: I1206 03:55:02.484304 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-bkfc6"] Dec 06 03:55:02 crc kubenswrapper[4802]: I1206 03:55:02.490314 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-74c6z"] Dec 06 03:55:02 crc kubenswrapper[4802]: I1206 03:55:02.517429 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/19f0c68c-32a3-41e2-a8ca-5ab0dc28d748-cert\") pod \"infra-operator-controller-manager-78d48bff9d-hpsfv\" (UID: \"19f0c68c-32a3-41e2-a8ca-5ab0dc28d748\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-hpsfv" Dec 06 03:55:02 crc kubenswrapper[4802]: E1206 03:55:02.517650 4802 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 06 03:55:02 crc kubenswrapper[4802]: E1206 03:55:02.517735 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/19f0c68c-32a3-41e2-a8ca-5ab0dc28d748-cert podName:19f0c68c-32a3-41e2-a8ca-5ab0dc28d748 nodeName:}" failed. No retries permitted until 2025-12-06 03:55:04.517716618 +0000 UTC m=+897.389625770 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/19f0c68c-32a3-41e2-a8ca-5ab0dc28d748-cert") pod "infra-operator-controller-manager-78d48bff9d-hpsfv" (UID: "19f0c68c-32a3-41e2-a8ca-5ab0dc28d748") : secret "infra-operator-webhook-server-cert" not found Dec 06 03:55:02 crc kubenswrapper[4802]: I1206 03:55:02.807404 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-4x9rb"] Dec 06 03:55:02 crc kubenswrapper[4802]: I1206 03:55:02.821695 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-8q2cg"] Dec 06 03:55:02 crc kubenswrapper[4802]: I1206 03:55:02.822226 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/05736c59-c4db-4cfe-addd-1b29c552596e-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4pwn82\" (UID: \"05736c59-c4db-4cfe-addd-1b29c552596e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4pwn82" Dec 06 03:55:02 crc kubenswrapper[4802]: E1206 03:55:02.822531 4802 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 03:55:02 crc kubenswrapper[4802]: E1206 03:55:02.822657 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/05736c59-c4db-4cfe-addd-1b29c552596e-cert podName:05736c59-c4db-4cfe-addd-1b29c552596e nodeName:}" failed. No retries permitted until 2025-12-06 03:55:04.822576462 +0000 UTC m=+897.694485624 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/05736c59-c4db-4cfe-addd-1b29c552596e-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4pwn82" (UID: "05736c59-c4db-4cfe-addd-1b29c552596e") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 03:55:02 crc kubenswrapper[4802]: W1206 03:55:02.827101 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod79efe004_97e7_4586_b48e_08ce9285bb79.slice/crio-034edfa5acd74e150404989ea977530fd804d2cf9fd2db772c7a2d2e1f40b8a5 WatchSource:0}: Error finding container 034edfa5acd74e150404989ea977530fd804d2cf9fd2db772c7a2d2e1f40b8a5: Status 404 returned error can't find the container with id 034edfa5acd74e150404989ea977530fd804d2cf9fd2db772c7a2d2e1f40b8a5 Dec 06 03:55:02 crc kubenswrapper[4802]: I1206 03:55:02.843379 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-q6k7m"] Dec 06 03:55:02 crc kubenswrapper[4802]: I1206 03:55:02.850620 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-mmh24"] Dec 06 03:55:02 crc kubenswrapper[4802]: I1206 03:55:02.871441 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-nd42v"] Dec 06 03:55:02 crc kubenswrapper[4802]: I1206 03:55:02.884425 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-b5wv6"] Dec 06 03:55:02 crc kubenswrapper[4802]: I1206 03:55:02.884486 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-w64dr"] Dec 06 03:55:02 crc kubenswrapper[4802]: W1206 03:55:02.896719 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod11885e74_7bc3_4207_b046_9e94ab51e499.slice/crio-9c93d2340594d59d4fb6826bfc6f13fcbd44a349479a2d312a4a4b6a6306fd23 WatchSource:0}: Error finding container 9c93d2340594d59d4fb6826bfc6f13fcbd44a349479a2d312a4a4b6a6306fd23: Status 404 returned error can't find the container with id 9c93d2340594d59d4fb6826bfc6f13fcbd44a349479a2d312a4a4b6a6306fd23 Dec 06 03:55:02 crc kubenswrapper[4802]: W1206 03:55:02.915431 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda3bed886_a616_4733_947a_288ada88dc74.slice/crio-01b1450b304075875de812131dc7c271af7db96ec363721e729fb19ab7da16c9 WatchSource:0}: Error finding container 01b1450b304075875de812131dc7c271af7db96ec363721e729fb19ab7da16c9: Status 404 returned error can't find the container with id 01b1450b304075875de812131dc7c271af7db96ec363721e729fb19ab7da16c9 Dec 06 03:55:02 crc kubenswrapper[4802]: W1206 03:55:02.916607 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod91b87773_2f3d_4e35_9a13_708f07cb2bbf.slice/crio-08fd37cc19b756bd75b87d6c815b64e9c306c35b079ffadaadd3cf9cf35fbdd7 WatchSource:0}: Error finding container 08fd37cc19b756bd75b87d6c815b64e9c306c35b079ffadaadd3cf9cf35fbdd7: Status 404 returned error can't find the container with id 08fd37cc19b756bd75b87d6c815b64e9c306c35b079ffadaadd3cf9cf35fbdd7 Dec 06 03:55:02 crc kubenswrapper[4802]: W1206 03:55:02.917700 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode9a14170_cafd_433c_83a4_22793b21d376.slice/crio-f31d0bba11ff58d97aaac137be7d0d8fd1762dcd9053b8ecd54d24976795fc18 WatchSource:0}: Error finding container f31d0bba11ff58d97aaac137be7d0d8fd1762dcd9053b8ecd54d24976795fc18: Status 404 returned error can't find the container with id f31d0bba11ff58d97aaac137be7d0d8fd1762dcd9053b8ecd54d24976795fc18 Dec 06 03:55:02 crc kubenswrapper[4802]: W1206 03:55:02.918145 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod18ad32ab_222c_41f9_8409_90fd51178f2d.slice/crio-9e2cafc23f48340643907554bfc094852ddfe1cf280c0cac041c53b30ae5710a WatchSource:0}: Error finding container 9e2cafc23f48340643907554bfc094852ddfe1cf280c0cac041c53b30ae5710a: Status 404 returned error can't find the container with id 9e2cafc23f48340643907554bfc094852ddfe1cf280c0cac041c53b30ae5710a Dec 06 03:55:02 crc kubenswrapper[4802]: E1206 03:55:02.920019 4802 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rrtdq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-w64dr_openstack-operators(91b87773-2f3d-4e35-9a13-708f07cb2bbf): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 06 03:55:02 crc kubenswrapper[4802]: E1206 03:55:02.921668 4802 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rrtdq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-w64dr_openstack-operators(91b87773-2f3d-4e35-9a13-708f07cb2bbf): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 06 03:55:02 crc kubenswrapper[4802]: E1206 03:55:02.921829 4802 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jsnpc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-q6k7m_openstack-operators(e9a14170-cafd-433c-83a4-22793b21d376): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 06 03:55:02 crc kubenswrapper[4802]: E1206 03:55:02.923506 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-w64dr" podUID="91b87773-2f3d-4e35-9a13-708f07cb2bbf" Dec 06 03:55:02 crc kubenswrapper[4802]: E1206 03:55:02.924494 4802 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jsnpc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-q6k7m_openstack-operators(e9a14170-cafd-433c-83a4-22793b21d376): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 06 03:55:02 crc kubenswrapper[4802]: E1206 03:55:02.925709 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-q6k7m" podUID="e9a14170-cafd-433c-83a4-22793b21d376" Dec 06 03:55:02 crc kubenswrapper[4802]: E1206 03:55:02.946046 4802 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lwnnp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-76cc84c6bb-nd42v_openstack-operators(18ad32ab-222c-41f9-8409-90fd51178f2d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 06 03:55:02 crc kubenswrapper[4802]: E1206 03:55:02.948574 4802 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lwnnp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-76cc84c6bb-nd42v_openstack-operators(18ad32ab-222c-41f9-8409-90fd51178f2d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 06 03:55:02 crc kubenswrapper[4802]: E1206 03:55:02.950582 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-nd42v" podUID="18ad32ab-222c-41f9-8409-90fd51178f2d" Dec 06 03:55:03 crc kubenswrapper[4802]: I1206 03:55:03.116150 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-q6k7m" event={"ID":"e9a14170-cafd-433c-83a4-22793b21d376","Type":"ContainerStarted","Data":"f31d0bba11ff58d97aaac137be7d0d8fd1762dcd9053b8ecd54d24976795fc18"} Dec 06 03:55:03 crc kubenswrapper[4802]: I1206 03:55:03.118383 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-74c6z" event={"ID":"1d3c0479-80bf-46af-a261-42b3834807f1","Type":"ContainerStarted","Data":"f3d2319e19793e681185072d1029bc070ca9abf204a4f9bc829d46983327a656"} Dec 06 03:55:03 crc kubenswrapper[4802]: E1206 03:55:03.118733 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-q6k7m" podUID="e9a14170-cafd-433c-83a4-22793b21d376" Dec 06 03:55:03 crc kubenswrapper[4802]: I1206 03:55:03.121493 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-gf27t" event={"ID":"5f8e1071-aa69-4e21-94c9-f533d55f2cac","Type":"ContainerStarted","Data":"2028a91c37d5477d6d4a37b7b995900f570d4d9fe42148b7a88f70b2655a19e5"} Dec 06 03:55:03 crc kubenswrapper[4802]: I1206 03:55:03.123585 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-4x9rb" event={"ID":"79efe004-97e7-4586-b48e-08ce9285bb79","Type":"ContainerStarted","Data":"034edfa5acd74e150404989ea977530fd804d2cf9fd2db772c7a2d2e1f40b8a5"} Dec 06 03:55:03 crc kubenswrapper[4802]: I1206 03:55:03.124892 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-h7629" event={"ID":"c2e52d5b-b8cf-41f2-933e-af31359dc69d","Type":"ContainerStarted","Data":"6cee2b35fe0ae09b08ec7d7ab47071afcb5786071c64d01c8f20477592612f44"} Dec 06 03:55:03 crc kubenswrapper[4802]: I1206 03:55:03.126842 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-nd42v" event={"ID":"18ad32ab-222c-41f9-8409-90fd51178f2d","Type":"ContainerStarted","Data":"9e2cafc23f48340643907554bfc094852ddfe1cf280c0cac041c53b30ae5710a"} Dec 06 03:55:03 crc kubenswrapper[4802]: I1206 03:55:03.132412 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-8q2cg" event={"ID":"119bce90-0fad-424d-89cf-2ca06d7126ae","Type":"ContainerStarted","Data":"7cab504ca8192730bfe1937c740d7a0fd4b9f4c512d0e62c592bd14a205ab020"} Dec 06 03:55:03 crc kubenswrapper[4802]: E1206 03:55:03.133220 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-nd42v" podUID="18ad32ab-222c-41f9-8409-90fd51178f2d" Dec 06 03:55:03 crc kubenswrapper[4802]: I1206 03:55:03.134545 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hlx24" event={"ID":"a076799b-6631-474a-b552-9cf2261c377a","Type":"ContainerStarted","Data":"adace6ac7dd390dbcfa346c1c0670c94466e283509aa6925354cc92a3361f57c"} Dec 06 03:55:03 crc kubenswrapper[4802]: I1206 03:55:03.136258 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-lk8bf" event={"ID":"8555a40e-f3d3-4627-8867-6db5ca5ed06a","Type":"ContainerStarted","Data":"6c2807d203ecbb81a1926f017cd7e6d45233fb8a3aa26a4718f671d1630bd158"} Dec 06 03:55:03 crc kubenswrapper[4802]: I1206 03:55:03.137551 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-bkfc6" event={"ID":"4460e15d-1d9a-4814-a7d2-042c6a84057f","Type":"ContainerStarted","Data":"2a3126b6297e5a341be861b96540b47b1cab8633b1d5d0965b46ad9ac8d817bc"} Dec 06 03:55:03 crc kubenswrapper[4802]: I1206 03:55:03.146853 4802 generic.go:334] "Generic (PLEG): container finished" podID="534f227e-60b2-4149-b8d6-3762d39388e2" containerID="b36ff830a7570abceab7ba8aa95d835040c694a9f61d4fb744160a0ef8f30280" exitCode=0 Dec 06 03:55:03 crc kubenswrapper[4802]: I1206 03:55:03.146951 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tvqdm" event={"ID":"534f227e-60b2-4149-b8d6-3762d39388e2","Type":"ContainerDied","Data":"b36ff830a7570abceab7ba8aa95d835040c694a9f61d4fb744160a0ef8f30280"} Dec 06 03:55:03 crc kubenswrapper[4802]: I1206 03:55:03.149531 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-c4sr8" event={"ID":"d71b7639-ee30-4323-8be0-a16d844d259f","Type":"ContainerStarted","Data":"cb8e0f2a5d151b2387cdc29734da3e44674304298a686ba3c30e2f51f6aa53bc"} Dec 06 03:55:03 crc kubenswrapper[4802]: I1206 03:55:03.158502 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-b5wv6" event={"ID":"a3bed886-a616-4733-947a-288ada88dc74","Type":"ContainerStarted","Data":"01b1450b304075875de812131dc7c271af7db96ec363721e729fb19ab7da16c9"} Dec 06 03:55:03 crc kubenswrapper[4802]: I1206 03:55:03.162640 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-mmh24" event={"ID":"11885e74-7bc3-4207-b046-9e94ab51e499","Type":"ContainerStarted","Data":"9c93d2340594d59d4fb6826bfc6f13fcbd44a349479a2d312a4a4b6a6306fd23"} Dec 06 03:55:03 crc kubenswrapper[4802]: I1206 03:55:03.164268 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-w64dr" event={"ID":"91b87773-2f3d-4e35-9a13-708f07cb2bbf","Type":"ContainerStarted","Data":"08fd37cc19b756bd75b87d6c815b64e9c306c35b079ffadaadd3cf9cf35fbdd7"} Dec 06 03:55:03 crc kubenswrapper[4802]: I1206 03:55:03.166326 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-4m2vm" event={"ID":"ef258927-21c4-4147-8fb5-3db5d78658a3","Type":"ContainerStarted","Data":"932532fd81445277457ae5fb53432f9d70736a7b8d30b57bd455a122908355e9"} Dec 06 03:55:03 crc kubenswrapper[4802]: E1206 03:55:03.169128 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-w64dr" podUID="91b87773-2f3d-4e35-9a13-708f07cb2bbf" Dec 06 03:55:03 crc kubenswrapper[4802]: I1206 03:55:03.227496 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5f6a5e2f-8db1-4d31-8a2d-5810c5e62083-webhook-certs\") pod \"openstack-operator-controller-manager-77f4d6cd-9s876\" (UID: \"5f6a5e2f-8db1-4d31-8a2d-5810c5e62083\") " pod="openstack-operators/openstack-operator-controller-manager-77f4d6cd-9s876" Dec 06 03:55:03 crc kubenswrapper[4802]: I1206 03:55:03.227622 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5f6a5e2f-8db1-4d31-8a2d-5810c5e62083-metrics-certs\") pod \"openstack-operator-controller-manager-77f4d6cd-9s876\" (UID: \"5f6a5e2f-8db1-4d31-8a2d-5810c5e62083\") " pod="openstack-operators/openstack-operator-controller-manager-77f4d6cd-9s876" Dec 06 03:55:03 crc kubenswrapper[4802]: E1206 03:55:03.227855 4802 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 06 03:55:03 crc kubenswrapper[4802]: E1206 03:55:03.227885 4802 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 06 03:55:03 crc kubenswrapper[4802]: E1206 03:55:03.227903 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5f6a5e2f-8db1-4d31-8a2d-5810c5e62083-webhook-certs podName:5f6a5e2f-8db1-4d31-8a2d-5810c5e62083 nodeName:}" failed. No retries permitted until 2025-12-06 03:55:05.227889564 +0000 UTC m=+898.099798716 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/5f6a5e2f-8db1-4d31-8a2d-5810c5e62083-webhook-certs") pod "openstack-operator-controller-manager-77f4d6cd-9s876" (UID: "5f6a5e2f-8db1-4d31-8a2d-5810c5e62083") : secret "webhook-server-cert" not found Dec 06 03:55:03 crc kubenswrapper[4802]: E1206 03:55:03.227967 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5f6a5e2f-8db1-4d31-8a2d-5810c5e62083-metrics-certs podName:5f6a5e2f-8db1-4d31-8a2d-5810c5e62083 nodeName:}" failed. No retries permitted until 2025-12-06 03:55:05.227950596 +0000 UTC m=+898.099859748 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5f6a5e2f-8db1-4d31-8a2d-5810c5e62083-metrics-certs") pod "openstack-operator-controller-manager-77f4d6cd-9s876" (UID: "5f6a5e2f-8db1-4d31-8a2d-5810c5e62083") : secret "metrics-server-cert" not found Dec 06 03:55:04 crc kubenswrapper[4802]: E1206 03:55:04.182355 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-nd42v" podUID="18ad32ab-222c-41f9-8409-90fd51178f2d" Dec 06 03:55:04 crc kubenswrapper[4802]: E1206 03:55:04.182929 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-q6k7m" podUID="e9a14170-cafd-433c-83a4-22793b21d376" Dec 06 03:55:04 crc kubenswrapper[4802]: E1206 03:55:04.183197 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-w64dr" podUID="91b87773-2f3d-4e35-9a13-708f07cb2bbf" Dec 06 03:55:04 crc kubenswrapper[4802]: I1206 03:55:04.550526 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/19f0c68c-32a3-41e2-a8ca-5ab0dc28d748-cert\") pod \"infra-operator-controller-manager-78d48bff9d-hpsfv\" (UID: \"19f0c68c-32a3-41e2-a8ca-5ab0dc28d748\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-hpsfv" Dec 06 03:55:04 crc kubenswrapper[4802]: E1206 03:55:04.550688 4802 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 06 03:55:04 crc kubenswrapper[4802]: E1206 03:55:04.550739 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/19f0c68c-32a3-41e2-a8ca-5ab0dc28d748-cert podName:19f0c68c-32a3-41e2-a8ca-5ab0dc28d748 nodeName:}" failed. No retries permitted until 2025-12-06 03:55:08.55072125 +0000 UTC m=+901.422630412 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/19f0c68c-32a3-41e2-a8ca-5ab0dc28d748-cert") pod "infra-operator-controller-manager-78d48bff9d-hpsfv" (UID: "19f0c68c-32a3-41e2-a8ca-5ab0dc28d748") : secret "infra-operator-webhook-server-cert" not found Dec 06 03:55:04 crc kubenswrapper[4802]: I1206 03:55:04.855616 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/05736c59-c4db-4cfe-addd-1b29c552596e-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4pwn82\" (UID: \"05736c59-c4db-4cfe-addd-1b29c552596e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4pwn82" Dec 06 03:55:04 crc kubenswrapper[4802]: E1206 03:55:04.855803 4802 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 03:55:04 crc kubenswrapper[4802]: E1206 03:55:04.855855 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/05736c59-c4db-4cfe-addd-1b29c552596e-cert podName:05736c59-c4db-4cfe-addd-1b29c552596e nodeName:}" failed. No retries permitted until 2025-12-06 03:55:08.855838211 +0000 UTC m=+901.727747363 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/05736c59-c4db-4cfe-addd-1b29c552596e-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4pwn82" (UID: "05736c59-c4db-4cfe-addd-1b29c552596e") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 03:55:05 crc kubenswrapper[4802]: I1206 03:55:05.261194 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5f6a5e2f-8db1-4d31-8a2d-5810c5e62083-webhook-certs\") pod \"openstack-operator-controller-manager-77f4d6cd-9s876\" (UID: \"5f6a5e2f-8db1-4d31-8a2d-5810c5e62083\") " pod="openstack-operators/openstack-operator-controller-manager-77f4d6cd-9s876" Dec 06 03:55:05 crc kubenswrapper[4802]: E1206 03:55:05.261369 4802 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 06 03:55:05 crc kubenswrapper[4802]: E1206 03:55:05.261800 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5f6a5e2f-8db1-4d31-8a2d-5810c5e62083-webhook-certs podName:5f6a5e2f-8db1-4d31-8a2d-5810c5e62083 nodeName:}" failed. No retries permitted until 2025-12-06 03:55:09.261782959 +0000 UTC m=+902.133692111 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/5f6a5e2f-8db1-4d31-8a2d-5810c5e62083-webhook-certs") pod "openstack-operator-controller-manager-77f4d6cd-9s876" (UID: "5f6a5e2f-8db1-4d31-8a2d-5810c5e62083") : secret "webhook-server-cert" not found Dec 06 03:55:05 crc kubenswrapper[4802]: I1206 03:55:05.261899 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5f6a5e2f-8db1-4d31-8a2d-5810c5e62083-metrics-certs\") pod \"openstack-operator-controller-manager-77f4d6cd-9s876\" (UID: \"5f6a5e2f-8db1-4d31-8a2d-5810c5e62083\") " pod="openstack-operators/openstack-operator-controller-manager-77f4d6cd-9s876" Dec 06 03:55:05 crc kubenswrapper[4802]: E1206 03:55:05.262247 4802 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 06 03:55:05 crc kubenswrapper[4802]: E1206 03:55:05.262451 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5f6a5e2f-8db1-4d31-8a2d-5810c5e62083-metrics-certs podName:5f6a5e2f-8db1-4d31-8a2d-5810c5e62083 nodeName:}" failed. No retries permitted until 2025-12-06 03:55:09.262419796 +0000 UTC m=+902.134328988 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5f6a5e2f-8db1-4d31-8a2d-5810c5e62083-metrics-certs") pod "openstack-operator-controller-manager-77f4d6cd-9s876" (UID: "5f6a5e2f-8db1-4d31-8a2d-5810c5e62083") : secret "metrics-server-cert" not found Dec 06 03:55:08 crc kubenswrapper[4802]: I1206 03:55:08.607378 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/19f0c68c-32a3-41e2-a8ca-5ab0dc28d748-cert\") pod \"infra-operator-controller-manager-78d48bff9d-hpsfv\" (UID: \"19f0c68c-32a3-41e2-a8ca-5ab0dc28d748\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-hpsfv" Dec 06 03:55:08 crc kubenswrapper[4802]: E1206 03:55:08.607667 4802 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 06 03:55:08 crc kubenswrapper[4802]: E1206 03:55:08.608221 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/19f0c68c-32a3-41e2-a8ca-5ab0dc28d748-cert podName:19f0c68c-32a3-41e2-a8ca-5ab0dc28d748 nodeName:}" failed. No retries permitted until 2025-12-06 03:55:16.608191698 +0000 UTC m=+909.480100890 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/19f0c68c-32a3-41e2-a8ca-5ab0dc28d748-cert") pod "infra-operator-controller-manager-78d48bff9d-hpsfv" (UID: "19f0c68c-32a3-41e2-a8ca-5ab0dc28d748") : secret "infra-operator-webhook-server-cert" not found Dec 06 03:55:08 crc kubenswrapper[4802]: I1206 03:55:08.911966 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/05736c59-c4db-4cfe-addd-1b29c552596e-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4pwn82\" (UID: \"05736c59-c4db-4cfe-addd-1b29c552596e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4pwn82" Dec 06 03:55:08 crc kubenswrapper[4802]: E1206 03:55:08.912217 4802 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 03:55:08 crc kubenswrapper[4802]: E1206 03:55:08.912297 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/05736c59-c4db-4cfe-addd-1b29c552596e-cert podName:05736c59-c4db-4cfe-addd-1b29c552596e nodeName:}" failed. No retries permitted until 2025-12-06 03:55:16.912281042 +0000 UTC m=+909.784190194 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/05736c59-c4db-4cfe-addd-1b29c552596e-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4pwn82" (UID: "05736c59-c4db-4cfe-addd-1b29c552596e") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 03:55:09 crc kubenswrapper[4802]: I1206 03:55:09.318181 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5f6a5e2f-8db1-4d31-8a2d-5810c5e62083-webhook-certs\") pod \"openstack-operator-controller-manager-77f4d6cd-9s876\" (UID: \"5f6a5e2f-8db1-4d31-8a2d-5810c5e62083\") " pod="openstack-operators/openstack-operator-controller-manager-77f4d6cd-9s876" Dec 06 03:55:09 crc kubenswrapper[4802]: I1206 03:55:09.318262 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5f6a5e2f-8db1-4d31-8a2d-5810c5e62083-metrics-certs\") pod \"openstack-operator-controller-manager-77f4d6cd-9s876\" (UID: \"5f6a5e2f-8db1-4d31-8a2d-5810c5e62083\") " pod="openstack-operators/openstack-operator-controller-manager-77f4d6cd-9s876" Dec 06 03:55:09 crc kubenswrapper[4802]: E1206 03:55:09.318437 4802 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 06 03:55:09 crc kubenswrapper[4802]: E1206 03:55:09.318498 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5f6a5e2f-8db1-4d31-8a2d-5810c5e62083-metrics-certs podName:5f6a5e2f-8db1-4d31-8a2d-5810c5e62083 nodeName:}" failed. No retries permitted until 2025-12-06 03:55:17.318482798 +0000 UTC m=+910.190391950 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5f6a5e2f-8db1-4d31-8a2d-5810c5e62083-metrics-certs") pod "openstack-operator-controller-manager-77f4d6cd-9s876" (UID: "5f6a5e2f-8db1-4d31-8a2d-5810c5e62083") : secret "metrics-server-cert" not found Dec 06 03:55:09 crc kubenswrapper[4802]: E1206 03:55:09.318829 4802 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 06 03:55:09 crc kubenswrapper[4802]: E1206 03:55:09.318859 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5f6a5e2f-8db1-4d31-8a2d-5810c5e62083-webhook-certs podName:5f6a5e2f-8db1-4d31-8a2d-5810c5e62083 nodeName:}" failed. No retries permitted until 2025-12-06 03:55:17.318849517 +0000 UTC m=+910.190758669 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/5f6a5e2f-8db1-4d31-8a2d-5810c5e62083-webhook-certs") pod "openstack-operator-controller-manager-77f4d6cd-9s876" (UID: "5f6a5e2f-8db1-4d31-8a2d-5810c5e62083") : secret "webhook-server-cert" not found Dec 06 03:55:13 crc kubenswrapper[4802]: I1206 03:55:13.283849 4802 patch_prober.go:28] interesting pod/machine-config-daemon-zpxxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 03:55:13 crc kubenswrapper[4802]: I1206 03:55:13.284247 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 03:55:16 crc kubenswrapper[4802]: I1206 03:55:16.647271 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/19f0c68c-32a3-41e2-a8ca-5ab0dc28d748-cert\") pod \"infra-operator-controller-manager-78d48bff9d-hpsfv\" (UID: \"19f0c68c-32a3-41e2-a8ca-5ab0dc28d748\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-hpsfv" Dec 06 03:55:16 crc kubenswrapper[4802]: I1206 03:55:16.654635 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/19f0c68c-32a3-41e2-a8ca-5ab0dc28d748-cert\") pod \"infra-operator-controller-manager-78d48bff9d-hpsfv\" (UID: \"19f0c68c-32a3-41e2-a8ca-5ab0dc28d748\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-hpsfv" Dec 06 03:55:16 crc kubenswrapper[4802]: I1206 03:55:16.771518 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-jrdns" Dec 06 03:55:16 crc kubenswrapper[4802]: I1206 03:55:16.778933 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-hpsfv" Dec 06 03:55:16 crc kubenswrapper[4802]: I1206 03:55:16.951941 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/05736c59-c4db-4cfe-addd-1b29c552596e-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4pwn82\" (UID: \"05736c59-c4db-4cfe-addd-1b29c552596e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4pwn82" Dec 06 03:55:16 crc kubenswrapper[4802]: E1206 03:55:16.952147 4802 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 03:55:16 crc kubenswrapper[4802]: E1206 03:55:16.952224 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/05736c59-c4db-4cfe-addd-1b29c552596e-cert podName:05736c59-c4db-4cfe-addd-1b29c552596e nodeName:}" failed. No retries permitted until 2025-12-06 03:55:32.952207937 +0000 UTC m=+925.824117089 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/05736c59-c4db-4cfe-addd-1b29c552596e-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4pwn82" (UID: "05736c59-c4db-4cfe-addd-1b29c552596e") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 03:55:17 crc kubenswrapper[4802]: I1206 03:55:17.359695 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5f6a5e2f-8db1-4d31-8a2d-5810c5e62083-metrics-certs\") pod \"openstack-operator-controller-manager-77f4d6cd-9s876\" (UID: \"5f6a5e2f-8db1-4d31-8a2d-5810c5e62083\") " pod="openstack-operators/openstack-operator-controller-manager-77f4d6cd-9s876" Dec 06 03:55:17 crc kubenswrapper[4802]: I1206 03:55:17.359878 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5f6a5e2f-8db1-4d31-8a2d-5810c5e62083-webhook-certs\") pod \"openstack-operator-controller-manager-77f4d6cd-9s876\" (UID: \"5f6a5e2f-8db1-4d31-8a2d-5810c5e62083\") " pod="openstack-operators/openstack-operator-controller-manager-77f4d6cd-9s876" Dec 06 03:55:17 crc kubenswrapper[4802]: I1206 03:55:17.364059 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5f6a5e2f-8db1-4d31-8a2d-5810c5e62083-metrics-certs\") pod \"openstack-operator-controller-manager-77f4d6cd-9s876\" (UID: \"5f6a5e2f-8db1-4d31-8a2d-5810c5e62083\") " pod="openstack-operators/openstack-operator-controller-manager-77f4d6cd-9s876" Dec 06 03:55:17 crc kubenswrapper[4802]: I1206 03:55:17.365887 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5f6a5e2f-8db1-4d31-8a2d-5810c5e62083-webhook-certs\") pod \"openstack-operator-controller-manager-77f4d6cd-9s876\" (UID: \"5f6a5e2f-8db1-4d31-8a2d-5810c5e62083\") " pod="openstack-operators/openstack-operator-controller-manager-77f4d6cd-9s876" Dec 06 03:55:17 crc kubenswrapper[4802]: I1206 03:55:17.492374 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-krrt8" Dec 06 03:55:17 crc kubenswrapper[4802]: I1206 03:55:17.500856 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-77f4d6cd-9s876" Dec 06 03:55:30 crc kubenswrapper[4802]: E1206 03:55:30.823977 4802 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168" Dec 06 03:55:30 crc kubenswrapper[4802]: E1206 03:55:30.824787 4802 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2lpxc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-lk8bf_openstack-operators(8555a40e-f3d3-4627-8867-6db5ca5ed06a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 03:55:33 crc kubenswrapper[4802]: I1206 03:55:33.041570 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/05736c59-c4db-4cfe-addd-1b29c552596e-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4pwn82\" (UID: \"05736c59-c4db-4cfe-addd-1b29c552596e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4pwn82" Dec 06 03:55:33 crc kubenswrapper[4802]: I1206 03:55:33.051070 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/05736c59-c4db-4cfe-addd-1b29c552596e-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4pwn82\" (UID: \"05736c59-c4db-4cfe-addd-1b29c552596e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4pwn82" Dec 06 03:55:33 crc kubenswrapper[4802]: I1206 03:55:33.220514 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-n55lf" Dec 06 03:55:33 crc kubenswrapper[4802]: I1206 03:55:33.226410 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4pwn82" Dec 06 03:55:34 crc kubenswrapper[4802]: I1206 03:55:34.624015 4802 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 03:55:35 crc kubenswrapper[4802]: E1206 03:55:35.488436 4802 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 06 03:55:35 crc kubenswrapper[4802]: E1206 03:55:35.488556 4802 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gtrjk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-tvqdm_openshift-marketplace(534f227e-60b2-4149-b8d6-3762d39388e2): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 03:55:35 crc kubenswrapper[4802]: E1206 03:55:35.489791 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-tvqdm" podUID="534f227e-60b2-4149-b8d6-3762d39388e2" Dec 06 03:55:35 crc kubenswrapper[4802]: E1206 03:55:35.734637 4802 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/cinder-operator@sha256:1d60701214b39cdb0fa70bbe5710f9b131139a9f4b482c2db4058a04daefb801" Dec 06 03:55:35 crc kubenswrapper[4802]: E1206 03:55:35.734814 4802 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/cinder-operator@sha256:1d60701214b39cdb0fa70bbe5710f9b131139a9f4b482c2db4058a04daefb801,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dmfnp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-operator-controller-manager-859b6ccc6-xk9cb_openstack-operators(309f12a2-5b02-4d5f-9e28-63116ff8bb47): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 03:55:36 crc kubenswrapper[4802]: E1206 03:55:36.520884 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-tvqdm" podUID="534f227e-60b2-4149-b8d6-3762d39388e2" Dec 06 03:55:36 crc kubenswrapper[4802]: E1206 03:55:36.605209 4802 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59" Dec 06 03:55:36 crc kubenswrapper[4802]: E1206 03:55:36.605448 4802 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-p9wb4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-8q2cg_openstack-operators(119bce90-0fad-424d-89cf-2ca06d7126ae): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 03:55:36 crc kubenswrapper[4802]: I1206 03:55:36.908659 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9qvb5"] Dec 06 03:55:36 crc kubenswrapper[4802]: I1206 03:55:36.912532 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9qvb5" Dec 06 03:55:36 crc kubenswrapper[4802]: I1206 03:55:36.916448 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9qvb5"] Dec 06 03:55:37 crc kubenswrapper[4802]: I1206 03:55:37.104197 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-269sf\" (UniqueName: \"kubernetes.io/projected/193a7f59-0c01-43e8-ab6c-d93b269d71ac-kube-api-access-269sf\") pod \"redhat-operators-9qvb5\" (UID: \"193a7f59-0c01-43e8-ab6c-d93b269d71ac\") " pod="openshift-marketplace/redhat-operators-9qvb5" Dec 06 03:55:37 crc kubenswrapper[4802]: I1206 03:55:37.104268 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/193a7f59-0c01-43e8-ab6c-d93b269d71ac-utilities\") pod \"redhat-operators-9qvb5\" (UID: \"193a7f59-0c01-43e8-ab6c-d93b269d71ac\") " pod="openshift-marketplace/redhat-operators-9qvb5" Dec 06 03:55:37 crc kubenswrapper[4802]: I1206 03:55:37.104382 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/193a7f59-0c01-43e8-ab6c-d93b269d71ac-catalog-content\") pod \"redhat-operators-9qvb5\" (UID: \"193a7f59-0c01-43e8-ab6c-d93b269d71ac\") " pod="openshift-marketplace/redhat-operators-9qvb5" Dec 06 03:55:37 crc kubenswrapper[4802]: I1206 03:55:37.205345 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-269sf\" (UniqueName: \"kubernetes.io/projected/193a7f59-0c01-43e8-ab6c-d93b269d71ac-kube-api-access-269sf\") pod \"redhat-operators-9qvb5\" (UID: \"193a7f59-0c01-43e8-ab6c-d93b269d71ac\") " pod="openshift-marketplace/redhat-operators-9qvb5" Dec 06 03:55:37 crc kubenswrapper[4802]: I1206 03:55:37.205408 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/193a7f59-0c01-43e8-ab6c-d93b269d71ac-utilities\") pod \"redhat-operators-9qvb5\" (UID: \"193a7f59-0c01-43e8-ab6c-d93b269d71ac\") " pod="openshift-marketplace/redhat-operators-9qvb5" Dec 06 03:55:37 crc kubenswrapper[4802]: I1206 03:55:37.205505 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/193a7f59-0c01-43e8-ab6c-d93b269d71ac-catalog-content\") pod \"redhat-operators-9qvb5\" (UID: \"193a7f59-0c01-43e8-ab6c-d93b269d71ac\") " pod="openshift-marketplace/redhat-operators-9qvb5" Dec 06 03:55:37 crc kubenswrapper[4802]: I1206 03:55:37.206033 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/193a7f59-0c01-43e8-ab6c-d93b269d71ac-catalog-content\") pod \"redhat-operators-9qvb5\" (UID: \"193a7f59-0c01-43e8-ab6c-d93b269d71ac\") " pod="openshift-marketplace/redhat-operators-9qvb5" Dec 06 03:55:37 crc kubenswrapper[4802]: I1206 03:55:37.206138 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/193a7f59-0c01-43e8-ab6c-d93b269d71ac-utilities\") pod \"redhat-operators-9qvb5\" (UID: \"193a7f59-0c01-43e8-ab6c-d93b269d71ac\") " pod="openshift-marketplace/redhat-operators-9qvb5" Dec 06 03:55:37 crc kubenswrapper[4802]: I1206 03:55:37.222245 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-269sf\" (UniqueName: \"kubernetes.io/projected/193a7f59-0c01-43e8-ab6c-d93b269d71ac-kube-api-access-269sf\") pod \"redhat-operators-9qvb5\" (UID: \"193a7f59-0c01-43e8-ab6c-d93b269d71ac\") " pod="openshift-marketplace/redhat-operators-9qvb5" Dec 06 03:55:37 crc kubenswrapper[4802]: I1206 03:55:37.235176 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9qvb5" Dec 06 03:55:38 crc kubenswrapper[4802]: E1206 03:55:38.750450 4802 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d" Dec 06 03:55:38 crc kubenswrapper[4802]: E1206 03:55:38.750688 4802 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tsfs2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-mmh24_openstack-operators(11885e74-7bc3-4207-b046-9e94ab51e499): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 03:55:40 crc kubenswrapper[4802]: E1206 03:55:40.579207 4802 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/manila-operator@sha256:2e59cfbeefc3aff0bb0a6ae9ce2235129f5173c98dd5ee8dac229ad4895faea9" Dec 06 03:55:40 crc kubenswrapper[4802]: E1206 03:55:40.579653 4802 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:2e59cfbeefc3aff0bb0a6ae9ce2235129f5173c98dd5ee8dac229ad4895faea9,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9wgdv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-7c79b5df47-74c6z_openstack-operators(1d3c0479-80bf-46af-a261-42b3834807f1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 03:55:41 crc kubenswrapper[4802]: E1206 03:55:41.634630 4802 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f" Dec 06 03:55:41 crc kubenswrapper[4802]: E1206 03:55:41.634872 4802 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wvl2m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-4x9rb_openstack-operators(79efe004-97e7-4586-b48e-08ce9285bb79): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 03:55:42 crc kubenswrapper[4802]: E1206 03:55:42.632808 4802 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7" Dec 06 03:55:42 crc kubenswrapper[4802]: E1206 03:55:42.633232 4802 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-l2ljb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-h7629_openstack-operators(c2e52d5b-b8cf-41f2-933e-af31359dc69d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 03:55:43 crc kubenswrapper[4802]: E1206 03:55:43.180373 4802 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Dec 06 03:55:43 crc kubenswrapper[4802]: E1206 03:55:43.180529 4802 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-x4sbz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-b5wv6_openstack-operators(a3bed886-a616-4733-947a-288ada88dc74): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 03:55:43 crc kubenswrapper[4802]: E1206 03:55:43.181846 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-b5wv6" podUID="a3bed886-a616-4733-947a-288ada88dc74" Dec 06 03:55:43 crc kubenswrapper[4802]: I1206 03:55:43.283420 4802 patch_prober.go:28] interesting pod/machine-config-daemon-zpxxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 03:55:43 crc kubenswrapper[4802]: I1206 03:55:43.283480 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 03:55:43 crc kubenswrapper[4802]: E1206 03:55:43.475237 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-b5wv6" podUID="a3bed886-a616-4733-947a-288ada88dc74" Dec 06 03:55:46 crc kubenswrapper[4802]: E1206 03:55:46.558126 4802 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670" Dec 06 03:55:46 crc kubenswrapper[4802]: E1206 03:55:46.558537 4802 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xwjn9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-4m2vm_openstack-operators(ef258927-21c4-4147-8fb5-3db5d78658a3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 03:55:47 crc kubenswrapper[4802]: I1206 03:55:47.000003 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-hpsfv"] Dec 06 03:55:47 crc kubenswrapper[4802]: I1206 03:55:47.035624 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-77f4d6cd-9s876"] Dec 06 03:55:47 crc kubenswrapper[4802]: I1206 03:55:47.147018 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4pwn82"] Dec 06 03:55:47 crc kubenswrapper[4802]: I1206 03:55:47.155052 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9qvb5"] Dec 06 03:55:47 crc kubenswrapper[4802]: W1206 03:55:47.380399 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod05736c59_c4db_4cfe_addd_1b29c552596e.slice/crio-7f7b358658959f899688dbaf6a5257a5174318340fb98aa14e4942a545d5c57e WatchSource:0}: Error finding container 7f7b358658959f899688dbaf6a5257a5174318340fb98aa14e4942a545d5c57e: Status 404 returned error can't find the container with id 7f7b358658959f899688dbaf6a5257a5174318340fb98aa14e4942a545d5c57e Dec 06 03:55:47 crc kubenswrapper[4802]: I1206 03:55:47.521657 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-55cb9d4b9c-srntd" event={"ID":"faf6c2c3-b028-431f-a302-3ad181dd93ae","Type":"ContainerStarted","Data":"a7013b05c2ba515b6d0d3586436ada6fd5596e81017044b99d76c66ee0a7d18e"} Dec 06 03:55:47 crc kubenswrapper[4802]: I1206 03:55:47.532848 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-77f4d6cd-9s876" event={"ID":"5f6a5e2f-8db1-4d31-8a2d-5810c5e62083","Type":"ContainerStarted","Data":"3a241a67ff83481c23be025c1c985d693218690ce921ccdc4549f72b75f323c3"} Dec 06 03:55:47 crc kubenswrapper[4802]: I1206 03:55:47.539841 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-hpsfv" event={"ID":"19f0c68c-32a3-41e2-a8ca-5ab0dc28d748","Type":"ContainerStarted","Data":"5b7988227400783f050fbbe826f758688bdc5a30fb6722b096bc3253a5aad88c"} Dec 06 03:55:47 crc kubenswrapper[4802]: I1206 03:55:47.584359 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9qvb5" event={"ID":"193a7f59-0c01-43e8-ab6c-d93b269d71ac","Type":"ContainerStarted","Data":"2deb936e417fcf271087df422ec759b88109da6ccfd43bc1456bb7fe6a63c492"} Dec 06 03:55:47 crc kubenswrapper[4802]: I1206 03:55:47.592131 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4pwn82" event={"ID":"05736c59-c4db-4cfe-addd-1b29c552596e","Type":"ContainerStarted","Data":"7f7b358658959f899688dbaf6a5257a5174318340fb98aa14e4942a545d5c57e"} Dec 06 03:55:48 crc kubenswrapper[4802]: I1206 03:55:48.631572 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-p8kv6" event={"ID":"1c3d06b1-b999-4c82-85b5-11bdd996770c","Type":"ContainerStarted","Data":"ebac8493dcbb2827bb8b688b448685bcaef7663b0788a074fcffa9b5bbd2bd57"} Dec 06 03:55:48 crc kubenswrapper[4802]: I1206 03:55:48.633345 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-c4sr8" event={"ID":"d71b7639-ee30-4323-8be0-a16d844d259f","Type":"ContainerStarted","Data":"3cd44ae958ba1378d6ccf244c366bd9034a68dbf05371af8f3f276a43c23925a"} Dec 06 03:55:48 crc kubenswrapper[4802]: I1206 03:55:48.635107 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-gf27t" event={"ID":"5f8e1071-aa69-4e21-94c9-f533d55f2cac","Type":"ContainerStarted","Data":"68cfb7ef1c751bb0e1aec1647c964977679f7c56255a525f48c8084a897e1e3d"} Dec 06 03:55:48 crc kubenswrapper[4802]: I1206 03:55:48.636924 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-wwgt6" event={"ID":"e2bc662f-efe3-4e89-9b0e-e40eba87e41b","Type":"ContainerStarted","Data":"ef7c428633daa909ee618dde9fd890f2a41cae2c107df0367070bba17f063534"} Dec 06 03:55:48 crc kubenswrapper[4802]: I1206 03:55:48.638420 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-w64dr" event={"ID":"91b87773-2f3d-4e35-9a13-708f07cb2bbf","Type":"ContainerStarted","Data":"8968eb32e973d1c85c031abac89f4aa099b4b3ce636fa65bfad4558491516fc3"} Dec 06 03:55:48 crc kubenswrapper[4802]: I1206 03:55:48.639659 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-bkfc6" event={"ID":"4460e15d-1d9a-4814-a7d2-042c6a84057f","Type":"ContainerStarted","Data":"995b7410e562b54da08fe182329e6a5bce26629cbf2998846d57d74921e2c900"} Dec 06 03:55:48 crc kubenswrapper[4802]: I1206 03:55:48.643268 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-q6k7m" event={"ID":"e9a14170-cafd-433c-83a4-22793b21d376","Type":"ContainerStarted","Data":"0f047b3ef245f033d59b757846e025c97c688c9eac3eb56b37beb5e1719ca744"} Dec 06 03:55:48 crc kubenswrapper[4802]: I1206 03:55:48.665729 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hlx24" event={"ID":"a076799b-6631-474a-b552-9cf2261c377a","Type":"ContainerStarted","Data":"2650729e91e8994f89be529c5ff86235b67ef4737736735844eca98d4a5ff728"} Dec 06 03:55:48 crc kubenswrapper[4802]: I1206 03:55:48.674436 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-zkp5c" event={"ID":"08b2817d-5061-4375-82d4-726108586852","Type":"ContainerStarted","Data":"47f1b9102af5333af1415c6eea8432c502573cab1c0634a1d823c4ca2f079663"} Dec 06 03:55:49 crc kubenswrapper[4802]: I1206 03:55:49.689460 4802 generic.go:334] "Generic (PLEG): container finished" podID="193a7f59-0c01-43e8-ab6c-d93b269d71ac" containerID="53b988db757d38e80c37944bb725bbc667f6b0992f62497a0e4e53a5028f242e" exitCode=0 Dec 06 03:55:49 crc kubenswrapper[4802]: I1206 03:55:49.689649 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9qvb5" event={"ID":"193a7f59-0c01-43e8-ab6c-d93b269d71ac","Type":"ContainerDied","Data":"53b988db757d38e80c37944bb725bbc667f6b0992f62497a0e4e53a5028f242e"} Dec 06 03:55:49 crc kubenswrapper[4802]: I1206 03:55:49.711761 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-77f4d6cd-9s876" event={"ID":"5f6a5e2f-8db1-4d31-8a2d-5810c5e62083","Type":"ContainerStarted","Data":"3f2658e84125ac26a4f03187a5ebb93532add0232540985327590983fd4de436"} Dec 06 03:55:49 crc kubenswrapper[4802]: I1206 03:55:49.711846 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-77f4d6cd-9s876" Dec 06 03:55:49 crc kubenswrapper[4802]: I1206 03:55:49.714617 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-nd42v" event={"ID":"18ad32ab-222c-41f9-8409-90fd51178f2d","Type":"ContainerStarted","Data":"7b9372f8bb6ca0e62e1ce5c2ff7147a165acea16d9a5ab2776faf577b4ef4030"} Dec 06 03:55:49 crc kubenswrapper[4802]: I1206 03:55:49.750303 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-77f4d6cd-9s876" podStartSLOduration=48.750284522 podStartE2EDuration="48.750284522s" podCreationTimestamp="2025-12-06 03:55:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:55:49.74400048 +0000 UTC m=+942.615909632" watchObservedRunningTime="2025-12-06 03:55:49.750284522 +0000 UTC m=+942.622193674" Dec 06 03:55:54 crc kubenswrapper[4802]: I1206 03:55:54.753602 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-hpsfv" event={"ID":"19f0c68c-32a3-41e2-a8ca-5ab0dc28d748","Type":"ContainerStarted","Data":"61d90a2a0bcaddb9feb985fa9b9cc3f7493082eb2ee6a48255fdd1237b4a773a"} Dec 06 03:55:54 crc kubenswrapper[4802]: I1206 03:55:54.754843 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-mmh24" event={"ID":"11885e74-7bc3-4207-b046-9e94ab51e499","Type":"ContainerStarted","Data":"02e7722641eb9920fa10b2fccd0ed035af2f2864802d13fa043f33d5c739fa03"} Dec 06 03:55:54 crc kubenswrapper[4802]: I1206 03:55:54.756969 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9qvb5" event={"ID":"193a7f59-0c01-43e8-ab6c-d93b269d71ac","Type":"ContainerStarted","Data":"96a6c278a56b5812c9cc2b39df19fc94083099ccebd282186ceaff2f6ba90a5d"} Dec 06 03:55:54 crc kubenswrapper[4802]: I1206 03:55:54.758223 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4pwn82" event={"ID":"05736c59-c4db-4cfe-addd-1b29c552596e","Type":"ContainerStarted","Data":"9c95cc837eb37c0167bf699d6105acb077c70ba57078712e336ea5717b28e4e3"} Dec 06 03:55:54 crc kubenswrapper[4802]: I1206 03:55:54.759522 4802 generic.go:334] "Generic (PLEG): container finished" podID="534f227e-60b2-4149-b8d6-3762d39388e2" containerID="d95123dbb389064ac97e3f8ccb55e7cb85786348f268d08c43845057d96f34b2" exitCode=0 Dec 06 03:55:54 crc kubenswrapper[4802]: I1206 03:55:54.759554 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tvqdm" event={"ID":"534f227e-60b2-4149-b8d6-3762d39388e2","Type":"ContainerDied","Data":"d95123dbb389064ac97e3f8ccb55e7cb85786348f268d08c43845057d96f34b2"} Dec 06 03:55:55 crc kubenswrapper[4802]: E1206 03:55:55.070862 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-mmh24" podUID="11885e74-7bc3-4207-b046-9e94ab51e499" Dec 06 03:55:55 crc kubenswrapper[4802]: E1206 03:55:55.134416 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-xk9cb" podUID="309f12a2-5b02-4d5f-9e28-63116ff8bb47" Dec 06 03:55:55 crc kubenswrapper[4802]: E1206 03:55:55.169258 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-h7629" podUID="c2e52d5b-b8cf-41f2-933e-af31359dc69d" Dec 06 03:55:55 crc kubenswrapper[4802]: E1206 03:55:55.187744 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-8q2cg" podUID="119bce90-0fad-424d-89cf-2ca06d7126ae" Dec 06 03:55:55 crc kubenswrapper[4802]: E1206 03:55:55.187895 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-74c6z" podUID="1d3c0479-80bf-46af-a261-42b3834807f1" Dec 06 03:55:55 crc kubenswrapper[4802]: E1206 03:55:55.187926 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-4m2vm" podUID="ef258927-21c4-4147-8fb5-3db5d78658a3" Dec 06 03:55:55 crc kubenswrapper[4802]: I1206 03:55:55.686063 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-x9d2c"] Dec 06 03:55:55 crc kubenswrapper[4802]: I1206 03:55:55.688310 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x9d2c" Dec 06 03:55:55 crc kubenswrapper[4802]: I1206 03:55:55.706344 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-x9d2c"] Dec 06 03:55:55 crc kubenswrapper[4802]: I1206 03:55:55.715784 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55948738-0680-414d-8a48-e6ab9bc43d5d-catalog-content\") pod \"certified-operators-x9d2c\" (UID: \"55948738-0680-414d-8a48-e6ab9bc43d5d\") " pod="openshift-marketplace/certified-operators-x9d2c" Dec 06 03:55:55 crc kubenswrapper[4802]: I1206 03:55:55.715855 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fjpg\" (UniqueName: \"kubernetes.io/projected/55948738-0680-414d-8a48-e6ab9bc43d5d-kube-api-access-2fjpg\") pod \"certified-operators-x9d2c\" (UID: \"55948738-0680-414d-8a48-e6ab9bc43d5d\") " pod="openshift-marketplace/certified-operators-x9d2c" Dec 06 03:55:55 crc kubenswrapper[4802]: I1206 03:55:55.715876 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55948738-0680-414d-8a48-e6ab9bc43d5d-utilities\") pod \"certified-operators-x9d2c\" (UID: \"55948738-0680-414d-8a48-e6ab9bc43d5d\") " pod="openshift-marketplace/certified-operators-x9d2c" Dec 06 03:55:55 crc kubenswrapper[4802]: E1206 03:55:55.765876 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/octavia-operator-controller-manager-998648c74-lk8bf" podUID="8555a40e-f3d3-4627-8867-6db5ca5ed06a" Dec 06 03:55:55 crc kubenswrapper[4802]: I1206 03:55:55.808102 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-8q2cg" event={"ID":"119bce90-0fad-424d-89cf-2ca06d7126ae","Type":"ContainerStarted","Data":"575eedcc8de2c05b68b5c3c7524969db30f924e9ff08b7897dd45c095dff2c76"} Dec 06 03:55:55 crc kubenswrapper[4802]: I1206 03:55:55.817162 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fjpg\" (UniqueName: \"kubernetes.io/projected/55948738-0680-414d-8a48-e6ab9bc43d5d-kube-api-access-2fjpg\") pod \"certified-operators-x9d2c\" (UID: \"55948738-0680-414d-8a48-e6ab9bc43d5d\") " pod="openshift-marketplace/certified-operators-x9d2c" Dec 06 03:55:55 crc kubenswrapper[4802]: I1206 03:55:55.817261 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55948738-0680-414d-8a48-e6ab9bc43d5d-utilities\") pod \"certified-operators-x9d2c\" (UID: \"55948738-0680-414d-8a48-e6ab9bc43d5d\") " pod="openshift-marketplace/certified-operators-x9d2c" Dec 06 03:55:55 crc kubenswrapper[4802]: I1206 03:55:55.817407 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55948738-0680-414d-8a48-e6ab9bc43d5d-catalog-content\") pod \"certified-operators-x9d2c\" (UID: \"55948738-0680-414d-8a48-e6ab9bc43d5d\") " pod="openshift-marketplace/certified-operators-x9d2c" Dec 06 03:55:55 crc kubenswrapper[4802]: I1206 03:55:55.817946 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55948738-0680-414d-8a48-e6ab9bc43d5d-utilities\") pod \"certified-operators-x9d2c\" (UID: \"55948738-0680-414d-8a48-e6ab9bc43d5d\") " pod="openshift-marketplace/certified-operators-x9d2c" Dec 06 03:55:55 crc kubenswrapper[4802]: I1206 03:55:55.818068 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-p8kv6" event={"ID":"1c3d06b1-b999-4c82-85b5-11bdd996770c","Type":"ContainerStarted","Data":"a2fe8acea6cf638725a8c42e4f50a106ad27743f6553080e9f459488ed6eb0d3"} Dec 06 03:55:55 crc kubenswrapper[4802]: I1206 03:55:55.818866 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55948738-0680-414d-8a48-e6ab9bc43d5d-catalog-content\") pod \"certified-operators-x9d2c\" (UID: \"55948738-0680-414d-8a48-e6ab9bc43d5d\") " pod="openshift-marketplace/certified-operators-x9d2c" Dec 06 03:55:55 crc kubenswrapper[4802]: I1206 03:55:55.818887 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-p8kv6" Dec 06 03:55:55 crc kubenswrapper[4802]: I1206 03:55:55.819660 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-4m2vm" event={"ID":"ef258927-21c4-4147-8fb5-3db5d78658a3","Type":"ContainerStarted","Data":"c71a041404c98033d8a7821e65ab5c5610b6e00207429267cc5e033d3782dc6f"} Dec 06 03:55:55 crc kubenswrapper[4802]: I1206 03:55:55.821475 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-p8kv6" Dec 06 03:55:55 crc kubenswrapper[4802]: I1206 03:55:55.826362 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-xk9cb" event={"ID":"309f12a2-5b02-4d5f-9e28-63116ff8bb47","Type":"ContainerStarted","Data":"2b8768423e4b362129dde0607b126c32f50cab8a19a709a67378eae03615253a"} Dec 06 03:55:55 crc kubenswrapper[4802]: E1206 03:55:55.827974 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670\\\"\"" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-4m2vm" podUID="ef258927-21c4-4147-8fb5-3db5d78658a3" Dec 06 03:55:55 crc kubenswrapper[4802]: I1206 03:55:55.833286 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-h7629" event={"ID":"c2e52d5b-b8cf-41f2-933e-af31359dc69d","Type":"ContainerStarted","Data":"48038234f4a2f4f8d9dd9c7f6640756005a52bd67f141ec909190e277d39ed55"} Dec 06 03:55:55 crc kubenswrapper[4802]: I1206 03:55:55.843204 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fjpg\" (UniqueName: \"kubernetes.io/projected/55948738-0680-414d-8a48-e6ab9bc43d5d-kube-api-access-2fjpg\") pod \"certified-operators-x9d2c\" (UID: \"55948738-0680-414d-8a48-e6ab9bc43d5d\") " pod="openshift-marketplace/certified-operators-x9d2c" Dec 06 03:55:55 crc kubenswrapper[4802]: I1206 03:55:55.865408 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-lk8bf" event={"ID":"8555a40e-f3d3-4627-8867-6db5ca5ed06a","Type":"ContainerStarted","Data":"2c993dedec54dcfc7f3eb2d7e12bd9415b0183a028c98c193fb2a5ab22db44ed"} Dec 06 03:55:55 crc kubenswrapper[4802]: I1206 03:55:55.869437 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-74c6z" event={"ID":"1d3c0479-80bf-46af-a261-42b3834807f1","Type":"ContainerStarted","Data":"34b9ef06529227555aada5010d2ff1183b49de4edc7a9d7d7ff20c7c4d84de8d"} Dec 06 03:55:55 crc kubenswrapper[4802]: E1206 03:55:55.893283 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/placement-operator-controller-manager-78f8948974-4x9rb" podUID="79efe004-97e7-4586-b48e-08ce9285bb79" Dec 06 03:55:55 crc kubenswrapper[4802]: I1206 03:55:55.948083 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-p8kv6" podStartSLOduration=3.6665233 podStartE2EDuration="55.948064207s" podCreationTimestamp="2025-12-06 03:55:00 +0000 UTC" firstStartedPulling="2025-12-06 03:55:01.438542766 +0000 UTC m=+894.310451918" lastFinishedPulling="2025-12-06 03:55:53.720083673 +0000 UTC m=+946.591992825" observedRunningTime="2025-12-06 03:55:55.918482393 +0000 UTC m=+948.790391545" watchObservedRunningTime="2025-12-06 03:55:55.948064207 +0000 UTC m=+948.819973359" Dec 06 03:55:56 crc kubenswrapper[4802]: I1206 03:55:56.039528 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x9d2c" Dec 06 03:55:56 crc kubenswrapper[4802]: I1206 03:55:56.801153 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-x9d2c"] Dec 06 03:55:56 crc kubenswrapper[4802]: W1206 03:55:56.840185 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod55948738_0680_414d_8a48_e6ab9bc43d5d.slice/crio-7b9890e2386e37d834ae73025b7153ec03656c044d68d687e2aa0804e97b9114 WatchSource:0}: Error finding container 7b9890e2386e37d834ae73025b7153ec03656c044d68d687e2aa0804e97b9114: Status 404 returned error can't find the container with id 7b9890e2386e37d834ae73025b7153ec03656c044d68d687e2aa0804e97b9114 Dec 06 03:55:56 crc kubenswrapper[4802]: I1206 03:55:56.908524 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x9d2c" event={"ID":"55948738-0680-414d-8a48-e6ab9bc43d5d","Type":"ContainerStarted","Data":"7b9890e2386e37d834ae73025b7153ec03656c044d68d687e2aa0804e97b9114"} Dec 06 03:55:56 crc kubenswrapper[4802]: I1206 03:55:56.915261 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hlx24" event={"ID":"a076799b-6631-474a-b552-9cf2261c377a","Type":"ContainerStarted","Data":"4999fa0dd98feb2492a8ea339258bd444b578a798c34e6704dcbde12b9d8804f"} Dec 06 03:55:56 crc kubenswrapper[4802]: I1206 03:55:56.915686 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hlx24" Dec 06 03:55:56 crc kubenswrapper[4802]: I1206 03:55:56.918135 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hlx24" Dec 06 03:55:56 crc kubenswrapper[4802]: I1206 03:55:56.919501 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-hpsfv" event={"ID":"19f0c68c-32a3-41e2-a8ca-5ab0dc28d748","Type":"ContainerStarted","Data":"975baf4779c19d1d1b00ed33b976805c3a0c28e1590e7aabb831c8d27b477831"} Dec 06 03:55:56 crc kubenswrapper[4802]: I1206 03:55:56.920021 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-hpsfv" Dec 06 03:55:56 crc kubenswrapper[4802]: I1206 03:55:56.935571 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-nd42v" event={"ID":"18ad32ab-222c-41f9-8409-90fd51178f2d","Type":"ContainerStarted","Data":"f100e98acda7b2564c63bea5ccf4c726483c840f1d74c78a061722380d91c70b"} Dec 06 03:55:56 crc kubenswrapper[4802]: I1206 03:55:56.936062 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-nd42v" Dec 06 03:55:56 crc kubenswrapper[4802]: I1206 03:55:56.938108 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-nd42v" Dec 06 03:55:56 crc kubenswrapper[4802]: I1206 03:55:56.946792 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hlx24" podStartSLOduration=3.514248716 podStartE2EDuration="56.946777565s" podCreationTimestamp="2025-12-06 03:55:00 +0000 UTC" firstStartedPulling="2025-12-06 03:55:02.469729045 +0000 UTC m=+895.341638197" lastFinishedPulling="2025-12-06 03:55:55.902257894 +0000 UTC m=+948.774167046" observedRunningTime="2025-12-06 03:55:56.946126227 +0000 UTC m=+949.818035379" watchObservedRunningTime="2025-12-06 03:55:56.946777565 +0000 UTC m=+949.818686717" Dec 06 03:55:56 crc kubenswrapper[4802]: I1206 03:55:56.948949 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-w64dr" event={"ID":"91b87773-2f3d-4e35-9a13-708f07cb2bbf","Type":"ContainerStarted","Data":"6e3cd35cb1864c8bc0926329312e7c869f41c88c6acd3f03201f0cf1c3e99806"} Dec 06 03:55:56 crc kubenswrapper[4802]: I1206 03:55:56.949704 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-w64dr" Dec 06 03:55:56 crc kubenswrapper[4802]: I1206 03:55:56.952263 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-w64dr" Dec 06 03:55:56 crc kubenswrapper[4802]: I1206 03:55:56.958434 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-c4sr8" event={"ID":"d71b7639-ee30-4323-8be0-a16d844d259f","Type":"ContainerStarted","Data":"a45d07da860ff3210facdbc4ddf30672af5310a0095eb5a05f08fcdfcc3996c0"} Dec 06 03:55:56 crc kubenswrapper[4802]: I1206 03:55:56.959331 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-c4sr8" Dec 06 03:55:56 crc kubenswrapper[4802]: I1206 03:55:56.969236 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-c4sr8" Dec 06 03:55:56 crc kubenswrapper[4802]: I1206 03:55:56.970720 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-nd42v" podStartSLOduration=5.741754197 podStartE2EDuration="56.970711224s" podCreationTimestamp="2025-12-06 03:55:00 +0000 UTC" firstStartedPulling="2025-12-06 03:55:02.945939491 +0000 UTC m=+895.817848643" lastFinishedPulling="2025-12-06 03:55:54.174896518 +0000 UTC m=+947.046805670" observedRunningTime="2025-12-06 03:55:56.969556513 +0000 UTC m=+949.841465665" watchObservedRunningTime="2025-12-06 03:55:56.970711224 +0000 UTC m=+949.842620376" Dec 06 03:55:56 crc kubenswrapper[4802]: I1206 03:55:56.989930 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-zkp5c" event={"ID":"08b2817d-5061-4375-82d4-726108586852","Type":"ContainerStarted","Data":"f3c5d3d0faf3092ba5ee5541e8232e349b9a085257201cb2ed17d9394f92c0c6"} Dec 06 03:55:56 crc kubenswrapper[4802]: I1206 03:55:56.991581 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-zkp5c" Dec 06 03:55:56 crc kubenswrapper[4802]: I1206 03:55:56.998010 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-zkp5c" Dec 06 03:55:57 crc kubenswrapper[4802]: I1206 03:55:57.020090 4802 generic.go:334] "Generic (PLEG): container finished" podID="193a7f59-0c01-43e8-ab6c-d93b269d71ac" containerID="96a6c278a56b5812c9cc2b39df19fc94083099ccebd282186ceaff2f6ba90a5d" exitCode=0 Dec 06 03:55:57 crc kubenswrapper[4802]: I1206 03:55:57.020150 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9qvb5" event={"ID":"193a7f59-0c01-43e8-ab6c-d93b269d71ac","Type":"ContainerDied","Data":"96a6c278a56b5812c9cc2b39df19fc94083099ccebd282186ceaff2f6ba90a5d"} Dec 06 03:55:57 crc kubenswrapper[4802]: I1206 03:55:57.042219 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-hpsfv" podStartSLOduration=50.820568242 podStartE2EDuration="57.042180965s" podCreationTimestamp="2025-12-06 03:55:00 +0000 UTC" firstStartedPulling="2025-12-06 03:55:47.3610405 +0000 UTC m=+940.232949652" lastFinishedPulling="2025-12-06 03:55:53.582653223 +0000 UTC m=+946.454562375" observedRunningTime="2025-12-06 03:55:57.02951395 +0000 UTC m=+949.901423102" watchObservedRunningTime="2025-12-06 03:55:57.042180965 +0000 UTC m=+949.914090107" Dec 06 03:55:57 crc kubenswrapper[4802]: I1206 03:55:57.065002 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-bkfc6" event={"ID":"4460e15d-1d9a-4814-a7d2-042c6a84057f","Type":"ContainerStarted","Data":"64aa8950458d9318047c58cf852c2dd43c35d8bbd21c9e0f0a678bd0b8218cb5"} Dec 06 03:55:57 crc kubenswrapper[4802]: I1206 03:55:57.065879 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-bkfc6" Dec 06 03:55:57 crc kubenswrapper[4802]: I1206 03:55:57.080122 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-w64dr" podStartSLOduration=3.88859406 podStartE2EDuration="56.080109114s" podCreationTimestamp="2025-12-06 03:55:01 +0000 UTC" firstStartedPulling="2025-12-06 03:55:02.919858533 +0000 UTC m=+895.791767685" lastFinishedPulling="2025-12-06 03:55:55.111373587 +0000 UTC m=+947.983282739" observedRunningTime="2025-12-06 03:55:57.079760924 +0000 UTC m=+949.951670076" watchObservedRunningTime="2025-12-06 03:55:57.080109114 +0000 UTC m=+949.952018266" Dec 06 03:55:57 crc kubenswrapper[4802]: I1206 03:55:57.105365 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-bkfc6" Dec 06 03:55:57 crc kubenswrapper[4802]: I1206 03:55:57.106694 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-zkp5c" podStartSLOduration=4.973673139 podStartE2EDuration="57.106685025s" podCreationTimestamp="2025-12-06 03:55:00 +0000 UTC" firstStartedPulling="2025-12-06 03:55:02.04219809 +0000 UTC m=+894.914107242" lastFinishedPulling="2025-12-06 03:55:54.175209976 +0000 UTC m=+947.047119128" observedRunningTime="2025-12-06 03:55:57.105537934 +0000 UTC m=+949.977447086" watchObservedRunningTime="2025-12-06 03:55:57.106685025 +0000 UTC m=+949.978594177" Dec 06 03:55:57 crc kubenswrapper[4802]: I1206 03:55:57.146134 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tvqdm" event={"ID":"534f227e-60b2-4149-b8d6-3762d39388e2","Type":"ContainerStarted","Data":"01aafdcd102e573b85c48da693239f7311705a492633c2c0523eae25042174c4"} Dec 06 03:55:57 crc kubenswrapper[4802]: I1206 03:55:57.184209 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-c4sr8" podStartSLOduration=3.693764709 podStartE2EDuration="57.184188909s" podCreationTimestamp="2025-12-06 03:55:00 +0000 UTC" firstStartedPulling="2025-12-06 03:55:02.209986825 +0000 UTC m=+895.081895977" lastFinishedPulling="2025-12-06 03:55:55.700411025 +0000 UTC m=+948.572320177" observedRunningTime="2025-12-06 03:55:57.168213496 +0000 UTC m=+950.040122648" watchObservedRunningTime="2025-12-06 03:55:57.184188909 +0000 UTC m=+950.056098061" Dec 06 03:55:57 crc kubenswrapper[4802]: I1206 03:55:57.211682 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-bkfc6" podStartSLOduration=5.96475005 podStartE2EDuration="57.211662024s" podCreationTimestamp="2025-12-06 03:55:00 +0000 UTC" firstStartedPulling="2025-12-06 03:55:02.477066005 +0000 UTC m=+895.348975157" lastFinishedPulling="2025-12-06 03:55:53.723977959 +0000 UTC m=+946.595887131" observedRunningTime="2025-12-06 03:55:57.206799392 +0000 UTC m=+950.078708554" watchObservedRunningTime="2025-12-06 03:55:57.211662024 +0000 UTC m=+950.083571176" Dec 06 03:55:57 crc kubenswrapper[4802]: I1206 03:55:57.212606 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-55cb9d4b9c-srntd" event={"ID":"faf6c2c3-b028-431f-a302-3ad181dd93ae","Type":"ContainerStarted","Data":"76a1d48952c55d3f65104f281f9c4593d5ca563b78733ad5e3cc5b4301b303e9"} Dec 06 03:55:57 crc kubenswrapper[4802]: I1206 03:55:57.225307 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-55cb9d4b9c-srntd" Dec 06 03:55:57 crc kubenswrapper[4802]: I1206 03:55:57.225360 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-55cb9d4b9c-srntd" Dec 06 03:55:57 crc kubenswrapper[4802]: I1206 03:55:57.244359 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-wwgt6" event={"ID":"e2bc662f-efe3-4e89-9b0e-e40eba87e41b","Type":"ContainerStarted","Data":"ef9a84c23fa5d019eb90ee6b9324620fd603dbe718a31fd4d4e3523c6a0b892b"} Dec 06 03:55:57 crc kubenswrapper[4802]: I1206 03:55:57.244407 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-wwgt6" Dec 06 03:55:57 crc kubenswrapper[4802]: I1206 03:55:57.251694 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-wwgt6" Dec 06 03:55:57 crc kubenswrapper[4802]: I1206 03:55:57.289201 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-55cb9d4b9c-srntd" podStartSLOduration=5.619038957 podStartE2EDuration="57.289176089s" podCreationTimestamp="2025-12-06 03:55:00 +0000 UTC" firstStartedPulling="2025-12-06 03:55:02.050013863 +0000 UTC m=+894.921923015" lastFinishedPulling="2025-12-06 03:55:53.720150985 +0000 UTC m=+946.592060147" observedRunningTime="2025-12-06 03:55:57.250685154 +0000 UTC m=+950.122594306" watchObservedRunningTime="2025-12-06 03:55:57.289176089 +0000 UTC m=+950.161085251" Dec 06 03:55:57 crc kubenswrapper[4802]: I1206 03:55:57.295348 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4pwn82" event={"ID":"05736c59-c4db-4cfe-addd-1b29c552596e","Type":"ContainerStarted","Data":"dffd4180d998107dffe2db51d0823c74f307c04e766797716702703064f68f8d"} Dec 06 03:55:57 crc kubenswrapper[4802]: I1206 03:55:57.295724 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4pwn82" Dec 06 03:55:57 crc kubenswrapper[4802]: I1206 03:55:57.316654 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-q6k7m" event={"ID":"e9a14170-cafd-433c-83a4-22793b21d376","Type":"ContainerStarted","Data":"c1f96eb17adbca90ef12933f86a62aaccb402090f51f2d42c48b456a7212babe"} Dec 06 03:55:57 crc kubenswrapper[4802]: I1206 03:55:57.319306 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-q6k7m" Dec 06 03:55:57 crc kubenswrapper[4802]: I1206 03:55:57.333758 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-wwgt6" podStartSLOduration=3.1152279959999998 podStartE2EDuration="57.333696527s" podCreationTimestamp="2025-12-06 03:55:00 +0000 UTC" firstStartedPulling="2025-12-06 03:55:01.685364615 +0000 UTC m=+894.557273767" lastFinishedPulling="2025-12-06 03:55:55.903833146 +0000 UTC m=+948.775742298" observedRunningTime="2025-12-06 03:55:57.298556573 +0000 UTC m=+950.170465735" watchObservedRunningTime="2025-12-06 03:55:57.333696527 +0000 UTC m=+950.205605669" Dec 06 03:55:57 crc kubenswrapper[4802]: I1206 03:55:57.353468 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-q6k7m" Dec 06 03:55:57 crc kubenswrapper[4802]: I1206 03:55:57.381823 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-gf27t" Dec 06 03:55:57 crc kubenswrapper[4802]: I1206 03:55:57.381855 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-gf27t" event={"ID":"5f8e1071-aa69-4e21-94c9-f533d55f2cac","Type":"ContainerStarted","Data":"d4cd65e10950cf736dc418447952d4dcdf6eeac0e7d2ad68292bb281dd5d224c"} Dec 06 03:55:57 crc kubenswrapper[4802]: I1206 03:55:57.398892 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-gf27t" Dec 06 03:55:57 crc kubenswrapper[4802]: I1206 03:55:57.410960 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-4x9rb" event={"ID":"79efe004-97e7-4586-b48e-08ce9285bb79","Type":"ContainerStarted","Data":"97a9638745efa72a9d48cc53d5c8286c84172b11bdd5be1a15b03315bd7e0892"} Dec 06 03:55:57 crc kubenswrapper[4802]: I1206 03:55:57.467236 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4pwn82" podStartSLOduration=51.201017957 podStartE2EDuration="57.467217871s" podCreationTimestamp="2025-12-06 03:55:00 +0000 UTC" firstStartedPulling="2025-12-06 03:55:47.382661366 +0000 UTC m=+940.254570518" lastFinishedPulling="2025-12-06 03:55:53.64886128 +0000 UTC m=+946.520770432" observedRunningTime="2025-12-06 03:55:57.463220713 +0000 UTC m=+950.335129865" watchObservedRunningTime="2025-12-06 03:55:57.467217871 +0000 UTC m=+950.339127023" Dec 06 03:55:57 crc kubenswrapper[4802]: I1206 03:55:57.491332 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-q6k7m" podStartSLOduration=5.369438974 podStartE2EDuration="56.491315875s" podCreationTimestamp="2025-12-06 03:55:01 +0000 UTC" firstStartedPulling="2025-12-06 03:55:02.921742974 +0000 UTC m=+895.793652126" lastFinishedPulling="2025-12-06 03:55:54.043619875 +0000 UTC m=+946.915529027" observedRunningTime="2025-12-06 03:55:57.490083651 +0000 UTC m=+950.361992803" watchObservedRunningTime="2025-12-06 03:55:57.491315875 +0000 UTC m=+950.363225027" Dec 06 03:55:57 crc kubenswrapper[4802]: I1206 03:55:57.514122 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-77f4d6cd-9s876" Dec 06 03:55:57 crc kubenswrapper[4802]: I1206 03:55:57.551370 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-gf27t" podStartSLOduration=5.7386008010000005 podStartE2EDuration="57.551348054s" podCreationTimestamp="2025-12-06 03:55:00 +0000 UTC" firstStartedPulling="2025-12-06 03:55:02.230882152 +0000 UTC m=+895.102791304" lastFinishedPulling="2025-12-06 03:55:54.043629405 +0000 UTC m=+946.915538557" observedRunningTime="2025-12-06 03:55:57.543187783 +0000 UTC m=+950.415096945" watchObservedRunningTime="2025-12-06 03:55:57.551348054 +0000 UTC m=+950.423257206" Dec 06 03:55:58 crc kubenswrapper[4802]: I1206 03:55:58.417642 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9qvb5" event={"ID":"193a7f59-0c01-43e8-ab6c-d93b269d71ac","Type":"ContainerStarted","Data":"7cb52d87b3671642efc6e44afb263c6d29ae4c5f72fa81eb887f110c026cf0e4"} Dec 06 03:55:58 crc kubenswrapper[4802]: I1206 03:55:58.420553 4802 generic.go:334] "Generic (PLEG): container finished" podID="55948738-0680-414d-8a48-e6ab9bc43d5d" containerID="8e90cd42ebbb0db411416e84438ef5d9762a9e843e63ceced8e0a7e24c3950a4" exitCode=0 Dec 06 03:55:58 crc kubenswrapper[4802]: I1206 03:55:58.420596 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x9d2c" event={"ID":"55948738-0680-414d-8a48-e6ab9bc43d5d","Type":"ContainerDied","Data":"8e90cd42ebbb0db411416e84438ef5d9762a9e843e63ceced8e0a7e24c3950a4"} Dec 06 03:55:58 crc kubenswrapper[4802]: I1206 03:55:58.423190 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-xk9cb" event={"ID":"309f12a2-5b02-4d5f-9e28-63116ff8bb47","Type":"ContainerStarted","Data":"44edd902602adf23167d17686ce980ffe3e91d393d1b0f444637914570af974f"} Dec 06 03:55:58 crc kubenswrapper[4802]: I1206 03:55:58.423370 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-xk9cb" Dec 06 03:55:58 crc kubenswrapper[4802]: I1206 03:55:58.443060 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-h7629" event={"ID":"c2e52d5b-b8cf-41f2-933e-af31359dc69d","Type":"ContainerStarted","Data":"39bdaa3285364309287d34c3432a313634adc81baef326c632f6138e86bce6da"} Dec 06 03:55:58 crc kubenswrapper[4802]: I1206 03:55:58.443649 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-h7629" Dec 06 03:55:58 crc kubenswrapper[4802]: I1206 03:55:58.444973 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-lk8bf" event={"ID":"8555a40e-f3d3-4627-8867-6db5ca5ed06a","Type":"ContainerStarted","Data":"f8766f268e63514aedf647d16622abb9e23efeed515945896dc90908abcec44f"} Dec 06 03:55:58 crc kubenswrapper[4802]: I1206 03:55:58.445320 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-lk8bf" Dec 06 03:55:58 crc kubenswrapper[4802]: I1206 03:55:58.445665 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9qvb5" podStartSLOduration=15.486928479 podStartE2EDuration="22.445642949s" podCreationTimestamp="2025-12-06 03:55:36 +0000 UTC" firstStartedPulling="2025-12-06 03:55:50.619118034 +0000 UTC m=+943.491027186" lastFinishedPulling="2025-12-06 03:55:57.577832504 +0000 UTC m=+950.449741656" observedRunningTime="2025-12-06 03:55:58.441712522 +0000 UTC m=+951.313621674" watchObservedRunningTime="2025-12-06 03:55:58.445642949 +0000 UTC m=+951.317552101" Dec 06 03:55:58 crc kubenswrapper[4802]: I1206 03:55:58.446453 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-74c6z" event={"ID":"1d3c0479-80bf-46af-a261-42b3834807f1","Type":"ContainerStarted","Data":"4d72baef38b21bf00dd6a54733d6a04aebb2994bb8cacffcc23117a4c4b01a02"} Dec 06 03:55:58 crc kubenswrapper[4802]: I1206 03:55:58.446838 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-74c6z" Dec 06 03:55:58 crc kubenswrapper[4802]: I1206 03:55:58.448967 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-mmh24" event={"ID":"11885e74-7bc3-4207-b046-9e94ab51e499","Type":"ContainerStarted","Data":"41d7f36fb694093282bb5f5b5434fcefdfdafff4ccbfe4a7554c8b9143f2eefa"} Dec 06 03:55:58 crc kubenswrapper[4802]: I1206 03:55:58.449194 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-mmh24" Dec 06 03:55:58 crc kubenswrapper[4802]: I1206 03:55:58.450637 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-4x9rb" event={"ID":"79efe004-97e7-4586-b48e-08ce9285bb79","Type":"ContainerStarted","Data":"a84780344459aa7510a6e64bfa31b1903b9c44fd61ae39d1291cd74c237abd51"} Dec 06 03:55:58 crc kubenswrapper[4802]: I1206 03:55:58.450687 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-4x9rb" Dec 06 03:55:58 crc kubenswrapper[4802]: I1206 03:55:58.452666 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-8q2cg" event={"ID":"119bce90-0fad-424d-89cf-2ca06d7126ae","Type":"ContainerStarted","Data":"0f48a399065dd452bb387b58673ba5946231e3e87a42147afaecf398c1c78065"} Dec 06 03:55:58 crc kubenswrapper[4802]: I1206 03:55:58.453059 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-8q2cg" Dec 06 03:55:58 crc kubenswrapper[4802]: I1206 03:55:58.457184 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-4m2vm" event={"ID":"ef258927-21c4-4147-8fb5-3db5d78658a3","Type":"ContainerStarted","Data":"a79afbd13293bf623664f92d94bc99b4e76545474eb96d39a1cf51f7340bec84"} Dec 06 03:55:58 crc kubenswrapper[4802]: I1206 03:55:58.457545 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-4m2vm" Dec 06 03:55:58 crc kubenswrapper[4802]: I1206 03:55:58.471744 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4pwn82" Dec 06 03:55:58 crc kubenswrapper[4802]: I1206 03:55:58.471903 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-hpsfv" Dec 06 03:55:58 crc kubenswrapper[4802]: I1206 03:55:58.550585 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-xk9cb" podStartSLOduration=3.096623771 podStartE2EDuration="58.550520795s" podCreationTimestamp="2025-12-06 03:55:00 +0000 UTC" firstStartedPulling="2025-12-06 03:55:01.393647977 +0000 UTC m=+894.265557129" lastFinishedPulling="2025-12-06 03:55:56.847545001 +0000 UTC m=+949.719454153" observedRunningTime="2025-12-06 03:55:58.49211629 +0000 UTC m=+951.364025442" watchObservedRunningTime="2025-12-06 03:55:58.550520795 +0000 UTC m=+951.422429947" Dec 06 03:55:58 crc kubenswrapper[4802]: I1206 03:55:58.618952 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-4x9rb" podStartSLOduration=3.393251102 podStartE2EDuration="58.618937022s" podCreationTimestamp="2025-12-06 03:55:00 +0000 UTC" firstStartedPulling="2025-12-06 03:55:02.833595132 +0000 UTC m=+895.705504284" lastFinishedPulling="2025-12-06 03:55:58.059281042 +0000 UTC m=+950.931190204" observedRunningTime="2025-12-06 03:55:58.614381939 +0000 UTC m=+951.486291091" watchObservedRunningTime="2025-12-06 03:55:58.618937022 +0000 UTC m=+951.490846174" Dec 06 03:55:58 crc kubenswrapper[4802]: I1206 03:55:58.662776 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-mmh24" podStartSLOduration=4.15030299 podStartE2EDuration="58.662758541s" podCreationTimestamp="2025-12-06 03:55:00 +0000 UTC" firstStartedPulling="2025-12-06 03:55:02.898239316 +0000 UTC m=+895.770148468" lastFinishedPulling="2025-12-06 03:55:57.410694867 +0000 UTC m=+950.282604019" observedRunningTime="2025-12-06 03:55:58.657323794 +0000 UTC m=+951.529232936" watchObservedRunningTime="2025-12-06 03:55:58.662758541 +0000 UTC m=+951.534667683" Dec 06 03:55:58 crc kubenswrapper[4802]: I1206 03:55:58.753317 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-tvqdm" podStartSLOduration=5.656935084 podStartE2EDuration="58.753287258s" podCreationTimestamp="2025-12-06 03:55:00 +0000 UTC" firstStartedPulling="2025-12-06 03:55:03.15039145 +0000 UTC m=+896.022300602" lastFinishedPulling="2025-12-06 03:55:56.246743624 +0000 UTC m=+949.118652776" observedRunningTime="2025-12-06 03:55:58.725847214 +0000 UTC m=+951.597756366" watchObservedRunningTime="2025-12-06 03:55:58.753287258 +0000 UTC m=+951.625196410" Dec 06 03:55:58 crc kubenswrapper[4802]: I1206 03:55:58.787085 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-74c6z" podStartSLOduration=3.856085836 podStartE2EDuration="58.787055715s" podCreationTimestamp="2025-12-06 03:55:00 +0000 UTC" firstStartedPulling="2025-12-06 03:55:02.47392356 +0000 UTC m=+895.345832712" lastFinishedPulling="2025-12-06 03:55:57.404893439 +0000 UTC m=+950.276802591" observedRunningTime="2025-12-06 03:55:58.754299866 +0000 UTC m=+951.626209018" watchObservedRunningTime="2025-12-06 03:55:58.787055715 +0000 UTC m=+951.658964857" Dec 06 03:55:58 crc kubenswrapper[4802]: I1206 03:55:58.802473 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-4m2vm" podStartSLOduration=3.215705244 podStartE2EDuration="58.802429153s" podCreationTimestamp="2025-12-06 03:55:00 +0000 UTC" firstStartedPulling="2025-12-06 03:55:02.471592126 +0000 UTC m=+895.343501278" lastFinishedPulling="2025-12-06 03:55:58.058316035 +0000 UTC m=+950.930225187" observedRunningTime="2025-12-06 03:55:58.786641504 +0000 UTC m=+951.658550656" watchObservedRunningTime="2025-12-06 03:55:58.802429153 +0000 UTC m=+951.674338325" Dec 06 03:55:58 crc kubenswrapper[4802]: I1206 03:55:58.843590 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-h7629" podStartSLOduration=3.923445444 podStartE2EDuration="58.84357135s" podCreationTimestamp="2025-12-06 03:55:00 +0000 UTC" firstStartedPulling="2025-12-06 03:55:02.486033368 +0000 UTC m=+895.357942520" lastFinishedPulling="2025-12-06 03:55:57.406159264 +0000 UTC m=+950.278068426" observedRunningTime="2025-12-06 03:55:58.840239618 +0000 UTC m=+951.712148770" watchObservedRunningTime="2025-12-06 03:55:58.84357135 +0000 UTC m=+951.715480502" Dec 06 03:55:58 crc kubenswrapper[4802]: I1206 03:55:58.876987 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-8q2cg" podStartSLOduration=4.932318947 podStartE2EDuration="58.876970556s" podCreationTimestamp="2025-12-06 03:55:00 +0000 UTC" firstStartedPulling="2025-12-06 03:55:02.901356691 +0000 UTC m=+895.773265843" lastFinishedPulling="2025-12-06 03:55:56.8460083 +0000 UTC m=+949.717917452" observedRunningTime="2025-12-06 03:55:58.874398216 +0000 UTC m=+951.746307368" watchObservedRunningTime="2025-12-06 03:55:58.876970556 +0000 UTC m=+951.748879708" Dec 06 03:55:58 crc kubenswrapper[4802]: I1206 03:55:58.905501 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-lk8bf" podStartSLOduration=4.534750726 podStartE2EDuration="58.90548225s" podCreationTimestamp="2025-12-06 03:55:00 +0000 UTC" firstStartedPulling="2025-12-06 03:55:02.478419631 +0000 UTC m=+895.350328783" lastFinishedPulling="2025-12-06 03:55:56.849151155 +0000 UTC m=+949.721060307" observedRunningTime="2025-12-06 03:55:58.899135447 +0000 UTC m=+951.771044599" watchObservedRunningTime="2025-12-06 03:55:58.90548225 +0000 UTC m=+951.777391402" Dec 06 03:56:00 crc kubenswrapper[4802]: I1206 03:56:00.481191 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x9d2c" event={"ID":"55948738-0680-414d-8a48-e6ab9bc43d5d","Type":"ContainerStarted","Data":"640b6f13afab03978f37668b165b3e575e3bf57eeb09ff333e0dc9991634f8d6"} Dec 06 03:56:00 crc kubenswrapper[4802]: I1206 03:56:00.484894 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-b5wv6" event={"ID":"a3bed886-a616-4733-947a-288ada88dc74","Type":"ContainerStarted","Data":"7514fa5805de7279f28aeb6b8ad141277c084732fc81a251b09f58af5615ac91"} Dec 06 03:56:00 crc kubenswrapper[4802]: I1206 03:56:00.520290 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-b5wv6" podStartSLOduration=3.334175152 podStartE2EDuration="59.5202719s" podCreationTimestamp="2025-12-06 03:55:01 +0000 UTC" firstStartedPulling="2025-12-06 03:55:02.918676811 +0000 UTC m=+895.790585963" lastFinishedPulling="2025-12-06 03:55:59.104773559 +0000 UTC m=+951.976682711" observedRunningTime="2025-12-06 03:56:00.519044057 +0000 UTC m=+953.390953209" watchObservedRunningTime="2025-12-06 03:56:00.5202719 +0000 UTC m=+953.392181052" Dec 06 03:56:01 crc kubenswrapper[4802]: I1206 03:56:01.087084 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-tvqdm" Dec 06 03:56:01 crc kubenswrapper[4802]: I1206 03:56:01.087138 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-tvqdm" Dec 06 03:56:01 crc kubenswrapper[4802]: I1206 03:56:01.157173 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-tvqdm" Dec 06 03:56:01 crc kubenswrapper[4802]: I1206 03:56:01.576853 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-tvqdm" Dec 06 03:56:02 crc kubenswrapper[4802]: I1206 03:56:02.193832 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tvqdm"] Dec 06 03:56:02 crc kubenswrapper[4802]: I1206 03:56:02.499372 4802 generic.go:334] "Generic (PLEG): container finished" podID="55948738-0680-414d-8a48-e6ab9bc43d5d" containerID="640b6f13afab03978f37668b165b3e575e3bf57eeb09ff333e0dc9991634f8d6" exitCode=0 Dec 06 03:56:02 crc kubenswrapper[4802]: I1206 03:56:02.499467 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x9d2c" event={"ID":"55948738-0680-414d-8a48-e6ab9bc43d5d","Type":"ContainerDied","Data":"640b6f13afab03978f37668b165b3e575e3bf57eeb09ff333e0dc9991634f8d6"} Dec 06 03:56:03 crc kubenswrapper[4802]: I1206 03:56:03.512330 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x9d2c" event={"ID":"55948738-0680-414d-8a48-e6ab9bc43d5d","Type":"ContainerStarted","Data":"aa80549935eb67c6790bb24a748dcdbe39019bc8f24d86f24d081df127a6de78"} Dec 06 03:56:03 crc kubenswrapper[4802]: I1206 03:56:03.512535 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-tvqdm" podUID="534f227e-60b2-4149-b8d6-3762d39388e2" containerName="registry-server" containerID="cri-o://01aafdcd102e573b85c48da693239f7311705a492633c2c0523eae25042174c4" gracePeriod=2 Dec 06 03:56:03 crc kubenswrapper[4802]: I1206 03:56:03.555703 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-x9d2c" podStartSLOduration=4.066281824 podStartE2EDuration="8.55568632s" podCreationTimestamp="2025-12-06 03:55:55 +0000 UTC" firstStartedPulling="2025-12-06 03:55:58.421873913 +0000 UTC m=+951.293783065" lastFinishedPulling="2025-12-06 03:56:02.911278409 +0000 UTC m=+955.783187561" observedRunningTime="2025-12-06 03:56:03.552949206 +0000 UTC m=+956.424858368" watchObservedRunningTime="2025-12-06 03:56:03.55568632 +0000 UTC m=+956.427595472" Dec 06 03:56:08 crc kubenswrapper[4802]: I1206 03:56:06.040974 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-x9d2c" Dec 06 03:56:08 crc kubenswrapper[4802]: I1206 03:56:06.042488 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-x9d2c" Dec 06 03:56:08 crc kubenswrapper[4802]: I1206 03:56:06.163479 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-x9d2c" Dec 06 03:56:08 crc kubenswrapper[4802]: I1206 03:56:06.538257 4802 generic.go:334] "Generic (PLEG): container finished" podID="534f227e-60b2-4149-b8d6-3762d39388e2" containerID="01aafdcd102e573b85c48da693239f7311705a492633c2c0523eae25042174c4" exitCode=0 Dec 06 03:56:08 crc kubenswrapper[4802]: I1206 03:56:06.538363 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tvqdm" event={"ID":"534f227e-60b2-4149-b8d6-3762d39388e2","Type":"ContainerDied","Data":"01aafdcd102e573b85c48da693239f7311705a492633c2c0523eae25042174c4"} Dec 06 03:56:08 crc kubenswrapper[4802]: I1206 03:56:07.236077 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9qvb5" Dec 06 03:56:08 crc kubenswrapper[4802]: I1206 03:56:07.236481 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9qvb5" Dec 06 03:56:08 crc kubenswrapper[4802]: I1206 03:56:07.296581 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9qvb5" Dec 06 03:56:08 crc kubenswrapper[4802]: I1206 03:56:07.605896 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9qvb5" Dec 06 03:56:08 crc kubenswrapper[4802]: I1206 03:56:08.807829 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-n5xnj"] Dec 06 03:56:08 crc kubenswrapper[4802]: I1206 03:56:08.809614 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n5xnj" Dec 06 03:56:08 crc kubenswrapper[4802]: I1206 03:56:08.817664 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n5xnj"] Dec 06 03:56:08 crc kubenswrapper[4802]: I1206 03:56:08.847711 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4v8jc\" (UniqueName: \"kubernetes.io/projected/66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa-kube-api-access-4v8jc\") pod \"community-operators-n5xnj\" (UID: \"66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa\") " pod="openshift-marketplace/community-operators-n5xnj" Dec 06 03:56:08 crc kubenswrapper[4802]: I1206 03:56:08.847904 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa-utilities\") pod \"community-operators-n5xnj\" (UID: \"66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa\") " pod="openshift-marketplace/community-operators-n5xnj" Dec 06 03:56:08 crc kubenswrapper[4802]: I1206 03:56:08.847956 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa-catalog-content\") pod \"community-operators-n5xnj\" (UID: \"66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa\") " pod="openshift-marketplace/community-operators-n5xnj" Dec 06 03:56:08 crc kubenswrapper[4802]: I1206 03:56:08.959686 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa-utilities\") pod \"community-operators-n5xnj\" (UID: \"66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa\") " pod="openshift-marketplace/community-operators-n5xnj" Dec 06 03:56:08 crc kubenswrapper[4802]: I1206 03:56:08.959739 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa-catalog-content\") pod \"community-operators-n5xnj\" (UID: \"66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa\") " pod="openshift-marketplace/community-operators-n5xnj" Dec 06 03:56:08 crc kubenswrapper[4802]: I1206 03:56:08.959800 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4v8jc\" (UniqueName: \"kubernetes.io/projected/66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa-kube-api-access-4v8jc\") pod \"community-operators-n5xnj\" (UID: \"66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa\") " pod="openshift-marketplace/community-operators-n5xnj" Dec 06 03:56:08 crc kubenswrapper[4802]: I1206 03:56:08.960205 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa-utilities\") pod \"community-operators-n5xnj\" (UID: \"66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa\") " pod="openshift-marketplace/community-operators-n5xnj" Dec 06 03:56:08 crc kubenswrapper[4802]: I1206 03:56:08.960384 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa-catalog-content\") pod \"community-operators-n5xnj\" (UID: \"66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa\") " pod="openshift-marketplace/community-operators-n5xnj" Dec 06 03:56:09 crc kubenswrapper[4802]: I1206 03:56:09.012150 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4v8jc\" (UniqueName: \"kubernetes.io/projected/66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa-kube-api-access-4v8jc\") pod \"community-operators-n5xnj\" (UID: \"66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa\") " pod="openshift-marketplace/community-operators-n5xnj" Dec 06 03:56:09 crc kubenswrapper[4802]: I1206 03:56:09.155893 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n5xnj" Dec 06 03:56:10 crc kubenswrapper[4802]: I1206 03:56:10.008837 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n5xnj"] Dec 06 03:56:10 crc kubenswrapper[4802]: W1206 03:56:10.010771 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod66ba9f3c_2e4d_4d46_860c_9ef14fdae3aa.slice/crio-00e7c44f859d877080b09608fcceb3067b179f86d293c8dd538a550cf8fd477b WatchSource:0}: Error finding container 00e7c44f859d877080b09608fcceb3067b179f86d293c8dd538a550cf8fd477b: Status 404 returned error can't find the container with id 00e7c44f859d877080b09608fcceb3067b179f86d293c8dd538a550cf8fd477b Dec 06 03:56:10 crc kubenswrapper[4802]: I1206 03:56:10.216484 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tvqdm" Dec 06 03:56:10 crc kubenswrapper[4802]: I1206 03:56:10.388794 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gtrjk\" (UniqueName: \"kubernetes.io/projected/534f227e-60b2-4149-b8d6-3762d39388e2-kube-api-access-gtrjk\") pod \"534f227e-60b2-4149-b8d6-3762d39388e2\" (UID: \"534f227e-60b2-4149-b8d6-3762d39388e2\") " Dec 06 03:56:10 crc kubenswrapper[4802]: I1206 03:56:10.388955 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/534f227e-60b2-4149-b8d6-3762d39388e2-utilities\") pod \"534f227e-60b2-4149-b8d6-3762d39388e2\" (UID: \"534f227e-60b2-4149-b8d6-3762d39388e2\") " Dec 06 03:56:10 crc kubenswrapper[4802]: I1206 03:56:10.389014 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/534f227e-60b2-4149-b8d6-3762d39388e2-catalog-content\") pod \"534f227e-60b2-4149-b8d6-3762d39388e2\" (UID: \"534f227e-60b2-4149-b8d6-3762d39388e2\") " Dec 06 03:56:10 crc kubenswrapper[4802]: I1206 03:56:10.389742 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/534f227e-60b2-4149-b8d6-3762d39388e2-utilities" (OuterVolumeSpecName: "utilities") pod "534f227e-60b2-4149-b8d6-3762d39388e2" (UID: "534f227e-60b2-4149-b8d6-3762d39388e2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:56:10 crc kubenswrapper[4802]: I1206 03:56:10.395607 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/534f227e-60b2-4149-b8d6-3762d39388e2-kube-api-access-gtrjk" (OuterVolumeSpecName: "kube-api-access-gtrjk") pod "534f227e-60b2-4149-b8d6-3762d39388e2" (UID: "534f227e-60b2-4149-b8d6-3762d39388e2"). InnerVolumeSpecName "kube-api-access-gtrjk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:56:10 crc kubenswrapper[4802]: I1206 03:56:10.422810 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/534f227e-60b2-4149-b8d6-3762d39388e2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "534f227e-60b2-4149-b8d6-3762d39388e2" (UID: "534f227e-60b2-4149-b8d6-3762d39388e2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:56:10 crc kubenswrapper[4802]: I1206 03:56:10.491530 4802 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/534f227e-60b2-4149-b8d6-3762d39388e2-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:10 crc kubenswrapper[4802]: I1206 03:56:10.491713 4802 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/534f227e-60b2-4149-b8d6-3762d39388e2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:10 crc kubenswrapper[4802]: I1206 03:56:10.491768 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gtrjk\" (UniqueName: \"kubernetes.io/projected/534f227e-60b2-4149-b8d6-3762d39388e2-kube-api-access-gtrjk\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:10 crc kubenswrapper[4802]: I1206 03:56:10.593696 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9qvb5"] Dec 06 03:56:10 crc kubenswrapper[4802]: I1206 03:56:10.615158 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n5xnj" event={"ID":"66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa","Type":"ContainerStarted","Data":"00e7c44f859d877080b09608fcceb3067b179f86d293c8dd538a550cf8fd477b"} Dec 06 03:56:10 crc kubenswrapper[4802]: I1206 03:56:10.619216 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tvqdm" event={"ID":"534f227e-60b2-4149-b8d6-3762d39388e2","Type":"ContainerDied","Data":"ff92ee3f82287267027418e234cbd91d2d4b2e88e37246c067858483f86ebfac"} Dec 06 03:56:10 crc kubenswrapper[4802]: I1206 03:56:10.619282 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tvqdm" Dec 06 03:56:10 crc kubenswrapper[4802]: I1206 03:56:10.619290 4802 scope.go:117] "RemoveContainer" containerID="01aafdcd102e573b85c48da693239f7311705a492633c2c0523eae25042174c4" Dec 06 03:56:10 crc kubenswrapper[4802]: I1206 03:56:10.619967 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9qvb5" podUID="193a7f59-0c01-43e8-ab6c-d93b269d71ac" containerName="registry-server" containerID="cri-o://7cb52d87b3671642efc6e44afb263c6d29ae4c5f72fa81eb887f110c026cf0e4" gracePeriod=2 Dec 06 03:56:10 crc kubenswrapper[4802]: I1206 03:56:10.677528 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tvqdm"] Dec 06 03:56:10 crc kubenswrapper[4802]: I1206 03:56:10.682680 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-tvqdm"] Dec 06 03:56:10 crc kubenswrapper[4802]: I1206 03:56:10.906025 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-xk9cb" Dec 06 03:56:11 crc kubenswrapper[4802]: I1206 03:56:11.272363 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-h7629" Dec 06 03:56:11 crc kubenswrapper[4802]: I1206 03:56:11.300375 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-74c6z" Dec 06 03:56:11 crc kubenswrapper[4802]: I1206 03:56:11.365343 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-4m2vm" Dec 06 03:56:11 crc kubenswrapper[4802]: I1206 03:56:11.390049 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-lk8bf" Dec 06 03:56:11 crc kubenswrapper[4802]: I1206 03:56:11.414206 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-8q2cg" Dec 06 03:56:11 crc kubenswrapper[4802]: I1206 03:56:11.447582 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-4x9rb" Dec 06 03:56:11 crc kubenswrapper[4802]: I1206 03:56:11.458678 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="534f227e-60b2-4149-b8d6-3762d39388e2" path="/var/lib/kubelet/pods/534f227e-60b2-4149-b8d6-3762d39388e2/volumes" Dec 06 03:56:11 crc kubenswrapper[4802]: I1206 03:56:11.468430 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-mmh24" Dec 06 03:56:12 crc kubenswrapper[4802]: E1206 03:56:12.119328 4802 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod193a7f59_0c01_43e8_ab6c_d93b269d71ac.slice/crio-7cb52d87b3671642efc6e44afb263c6d29ae4c5f72fa81eb887f110c026cf0e4.scope\": RecentStats: unable to find data in memory cache]" Dec 06 03:56:13 crc kubenswrapper[4802]: I1206 03:56:13.283594 4802 patch_prober.go:28] interesting pod/machine-config-daemon-zpxxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 03:56:13 crc kubenswrapper[4802]: I1206 03:56:13.283679 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 03:56:13 crc kubenswrapper[4802]: I1206 03:56:13.283736 4802 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" Dec 06 03:56:13 crc kubenswrapper[4802]: I1206 03:56:13.284617 4802 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e13987c26ea7eba2b08231bba48c1e97a8d8b6df1eb4749d8e02e423ceb644c9"} pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 03:56:13 crc kubenswrapper[4802]: I1206 03:56:13.284711 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" containerID="cri-o://e13987c26ea7eba2b08231bba48c1e97a8d8b6df1eb4749d8e02e423ceb644c9" gracePeriod=600 Dec 06 03:56:14 crc kubenswrapper[4802]: I1206 03:56:14.535932 4802 scope.go:117] "RemoveContainer" containerID="d95123dbb389064ac97e3f8ccb55e7cb85786348f268d08c43845057d96f34b2" Dec 06 03:56:14 crc kubenswrapper[4802]: I1206 03:56:14.615648 4802 scope.go:117] "RemoveContainer" containerID="b36ff830a7570abceab7ba8aa95d835040c694a9f61d4fb744160a0ef8f30280" Dec 06 03:56:15 crc kubenswrapper[4802]: I1206 03:56:15.670174 4802 generic.go:334] "Generic (PLEG): container finished" podID="66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa" containerID="22449ccc96a78c21ea51a33ed973838df22bdd5e3a3ca6334f607f011421ea11" exitCode=0 Dec 06 03:56:15 crc kubenswrapper[4802]: I1206 03:56:15.670427 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n5xnj" event={"ID":"66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa","Type":"ContainerDied","Data":"22449ccc96a78c21ea51a33ed973838df22bdd5e3a3ca6334f607f011421ea11"} Dec 06 03:56:15 crc kubenswrapper[4802]: I1206 03:56:15.757985 4802 generic.go:334] "Generic (PLEG): container finished" podID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerID="e13987c26ea7eba2b08231bba48c1e97a8d8b6df1eb4749d8e02e423ceb644c9" exitCode=0 Dec 06 03:56:15 crc kubenswrapper[4802]: I1206 03:56:15.758075 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" event={"ID":"6df38316-e0d3-4018-8d27-3620eba3a68d","Type":"ContainerDied","Data":"e13987c26ea7eba2b08231bba48c1e97a8d8b6df1eb4749d8e02e423ceb644c9"} Dec 06 03:56:15 crc kubenswrapper[4802]: I1206 03:56:15.758108 4802 scope.go:117] "RemoveContainer" containerID="57ad81d5a8f13aaa4dc17e3a8d71a6ae21755975dbb001e2d8346604e492cf3d" Dec 06 03:56:15 crc kubenswrapper[4802]: I1206 03:56:15.772905 4802 generic.go:334] "Generic (PLEG): container finished" podID="193a7f59-0c01-43e8-ab6c-d93b269d71ac" containerID="7cb52d87b3671642efc6e44afb263c6d29ae4c5f72fa81eb887f110c026cf0e4" exitCode=0 Dec 06 03:56:15 crc kubenswrapper[4802]: I1206 03:56:15.772953 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9qvb5" event={"ID":"193a7f59-0c01-43e8-ab6c-d93b269d71ac","Type":"ContainerDied","Data":"7cb52d87b3671642efc6e44afb263c6d29ae4c5f72fa81eb887f110c026cf0e4"} Dec 06 03:56:16 crc kubenswrapper[4802]: I1206 03:56:16.023937 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9qvb5" Dec 06 03:56:16 crc kubenswrapper[4802]: I1206 03:56:16.085031 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-x9d2c" Dec 06 03:56:16 crc kubenswrapper[4802]: I1206 03:56:16.087539 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/193a7f59-0c01-43e8-ab6c-d93b269d71ac-catalog-content\") pod \"193a7f59-0c01-43e8-ab6c-d93b269d71ac\" (UID: \"193a7f59-0c01-43e8-ab6c-d93b269d71ac\") " Dec 06 03:56:16 crc kubenswrapper[4802]: I1206 03:56:16.087698 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/193a7f59-0c01-43e8-ab6c-d93b269d71ac-utilities\") pod \"193a7f59-0c01-43e8-ab6c-d93b269d71ac\" (UID: \"193a7f59-0c01-43e8-ab6c-d93b269d71ac\") " Dec 06 03:56:16 crc kubenswrapper[4802]: I1206 03:56:16.088229 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-269sf\" (UniqueName: \"kubernetes.io/projected/193a7f59-0c01-43e8-ab6c-d93b269d71ac-kube-api-access-269sf\") pod \"193a7f59-0c01-43e8-ab6c-d93b269d71ac\" (UID: \"193a7f59-0c01-43e8-ab6c-d93b269d71ac\") " Dec 06 03:56:16 crc kubenswrapper[4802]: I1206 03:56:16.089514 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/193a7f59-0c01-43e8-ab6c-d93b269d71ac-utilities" (OuterVolumeSpecName: "utilities") pod "193a7f59-0c01-43e8-ab6c-d93b269d71ac" (UID: "193a7f59-0c01-43e8-ab6c-d93b269d71ac"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:56:16 crc kubenswrapper[4802]: I1206 03:56:16.092703 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/193a7f59-0c01-43e8-ab6c-d93b269d71ac-kube-api-access-269sf" (OuterVolumeSpecName: "kube-api-access-269sf") pod "193a7f59-0c01-43e8-ab6c-d93b269d71ac" (UID: "193a7f59-0c01-43e8-ab6c-d93b269d71ac"). InnerVolumeSpecName "kube-api-access-269sf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:56:16 crc kubenswrapper[4802]: I1206 03:56:16.190452 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/193a7f59-0c01-43e8-ab6c-d93b269d71ac-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "193a7f59-0c01-43e8-ab6c-d93b269d71ac" (UID: "193a7f59-0c01-43e8-ab6c-d93b269d71ac"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:56:16 crc kubenswrapper[4802]: I1206 03:56:16.190708 4802 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/193a7f59-0c01-43e8-ab6c-d93b269d71ac-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:16 crc kubenswrapper[4802]: I1206 03:56:16.190766 4802 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/193a7f59-0c01-43e8-ab6c-d93b269d71ac-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:16 crc kubenswrapper[4802]: I1206 03:56:16.190779 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-269sf\" (UniqueName: \"kubernetes.io/projected/193a7f59-0c01-43e8-ab6c-d93b269d71ac-kube-api-access-269sf\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:16 crc kubenswrapper[4802]: I1206 03:56:16.782523 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9qvb5" event={"ID":"193a7f59-0c01-43e8-ab6c-d93b269d71ac","Type":"ContainerDied","Data":"2deb936e417fcf271087df422ec759b88109da6ccfd43bc1456bb7fe6a63c492"} Dec 06 03:56:16 crc kubenswrapper[4802]: I1206 03:56:16.782562 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9qvb5" Dec 06 03:56:16 crc kubenswrapper[4802]: I1206 03:56:16.782570 4802 scope.go:117] "RemoveContainer" containerID="7cb52d87b3671642efc6e44afb263c6d29ae4c5f72fa81eb887f110c026cf0e4" Dec 06 03:56:16 crc kubenswrapper[4802]: I1206 03:56:16.785274 4802 generic.go:334] "Generic (PLEG): container finished" podID="66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa" containerID="b9e568af7c79aacbf4797181e11d497576752abfab9922004b5880d7204e93c7" exitCode=0 Dec 06 03:56:16 crc kubenswrapper[4802]: I1206 03:56:16.785376 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n5xnj" event={"ID":"66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa","Type":"ContainerDied","Data":"b9e568af7c79aacbf4797181e11d497576752abfab9922004b5880d7204e93c7"} Dec 06 03:56:16 crc kubenswrapper[4802]: I1206 03:56:16.791292 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" event={"ID":"6df38316-e0d3-4018-8d27-3620eba3a68d","Type":"ContainerStarted","Data":"04d0056d567a84a69f17a8c41df89172171e22968697aae49e1b7d82534da79c"} Dec 06 03:56:16 crc kubenswrapper[4802]: I1206 03:56:16.834454 4802 scope.go:117] "RemoveContainer" containerID="96a6c278a56b5812c9cc2b39df19fc94083099ccebd282186ceaff2f6ba90a5d" Dec 06 03:56:16 crc kubenswrapper[4802]: I1206 03:56:16.846168 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9qvb5"] Dec 06 03:56:16 crc kubenswrapper[4802]: I1206 03:56:16.851569 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9qvb5"] Dec 06 03:56:16 crc kubenswrapper[4802]: I1206 03:56:16.858324 4802 scope.go:117] "RemoveContainer" containerID="53b988db757d38e80c37944bb725bbc667f6b0992f62497a0e4e53a5028f242e" Dec 06 03:56:17 crc kubenswrapper[4802]: I1206 03:56:17.462418 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="193a7f59-0c01-43e8-ab6c-d93b269d71ac" path="/var/lib/kubelet/pods/193a7f59-0c01-43e8-ab6c-d93b269d71ac/volumes" Dec 06 03:56:17 crc kubenswrapper[4802]: I1206 03:56:17.811998 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n5xnj" event={"ID":"66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa","Type":"ContainerStarted","Data":"446b95da2a728878e0bbb5342f503f35d9921dc555f3c506697957799edeb012"} Dec 06 03:56:17 crc kubenswrapper[4802]: I1206 03:56:17.843792 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-n5xnj" podStartSLOduration=8.316587638 podStartE2EDuration="9.84377412s" podCreationTimestamp="2025-12-06 03:56:08 +0000 UTC" firstStartedPulling="2025-12-06 03:56:15.672700071 +0000 UTC m=+968.544609223" lastFinishedPulling="2025-12-06 03:56:17.199886523 +0000 UTC m=+970.071795705" observedRunningTime="2025-12-06 03:56:17.841069587 +0000 UTC m=+970.712978739" watchObservedRunningTime="2025-12-06 03:56:17.84377412 +0000 UTC m=+970.715683262" Dec 06 03:56:18 crc kubenswrapper[4802]: I1206 03:56:18.395601 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-x9d2c"] Dec 06 03:56:18 crc kubenswrapper[4802]: I1206 03:56:18.396110 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-x9d2c" podUID="55948738-0680-414d-8a48-e6ab9bc43d5d" containerName="registry-server" containerID="cri-o://aa80549935eb67c6790bb24a748dcdbe39019bc8f24d86f24d081df127a6de78" gracePeriod=2 Dec 06 03:56:18 crc kubenswrapper[4802]: I1206 03:56:18.805513 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x9d2c" Dec 06 03:56:18 crc kubenswrapper[4802]: I1206 03:56:18.818513 4802 generic.go:334] "Generic (PLEG): container finished" podID="55948738-0680-414d-8a48-e6ab9bc43d5d" containerID="aa80549935eb67c6790bb24a748dcdbe39019bc8f24d86f24d081df127a6de78" exitCode=0 Dec 06 03:56:18 crc kubenswrapper[4802]: I1206 03:56:18.818610 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x9d2c" event={"ID":"55948738-0680-414d-8a48-e6ab9bc43d5d","Type":"ContainerDied","Data":"aa80549935eb67c6790bb24a748dcdbe39019bc8f24d86f24d081df127a6de78"} Dec 06 03:56:18 crc kubenswrapper[4802]: I1206 03:56:18.818620 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x9d2c" Dec 06 03:56:18 crc kubenswrapper[4802]: I1206 03:56:18.818646 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x9d2c" event={"ID":"55948738-0680-414d-8a48-e6ab9bc43d5d","Type":"ContainerDied","Data":"7b9890e2386e37d834ae73025b7153ec03656c044d68d687e2aa0804e97b9114"} Dec 06 03:56:18 crc kubenswrapper[4802]: I1206 03:56:18.818668 4802 scope.go:117] "RemoveContainer" containerID="aa80549935eb67c6790bb24a748dcdbe39019bc8f24d86f24d081df127a6de78" Dec 06 03:56:18 crc kubenswrapper[4802]: I1206 03:56:18.824537 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55948738-0680-414d-8a48-e6ab9bc43d5d-utilities\") pod \"55948738-0680-414d-8a48-e6ab9bc43d5d\" (UID: \"55948738-0680-414d-8a48-e6ab9bc43d5d\") " Dec 06 03:56:18 crc kubenswrapper[4802]: I1206 03:56:18.824646 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55948738-0680-414d-8a48-e6ab9bc43d5d-catalog-content\") pod \"55948738-0680-414d-8a48-e6ab9bc43d5d\" (UID: \"55948738-0680-414d-8a48-e6ab9bc43d5d\") " Dec 06 03:56:18 crc kubenswrapper[4802]: I1206 03:56:18.824676 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2fjpg\" (UniqueName: \"kubernetes.io/projected/55948738-0680-414d-8a48-e6ab9bc43d5d-kube-api-access-2fjpg\") pod \"55948738-0680-414d-8a48-e6ab9bc43d5d\" (UID: \"55948738-0680-414d-8a48-e6ab9bc43d5d\") " Dec 06 03:56:18 crc kubenswrapper[4802]: I1206 03:56:18.825461 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55948738-0680-414d-8a48-e6ab9bc43d5d-utilities" (OuterVolumeSpecName: "utilities") pod "55948738-0680-414d-8a48-e6ab9bc43d5d" (UID: "55948738-0680-414d-8a48-e6ab9bc43d5d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:56:18 crc kubenswrapper[4802]: I1206 03:56:18.829994 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55948738-0680-414d-8a48-e6ab9bc43d5d-kube-api-access-2fjpg" (OuterVolumeSpecName: "kube-api-access-2fjpg") pod "55948738-0680-414d-8a48-e6ab9bc43d5d" (UID: "55948738-0680-414d-8a48-e6ab9bc43d5d"). InnerVolumeSpecName "kube-api-access-2fjpg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:56:18 crc kubenswrapper[4802]: I1206 03:56:18.855619 4802 scope.go:117] "RemoveContainer" containerID="640b6f13afab03978f37668b165b3e575e3bf57eeb09ff333e0dc9991634f8d6" Dec 06 03:56:18 crc kubenswrapper[4802]: I1206 03:56:18.882552 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55948738-0680-414d-8a48-e6ab9bc43d5d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "55948738-0680-414d-8a48-e6ab9bc43d5d" (UID: "55948738-0680-414d-8a48-e6ab9bc43d5d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:56:18 crc kubenswrapper[4802]: I1206 03:56:18.887422 4802 scope.go:117] "RemoveContainer" containerID="8e90cd42ebbb0db411416e84438ef5d9762a9e843e63ceced8e0a7e24c3950a4" Dec 06 03:56:18 crc kubenswrapper[4802]: I1206 03:56:18.905799 4802 scope.go:117] "RemoveContainer" containerID="aa80549935eb67c6790bb24a748dcdbe39019bc8f24d86f24d081df127a6de78" Dec 06 03:56:18 crc kubenswrapper[4802]: E1206 03:56:18.906291 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa80549935eb67c6790bb24a748dcdbe39019bc8f24d86f24d081df127a6de78\": container with ID starting with aa80549935eb67c6790bb24a748dcdbe39019bc8f24d86f24d081df127a6de78 not found: ID does not exist" containerID="aa80549935eb67c6790bb24a748dcdbe39019bc8f24d86f24d081df127a6de78" Dec 06 03:56:18 crc kubenswrapper[4802]: I1206 03:56:18.906327 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa80549935eb67c6790bb24a748dcdbe39019bc8f24d86f24d081df127a6de78"} err="failed to get container status \"aa80549935eb67c6790bb24a748dcdbe39019bc8f24d86f24d081df127a6de78\": rpc error: code = NotFound desc = could not find container \"aa80549935eb67c6790bb24a748dcdbe39019bc8f24d86f24d081df127a6de78\": container with ID starting with aa80549935eb67c6790bb24a748dcdbe39019bc8f24d86f24d081df127a6de78 not found: ID does not exist" Dec 06 03:56:18 crc kubenswrapper[4802]: I1206 03:56:18.906352 4802 scope.go:117] "RemoveContainer" containerID="640b6f13afab03978f37668b165b3e575e3bf57eeb09ff333e0dc9991634f8d6" Dec 06 03:56:18 crc kubenswrapper[4802]: E1206 03:56:18.906670 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"640b6f13afab03978f37668b165b3e575e3bf57eeb09ff333e0dc9991634f8d6\": container with ID starting with 640b6f13afab03978f37668b165b3e575e3bf57eeb09ff333e0dc9991634f8d6 not found: ID does not exist" containerID="640b6f13afab03978f37668b165b3e575e3bf57eeb09ff333e0dc9991634f8d6" Dec 06 03:56:18 crc kubenswrapper[4802]: I1206 03:56:18.906722 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"640b6f13afab03978f37668b165b3e575e3bf57eeb09ff333e0dc9991634f8d6"} err="failed to get container status \"640b6f13afab03978f37668b165b3e575e3bf57eeb09ff333e0dc9991634f8d6\": rpc error: code = NotFound desc = could not find container \"640b6f13afab03978f37668b165b3e575e3bf57eeb09ff333e0dc9991634f8d6\": container with ID starting with 640b6f13afab03978f37668b165b3e575e3bf57eeb09ff333e0dc9991634f8d6 not found: ID does not exist" Dec 06 03:56:18 crc kubenswrapper[4802]: I1206 03:56:18.906775 4802 scope.go:117] "RemoveContainer" containerID="8e90cd42ebbb0db411416e84438ef5d9762a9e843e63ceced8e0a7e24c3950a4" Dec 06 03:56:18 crc kubenswrapper[4802]: E1206 03:56:18.907305 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e90cd42ebbb0db411416e84438ef5d9762a9e843e63ceced8e0a7e24c3950a4\": container with ID starting with 8e90cd42ebbb0db411416e84438ef5d9762a9e843e63ceced8e0a7e24c3950a4 not found: ID does not exist" containerID="8e90cd42ebbb0db411416e84438ef5d9762a9e843e63ceced8e0a7e24c3950a4" Dec 06 03:56:18 crc kubenswrapper[4802]: I1206 03:56:18.907332 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e90cd42ebbb0db411416e84438ef5d9762a9e843e63ceced8e0a7e24c3950a4"} err="failed to get container status \"8e90cd42ebbb0db411416e84438ef5d9762a9e843e63ceced8e0a7e24c3950a4\": rpc error: code = NotFound desc = could not find container \"8e90cd42ebbb0db411416e84438ef5d9762a9e843e63ceced8e0a7e24c3950a4\": container with ID starting with 8e90cd42ebbb0db411416e84438ef5d9762a9e843e63ceced8e0a7e24c3950a4 not found: ID does not exist" Dec 06 03:56:18 crc kubenswrapper[4802]: I1206 03:56:18.926534 4802 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55948738-0680-414d-8a48-e6ab9bc43d5d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:18 crc kubenswrapper[4802]: I1206 03:56:18.926566 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2fjpg\" (UniqueName: \"kubernetes.io/projected/55948738-0680-414d-8a48-e6ab9bc43d5d-kube-api-access-2fjpg\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:18 crc kubenswrapper[4802]: I1206 03:56:18.926576 4802 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55948738-0680-414d-8a48-e6ab9bc43d5d-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:19 crc kubenswrapper[4802]: I1206 03:56:19.147323 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-x9d2c"] Dec 06 03:56:19 crc kubenswrapper[4802]: I1206 03:56:19.153442 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-x9d2c"] Dec 06 03:56:19 crc kubenswrapper[4802]: I1206 03:56:19.156872 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-n5xnj" Dec 06 03:56:19 crc kubenswrapper[4802]: I1206 03:56:19.156978 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-n5xnj" Dec 06 03:56:19 crc kubenswrapper[4802]: I1206 03:56:19.461033 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55948738-0680-414d-8a48-e6ab9bc43d5d" path="/var/lib/kubelet/pods/55948738-0680-414d-8a48-e6ab9bc43d5d/volumes" Dec 06 03:56:20 crc kubenswrapper[4802]: I1206 03:56:20.203402 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-n5xnj" podUID="66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa" containerName="registry-server" probeResult="failure" output=< Dec 06 03:56:20 crc kubenswrapper[4802]: timeout: failed to connect service ":50051" within 1s Dec 06 03:56:20 crc kubenswrapper[4802]: > Dec 06 03:56:27 crc kubenswrapper[4802]: I1206 03:56:27.695787 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-jg5xj"] Dec 06 03:56:27 crc kubenswrapper[4802]: E1206 03:56:27.697698 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="193a7f59-0c01-43e8-ab6c-d93b269d71ac" containerName="extract-utilities" Dec 06 03:56:27 crc kubenswrapper[4802]: I1206 03:56:27.697802 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="193a7f59-0c01-43e8-ab6c-d93b269d71ac" containerName="extract-utilities" Dec 06 03:56:27 crc kubenswrapper[4802]: E1206 03:56:27.697902 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="534f227e-60b2-4149-b8d6-3762d39388e2" containerName="extract-utilities" Dec 06 03:56:27 crc kubenswrapper[4802]: I1206 03:56:27.697991 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="534f227e-60b2-4149-b8d6-3762d39388e2" containerName="extract-utilities" Dec 06 03:56:27 crc kubenswrapper[4802]: E1206 03:56:27.698084 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55948738-0680-414d-8a48-e6ab9bc43d5d" containerName="registry-server" Dec 06 03:56:27 crc kubenswrapper[4802]: I1206 03:56:27.698157 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="55948738-0680-414d-8a48-e6ab9bc43d5d" containerName="registry-server" Dec 06 03:56:27 crc kubenswrapper[4802]: E1206 03:56:27.698246 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55948738-0680-414d-8a48-e6ab9bc43d5d" containerName="extract-content" Dec 06 03:56:27 crc kubenswrapper[4802]: I1206 03:56:27.698310 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="55948738-0680-414d-8a48-e6ab9bc43d5d" containerName="extract-content" Dec 06 03:56:27 crc kubenswrapper[4802]: E1206 03:56:27.698388 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="534f227e-60b2-4149-b8d6-3762d39388e2" containerName="registry-server" Dec 06 03:56:27 crc kubenswrapper[4802]: I1206 03:56:27.698508 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="534f227e-60b2-4149-b8d6-3762d39388e2" containerName="registry-server" Dec 06 03:56:27 crc kubenswrapper[4802]: E1206 03:56:27.698597 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="193a7f59-0c01-43e8-ab6c-d93b269d71ac" containerName="registry-server" Dec 06 03:56:27 crc kubenswrapper[4802]: I1206 03:56:27.698651 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="193a7f59-0c01-43e8-ab6c-d93b269d71ac" containerName="registry-server" Dec 06 03:56:27 crc kubenswrapper[4802]: E1206 03:56:27.698714 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55948738-0680-414d-8a48-e6ab9bc43d5d" containerName="extract-utilities" Dec 06 03:56:27 crc kubenswrapper[4802]: I1206 03:56:27.698787 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="55948738-0680-414d-8a48-e6ab9bc43d5d" containerName="extract-utilities" Dec 06 03:56:27 crc kubenswrapper[4802]: E1206 03:56:27.698876 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="534f227e-60b2-4149-b8d6-3762d39388e2" containerName="extract-content" Dec 06 03:56:27 crc kubenswrapper[4802]: I1206 03:56:27.698949 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="534f227e-60b2-4149-b8d6-3762d39388e2" containerName="extract-content" Dec 06 03:56:27 crc kubenswrapper[4802]: E1206 03:56:27.699031 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="193a7f59-0c01-43e8-ab6c-d93b269d71ac" containerName="extract-content" Dec 06 03:56:27 crc kubenswrapper[4802]: I1206 03:56:27.699104 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="193a7f59-0c01-43e8-ab6c-d93b269d71ac" containerName="extract-content" Dec 06 03:56:27 crc kubenswrapper[4802]: I1206 03:56:27.699308 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="534f227e-60b2-4149-b8d6-3762d39388e2" containerName="registry-server" Dec 06 03:56:27 crc kubenswrapper[4802]: I1206 03:56:27.699397 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="55948738-0680-414d-8a48-e6ab9bc43d5d" containerName="registry-server" Dec 06 03:56:27 crc kubenswrapper[4802]: I1206 03:56:27.699465 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="193a7f59-0c01-43e8-ab6c-d93b269d71ac" containerName="registry-server" Dec 06 03:56:27 crc kubenswrapper[4802]: I1206 03:56:27.700354 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-jg5xj" Dec 06 03:56:27 crc kubenswrapper[4802]: I1206 03:56:27.703139 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-jg5xj"] Dec 06 03:56:27 crc kubenswrapper[4802]: I1206 03:56:27.705384 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 06 03:56:27 crc kubenswrapper[4802]: I1206 03:56:27.705592 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 06 03:56:27 crc kubenswrapper[4802]: I1206 03:56:27.708325 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-p5q2n" Dec 06 03:56:27 crc kubenswrapper[4802]: I1206 03:56:27.708641 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 06 03:56:27 crc kubenswrapper[4802]: I1206 03:56:27.747066 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5rch\" (UniqueName: \"kubernetes.io/projected/cf7fcfab-7fcb-4e45-87fd-207e25fb80b1-kube-api-access-r5rch\") pod \"dnsmasq-dns-675f4bcbfc-jg5xj\" (UID: \"cf7fcfab-7fcb-4e45-87fd-207e25fb80b1\") " pod="openstack/dnsmasq-dns-675f4bcbfc-jg5xj" Dec 06 03:56:27 crc kubenswrapper[4802]: I1206 03:56:27.747353 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf7fcfab-7fcb-4e45-87fd-207e25fb80b1-config\") pod \"dnsmasq-dns-675f4bcbfc-jg5xj\" (UID: \"cf7fcfab-7fcb-4e45-87fd-207e25fb80b1\") " pod="openstack/dnsmasq-dns-675f4bcbfc-jg5xj" Dec 06 03:56:27 crc kubenswrapper[4802]: I1206 03:56:27.767463 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-pg5kn"] Dec 06 03:56:27 crc kubenswrapper[4802]: I1206 03:56:27.769979 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-pg5kn" Dec 06 03:56:27 crc kubenswrapper[4802]: I1206 03:56:27.774401 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 06 03:56:27 crc kubenswrapper[4802]: I1206 03:56:27.794447 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-pg5kn"] Dec 06 03:56:27 crc kubenswrapper[4802]: I1206 03:56:27.848234 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5rch\" (UniqueName: \"kubernetes.io/projected/cf7fcfab-7fcb-4e45-87fd-207e25fb80b1-kube-api-access-r5rch\") pod \"dnsmasq-dns-675f4bcbfc-jg5xj\" (UID: \"cf7fcfab-7fcb-4e45-87fd-207e25fb80b1\") " pod="openstack/dnsmasq-dns-675f4bcbfc-jg5xj" Dec 06 03:56:27 crc kubenswrapper[4802]: I1206 03:56:27.848288 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b463579f-6ba7-499f-b1e3-65a87f74026e-config\") pod \"dnsmasq-dns-78dd6ddcc-pg5kn\" (UID: \"b463579f-6ba7-499f-b1e3-65a87f74026e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-pg5kn" Dec 06 03:56:27 crc kubenswrapper[4802]: I1206 03:56:27.848337 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zdgr\" (UniqueName: \"kubernetes.io/projected/b463579f-6ba7-499f-b1e3-65a87f74026e-kube-api-access-5zdgr\") pod \"dnsmasq-dns-78dd6ddcc-pg5kn\" (UID: \"b463579f-6ba7-499f-b1e3-65a87f74026e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-pg5kn" Dec 06 03:56:27 crc kubenswrapper[4802]: I1206 03:56:27.848368 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf7fcfab-7fcb-4e45-87fd-207e25fb80b1-config\") pod \"dnsmasq-dns-675f4bcbfc-jg5xj\" (UID: \"cf7fcfab-7fcb-4e45-87fd-207e25fb80b1\") " pod="openstack/dnsmasq-dns-675f4bcbfc-jg5xj" Dec 06 03:56:27 crc kubenswrapper[4802]: I1206 03:56:27.848606 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b463579f-6ba7-499f-b1e3-65a87f74026e-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-pg5kn\" (UID: \"b463579f-6ba7-499f-b1e3-65a87f74026e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-pg5kn" Dec 06 03:56:27 crc kubenswrapper[4802]: I1206 03:56:27.849158 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf7fcfab-7fcb-4e45-87fd-207e25fb80b1-config\") pod \"dnsmasq-dns-675f4bcbfc-jg5xj\" (UID: \"cf7fcfab-7fcb-4e45-87fd-207e25fb80b1\") " pod="openstack/dnsmasq-dns-675f4bcbfc-jg5xj" Dec 06 03:56:27 crc kubenswrapper[4802]: I1206 03:56:27.870838 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5rch\" (UniqueName: \"kubernetes.io/projected/cf7fcfab-7fcb-4e45-87fd-207e25fb80b1-kube-api-access-r5rch\") pod \"dnsmasq-dns-675f4bcbfc-jg5xj\" (UID: \"cf7fcfab-7fcb-4e45-87fd-207e25fb80b1\") " pod="openstack/dnsmasq-dns-675f4bcbfc-jg5xj" Dec 06 03:56:27 crc kubenswrapper[4802]: I1206 03:56:27.949538 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b463579f-6ba7-499f-b1e3-65a87f74026e-config\") pod \"dnsmasq-dns-78dd6ddcc-pg5kn\" (UID: \"b463579f-6ba7-499f-b1e3-65a87f74026e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-pg5kn" Dec 06 03:56:27 crc kubenswrapper[4802]: I1206 03:56:27.949627 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zdgr\" (UniqueName: \"kubernetes.io/projected/b463579f-6ba7-499f-b1e3-65a87f74026e-kube-api-access-5zdgr\") pod \"dnsmasq-dns-78dd6ddcc-pg5kn\" (UID: \"b463579f-6ba7-499f-b1e3-65a87f74026e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-pg5kn" Dec 06 03:56:27 crc kubenswrapper[4802]: I1206 03:56:27.949698 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b463579f-6ba7-499f-b1e3-65a87f74026e-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-pg5kn\" (UID: \"b463579f-6ba7-499f-b1e3-65a87f74026e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-pg5kn" Dec 06 03:56:27 crc kubenswrapper[4802]: I1206 03:56:27.950693 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b463579f-6ba7-499f-b1e3-65a87f74026e-config\") pod \"dnsmasq-dns-78dd6ddcc-pg5kn\" (UID: \"b463579f-6ba7-499f-b1e3-65a87f74026e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-pg5kn" Dec 06 03:56:27 crc kubenswrapper[4802]: I1206 03:56:27.950767 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b463579f-6ba7-499f-b1e3-65a87f74026e-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-pg5kn\" (UID: \"b463579f-6ba7-499f-b1e3-65a87f74026e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-pg5kn" Dec 06 03:56:27 crc kubenswrapper[4802]: I1206 03:56:27.969123 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zdgr\" (UniqueName: \"kubernetes.io/projected/b463579f-6ba7-499f-b1e3-65a87f74026e-kube-api-access-5zdgr\") pod \"dnsmasq-dns-78dd6ddcc-pg5kn\" (UID: \"b463579f-6ba7-499f-b1e3-65a87f74026e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-pg5kn" Dec 06 03:56:28 crc kubenswrapper[4802]: I1206 03:56:28.054561 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-jg5xj" Dec 06 03:56:28 crc kubenswrapper[4802]: I1206 03:56:28.082812 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-pg5kn" Dec 06 03:56:28 crc kubenswrapper[4802]: I1206 03:56:28.502562 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-pg5kn"] Dec 06 03:56:28 crc kubenswrapper[4802]: I1206 03:56:28.558574 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-jg5xj"] Dec 06 03:56:28 crc kubenswrapper[4802]: W1206 03:56:28.563565 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcf7fcfab_7fcb_4e45_87fd_207e25fb80b1.slice/crio-d2c9a022b9615bf0435a6bc22c9ecfcb642ef14a8fb22374da7eb0d4dc2e968d WatchSource:0}: Error finding container d2c9a022b9615bf0435a6bc22c9ecfcb642ef14a8fb22374da7eb0d4dc2e968d: Status 404 returned error can't find the container with id d2c9a022b9615bf0435a6bc22c9ecfcb642ef14a8fb22374da7eb0d4dc2e968d Dec 06 03:56:28 crc kubenswrapper[4802]: I1206 03:56:28.915622 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-pg5kn" event={"ID":"b463579f-6ba7-499f-b1e3-65a87f74026e","Type":"ContainerStarted","Data":"34afa2baa744284f63890657da95d6001a9024dc15edca350eab649edc97ebce"} Dec 06 03:56:28 crc kubenswrapper[4802]: I1206 03:56:28.917032 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-jg5xj" event={"ID":"cf7fcfab-7fcb-4e45-87fd-207e25fb80b1","Type":"ContainerStarted","Data":"d2c9a022b9615bf0435a6bc22c9ecfcb642ef14a8fb22374da7eb0d4dc2e968d"} Dec 06 03:56:29 crc kubenswrapper[4802]: I1206 03:56:29.212081 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-n5xnj" Dec 06 03:56:29 crc kubenswrapper[4802]: I1206 03:56:29.259546 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-n5xnj" Dec 06 03:56:29 crc kubenswrapper[4802]: I1206 03:56:29.447116 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-n5xnj"] Dec 06 03:56:30 crc kubenswrapper[4802]: I1206 03:56:30.564030 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-jg5xj"] Dec 06 03:56:30 crc kubenswrapper[4802]: I1206 03:56:30.596585 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-wtbxc"] Dec 06 03:56:30 crc kubenswrapper[4802]: I1206 03:56:30.597852 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-wtbxc" Dec 06 03:56:30 crc kubenswrapper[4802]: I1206 03:56:30.620585 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-wtbxc"] Dec 06 03:56:30 crc kubenswrapper[4802]: I1206 03:56:30.698608 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e78545cd-e888-4838-9b9b-6008a2c60a40-config\") pod \"dnsmasq-dns-666b6646f7-wtbxc\" (UID: \"e78545cd-e888-4838-9b9b-6008a2c60a40\") " pod="openstack/dnsmasq-dns-666b6646f7-wtbxc" Dec 06 03:56:30 crc kubenswrapper[4802]: I1206 03:56:30.698652 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e78545cd-e888-4838-9b9b-6008a2c60a40-dns-svc\") pod \"dnsmasq-dns-666b6646f7-wtbxc\" (UID: \"e78545cd-e888-4838-9b9b-6008a2c60a40\") " pod="openstack/dnsmasq-dns-666b6646f7-wtbxc" Dec 06 03:56:30 crc kubenswrapper[4802]: I1206 03:56:30.698691 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6f5m\" (UniqueName: \"kubernetes.io/projected/e78545cd-e888-4838-9b9b-6008a2c60a40-kube-api-access-m6f5m\") pod \"dnsmasq-dns-666b6646f7-wtbxc\" (UID: \"e78545cd-e888-4838-9b9b-6008a2c60a40\") " pod="openstack/dnsmasq-dns-666b6646f7-wtbxc" Dec 06 03:56:30 crc kubenswrapper[4802]: I1206 03:56:30.804188 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e78545cd-e888-4838-9b9b-6008a2c60a40-config\") pod \"dnsmasq-dns-666b6646f7-wtbxc\" (UID: \"e78545cd-e888-4838-9b9b-6008a2c60a40\") " pod="openstack/dnsmasq-dns-666b6646f7-wtbxc" Dec 06 03:56:30 crc kubenswrapper[4802]: I1206 03:56:30.804243 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e78545cd-e888-4838-9b9b-6008a2c60a40-dns-svc\") pod \"dnsmasq-dns-666b6646f7-wtbxc\" (UID: \"e78545cd-e888-4838-9b9b-6008a2c60a40\") " pod="openstack/dnsmasq-dns-666b6646f7-wtbxc" Dec 06 03:56:30 crc kubenswrapper[4802]: I1206 03:56:30.804273 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6f5m\" (UniqueName: \"kubernetes.io/projected/e78545cd-e888-4838-9b9b-6008a2c60a40-kube-api-access-m6f5m\") pod \"dnsmasq-dns-666b6646f7-wtbxc\" (UID: \"e78545cd-e888-4838-9b9b-6008a2c60a40\") " pod="openstack/dnsmasq-dns-666b6646f7-wtbxc" Dec 06 03:56:30 crc kubenswrapper[4802]: I1206 03:56:30.805673 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e78545cd-e888-4838-9b9b-6008a2c60a40-config\") pod \"dnsmasq-dns-666b6646f7-wtbxc\" (UID: \"e78545cd-e888-4838-9b9b-6008a2c60a40\") " pod="openstack/dnsmasq-dns-666b6646f7-wtbxc" Dec 06 03:56:30 crc kubenswrapper[4802]: I1206 03:56:30.806312 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e78545cd-e888-4838-9b9b-6008a2c60a40-dns-svc\") pod \"dnsmasq-dns-666b6646f7-wtbxc\" (UID: \"e78545cd-e888-4838-9b9b-6008a2c60a40\") " pod="openstack/dnsmasq-dns-666b6646f7-wtbxc" Dec 06 03:56:30 crc kubenswrapper[4802]: I1206 03:56:30.824862 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6f5m\" (UniqueName: \"kubernetes.io/projected/e78545cd-e888-4838-9b9b-6008a2c60a40-kube-api-access-m6f5m\") pod \"dnsmasq-dns-666b6646f7-wtbxc\" (UID: \"e78545cd-e888-4838-9b9b-6008a2c60a40\") " pod="openstack/dnsmasq-dns-666b6646f7-wtbxc" Dec 06 03:56:30 crc kubenswrapper[4802]: I1206 03:56:30.879188 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-pg5kn"] Dec 06 03:56:30 crc kubenswrapper[4802]: I1206 03:56:30.901914 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-c5jjr"] Dec 06 03:56:30 crc kubenswrapper[4802]: I1206 03:56:30.902963 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-c5jjr" Dec 06 03:56:30 crc kubenswrapper[4802]: I1206 03:56:30.919286 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-c5jjr"] Dec 06 03:56:30 crc kubenswrapper[4802]: I1206 03:56:30.943112 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-wtbxc" Dec 06 03:56:30 crc kubenswrapper[4802]: I1206 03:56:30.998363 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-n5xnj" podUID="66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa" containerName="registry-server" containerID="cri-o://446b95da2a728878e0bbb5342f503f35d9921dc555f3c506697957799edeb012" gracePeriod=2 Dec 06 03:56:31 crc kubenswrapper[4802]: I1206 03:56:31.011459 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68de0f89-2871-4ae7-aa90-e002255f92b8-config\") pod \"dnsmasq-dns-57d769cc4f-c5jjr\" (UID: \"68de0f89-2871-4ae7-aa90-e002255f92b8\") " pod="openstack/dnsmasq-dns-57d769cc4f-c5jjr" Dec 06 03:56:31 crc kubenswrapper[4802]: I1206 03:56:31.012665 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68de0f89-2871-4ae7-aa90-e002255f92b8-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-c5jjr\" (UID: \"68de0f89-2871-4ae7-aa90-e002255f92b8\") " pod="openstack/dnsmasq-dns-57d769cc4f-c5jjr" Dec 06 03:56:31 crc kubenswrapper[4802]: I1206 03:56:31.013056 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-756r8\" (UniqueName: \"kubernetes.io/projected/68de0f89-2871-4ae7-aa90-e002255f92b8-kube-api-access-756r8\") pod \"dnsmasq-dns-57d769cc4f-c5jjr\" (UID: \"68de0f89-2871-4ae7-aa90-e002255f92b8\") " pod="openstack/dnsmasq-dns-57d769cc4f-c5jjr" Dec 06 03:56:31 crc kubenswrapper[4802]: I1206 03:56:31.116567 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68de0f89-2871-4ae7-aa90-e002255f92b8-config\") pod \"dnsmasq-dns-57d769cc4f-c5jjr\" (UID: \"68de0f89-2871-4ae7-aa90-e002255f92b8\") " pod="openstack/dnsmasq-dns-57d769cc4f-c5jjr" Dec 06 03:56:31 crc kubenswrapper[4802]: I1206 03:56:31.116707 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68de0f89-2871-4ae7-aa90-e002255f92b8-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-c5jjr\" (UID: \"68de0f89-2871-4ae7-aa90-e002255f92b8\") " pod="openstack/dnsmasq-dns-57d769cc4f-c5jjr" Dec 06 03:56:31 crc kubenswrapper[4802]: I1206 03:56:31.116783 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-756r8\" (UniqueName: \"kubernetes.io/projected/68de0f89-2871-4ae7-aa90-e002255f92b8-kube-api-access-756r8\") pod \"dnsmasq-dns-57d769cc4f-c5jjr\" (UID: \"68de0f89-2871-4ae7-aa90-e002255f92b8\") " pod="openstack/dnsmasq-dns-57d769cc4f-c5jjr" Dec 06 03:56:31 crc kubenswrapper[4802]: I1206 03:56:31.117883 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68de0f89-2871-4ae7-aa90-e002255f92b8-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-c5jjr\" (UID: \"68de0f89-2871-4ae7-aa90-e002255f92b8\") " pod="openstack/dnsmasq-dns-57d769cc4f-c5jjr" Dec 06 03:56:31 crc kubenswrapper[4802]: I1206 03:56:31.118475 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68de0f89-2871-4ae7-aa90-e002255f92b8-config\") pod \"dnsmasq-dns-57d769cc4f-c5jjr\" (UID: \"68de0f89-2871-4ae7-aa90-e002255f92b8\") " pod="openstack/dnsmasq-dns-57d769cc4f-c5jjr" Dec 06 03:56:31 crc kubenswrapper[4802]: I1206 03:56:31.133303 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-756r8\" (UniqueName: \"kubernetes.io/projected/68de0f89-2871-4ae7-aa90-e002255f92b8-kube-api-access-756r8\") pod \"dnsmasq-dns-57d769cc4f-c5jjr\" (UID: \"68de0f89-2871-4ae7-aa90-e002255f92b8\") " pod="openstack/dnsmasq-dns-57d769cc4f-c5jjr" Dec 06 03:56:31 crc kubenswrapper[4802]: I1206 03:56:31.226777 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-c5jjr" Dec 06 03:56:31 crc kubenswrapper[4802]: I1206 03:56:31.566797 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-wtbxc"] Dec 06 03:56:31 crc kubenswrapper[4802]: I1206 03:56:31.614812 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n5xnj" Dec 06 03:56:31 crc kubenswrapper[4802]: I1206 03:56:31.725592 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa-catalog-content\") pod \"66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa\" (UID: \"66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa\") " Dec 06 03:56:31 crc kubenswrapper[4802]: I1206 03:56:31.725678 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4v8jc\" (UniqueName: \"kubernetes.io/projected/66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa-kube-api-access-4v8jc\") pod \"66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa\" (UID: \"66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa\") " Dec 06 03:56:31 crc kubenswrapper[4802]: I1206 03:56:31.725699 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa-utilities\") pod \"66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa\" (UID: \"66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa\") " Dec 06 03:56:31 crc kubenswrapper[4802]: I1206 03:56:31.732496 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa-utilities" (OuterVolumeSpecName: "utilities") pod "66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa" (UID: "66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:56:31 crc kubenswrapper[4802]: I1206 03:56:31.734037 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa-kube-api-access-4v8jc" (OuterVolumeSpecName: "kube-api-access-4v8jc") pod "66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa" (UID: "66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa"). InnerVolumeSpecName "kube-api-access-4v8jc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:56:31 crc kubenswrapper[4802]: I1206 03:56:31.751398 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 03:56:31 crc kubenswrapper[4802]: E1206 03:56:31.751735 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa" containerName="registry-server" Dec 06 03:56:31 crc kubenswrapper[4802]: I1206 03:56:31.751801 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa" containerName="registry-server" Dec 06 03:56:31 crc kubenswrapper[4802]: E1206 03:56:31.751823 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa" containerName="extract-utilities" Dec 06 03:56:31 crc kubenswrapper[4802]: I1206 03:56:31.751830 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa" containerName="extract-utilities" Dec 06 03:56:31 crc kubenswrapper[4802]: E1206 03:56:31.751847 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa" containerName="extract-content" Dec 06 03:56:31 crc kubenswrapper[4802]: I1206 03:56:31.751853 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa" containerName="extract-content" Dec 06 03:56:31 crc kubenswrapper[4802]: I1206 03:56:31.751998 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa" containerName="registry-server" Dec 06 03:56:31 crc kubenswrapper[4802]: I1206 03:56:31.753344 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 06 03:56:31 crc kubenswrapper[4802]: I1206 03:56:31.755723 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-8mwrg" Dec 06 03:56:31 crc kubenswrapper[4802]: I1206 03:56:31.757910 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 06 03:56:31 crc kubenswrapper[4802]: I1206 03:56:31.758086 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 06 03:56:31 crc kubenswrapper[4802]: I1206 03:56:31.761898 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 06 03:56:31 crc kubenswrapper[4802]: I1206 03:56:31.762281 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 06 03:56:31 crc kubenswrapper[4802]: I1206 03:56:31.763720 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 06 03:56:31 crc kubenswrapper[4802]: I1206 03:56:31.764280 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 06 03:56:31 crc kubenswrapper[4802]: I1206 03:56:31.780661 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 03:56:31 crc kubenswrapper[4802]: I1206 03:56:31.802161 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa" (UID: "66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:56:31 crc kubenswrapper[4802]: I1206 03:56:31.827602 4802 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:31 crc kubenswrapper[4802]: I1206 03:56:31.827626 4802 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:31 crc kubenswrapper[4802]: I1206 03:56:31.827637 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4v8jc\" (UniqueName: \"kubernetes.io/projected/66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa-kube-api-access-4v8jc\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:31 crc kubenswrapper[4802]: I1206 03:56:31.855418 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-c5jjr"] Dec 06 03:56:31 crc kubenswrapper[4802]: W1206 03:56:31.864644 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod68de0f89_2871_4ae7_aa90_e002255f92b8.slice/crio-7db130bbc8e83573ea51f3edf21192810e30f562805639dd8b72344a90c1ca41 WatchSource:0}: Error finding container 7db130bbc8e83573ea51f3edf21192810e30f562805639dd8b72344a90c1ca41: Status 404 returned error can't find the container with id 7db130bbc8e83573ea51f3edf21192810e30f562805639dd8b72344a90c1ca41 Dec 06 03:56:31 crc kubenswrapper[4802]: I1206 03:56:31.929844 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"7b4f22a1-41dc-4a75-8999-53df29d2fb5d\") " pod="openstack/rabbitmq-server-0" Dec 06 03:56:31 crc kubenswrapper[4802]: I1206 03:56:31.929983 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"7b4f22a1-41dc-4a75-8999-53df29d2fb5d\") " pod="openstack/rabbitmq-server-0" Dec 06 03:56:31 crc kubenswrapper[4802]: I1206 03:56:31.930226 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"7b4f22a1-41dc-4a75-8999-53df29d2fb5d\") " pod="openstack/rabbitmq-server-0" Dec 06 03:56:31 crc kubenswrapper[4802]: I1206 03:56:31.930383 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"7b4f22a1-41dc-4a75-8999-53df29d2fb5d\") " pod="openstack/rabbitmq-server-0" Dec 06 03:56:31 crc kubenswrapper[4802]: I1206 03:56:31.930443 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-server-0\" (UID: \"7b4f22a1-41dc-4a75-8999-53df29d2fb5d\") " pod="openstack/rabbitmq-server-0" Dec 06 03:56:31 crc kubenswrapper[4802]: I1206 03:56:31.930534 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"7b4f22a1-41dc-4a75-8999-53df29d2fb5d\") " pod="openstack/rabbitmq-server-0" Dec 06 03:56:31 crc kubenswrapper[4802]: I1206 03:56:31.930663 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"7b4f22a1-41dc-4a75-8999-53df29d2fb5d\") " pod="openstack/rabbitmq-server-0" Dec 06 03:56:31 crc kubenswrapper[4802]: I1206 03:56:31.930741 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-config-data\") pod \"rabbitmq-server-0\" (UID: \"7b4f22a1-41dc-4a75-8999-53df29d2fb5d\") " pod="openstack/rabbitmq-server-0" Dec 06 03:56:31 crc kubenswrapper[4802]: I1206 03:56:31.930824 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-server-conf\") pod \"rabbitmq-server-0\" (UID: \"7b4f22a1-41dc-4a75-8999-53df29d2fb5d\") " pod="openstack/rabbitmq-server-0" Dec 06 03:56:31 crc kubenswrapper[4802]: I1206 03:56:31.930914 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjz5x\" (UniqueName: \"kubernetes.io/projected/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-kube-api-access-cjz5x\") pod \"rabbitmq-server-0\" (UID: \"7b4f22a1-41dc-4a75-8999-53df29d2fb5d\") " pod="openstack/rabbitmq-server-0" Dec 06 03:56:31 crc kubenswrapper[4802]: I1206 03:56:31.930957 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-pod-info\") pod \"rabbitmq-server-0\" (UID: \"7b4f22a1-41dc-4a75-8999-53df29d2fb5d\") " pod="openstack/rabbitmq-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.008688 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-wtbxc" event={"ID":"e78545cd-e888-4838-9b9b-6008a2c60a40","Type":"ContainerStarted","Data":"144186cebdc968c870d90cf997757a5dbce30f6d108b3c222507de2edbeb34cf"} Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.011039 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-c5jjr" event={"ID":"68de0f89-2871-4ae7-aa90-e002255f92b8","Type":"ContainerStarted","Data":"7db130bbc8e83573ea51f3edf21192810e30f562805639dd8b72344a90c1ca41"} Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.014590 4802 generic.go:334] "Generic (PLEG): container finished" podID="66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa" containerID="446b95da2a728878e0bbb5342f503f35d9921dc555f3c506697957799edeb012" exitCode=0 Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.014646 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n5xnj" event={"ID":"66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa","Type":"ContainerDied","Data":"446b95da2a728878e0bbb5342f503f35d9921dc555f3c506697957799edeb012"} Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.014715 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n5xnj" event={"ID":"66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa","Type":"ContainerDied","Data":"00e7c44f859d877080b09608fcceb3067b179f86d293c8dd538a550cf8fd477b"} Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.014734 4802 scope.go:117] "RemoveContainer" containerID="446b95da2a728878e0bbb5342f503f35d9921dc555f3c506697957799edeb012" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.014823 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n5xnj" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.031835 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"7b4f22a1-41dc-4a75-8999-53df29d2fb5d\") " pod="openstack/rabbitmq-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.031879 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-server-0\" (UID: \"7b4f22a1-41dc-4a75-8999-53df29d2fb5d\") " pod="openstack/rabbitmq-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.031913 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"7b4f22a1-41dc-4a75-8999-53df29d2fb5d\") " pod="openstack/rabbitmq-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.031951 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"7b4f22a1-41dc-4a75-8999-53df29d2fb5d\") " pod="openstack/rabbitmq-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.031971 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-config-data\") pod \"rabbitmq-server-0\" (UID: \"7b4f22a1-41dc-4a75-8999-53df29d2fb5d\") " pod="openstack/rabbitmq-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.031987 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-server-conf\") pod \"rabbitmq-server-0\" (UID: \"7b4f22a1-41dc-4a75-8999-53df29d2fb5d\") " pod="openstack/rabbitmq-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.032013 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjz5x\" (UniqueName: \"kubernetes.io/projected/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-kube-api-access-cjz5x\") pod \"rabbitmq-server-0\" (UID: \"7b4f22a1-41dc-4a75-8999-53df29d2fb5d\") " pod="openstack/rabbitmq-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.032436 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"7b4f22a1-41dc-4a75-8999-53df29d2fb5d\") " pod="openstack/rabbitmq-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.032491 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-pod-info\") pod \"rabbitmq-server-0\" (UID: \"7b4f22a1-41dc-4a75-8999-53df29d2fb5d\") " pod="openstack/rabbitmq-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.032547 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"7b4f22a1-41dc-4a75-8999-53df29d2fb5d\") " pod="openstack/rabbitmq-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.032569 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"7b4f22a1-41dc-4a75-8999-53df29d2fb5d\") " pod="openstack/rabbitmq-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.032598 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"7b4f22a1-41dc-4a75-8999-53df29d2fb5d\") " pod="openstack/rabbitmq-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.033465 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"7b4f22a1-41dc-4a75-8999-53df29d2fb5d\") " pod="openstack/rabbitmq-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.034038 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-server-conf\") pod \"rabbitmq-server-0\" (UID: \"7b4f22a1-41dc-4a75-8999-53df29d2fb5d\") " pod="openstack/rabbitmq-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.034129 4802 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-server-0\" (UID: \"7b4f22a1-41dc-4a75-8999-53df29d2fb5d\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/rabbitmq-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.034321 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"7b4f22a1-41dc-4a75-8999-53df29d2fb5d\") " pod="openstack/rabbitmq-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.036181 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-config-data\") pod \"rabbitmq-server-0\" (UID: \"7b4f22a1-41dc-4a75-8999-53df29d2fb5d\") " pod="openstack/rabbitmq-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.037778 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"7b4f22a1-41dc-4a75-8999-53df29d2fb5d\") " pod="openstack/rabbitmq-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.041558 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-pod-info\") pod \"rabbitmq-server-0\" (UID: \"7b4f22a1-41dc-4a75-8999-53df29d2fb5d\") " pod="openstack/rabbitmq-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.044601 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"7b4f22a1-41dc-4a75-8999-53df29d2fb5d\") " pod="openstack/rabbitmq-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.051696 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjz5x\" (UniqueName: \"kubernetes.io/projected/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-kube-api-access-cjz5x\") pod \"rabbitmq-server-0\" (UID: \"7b4f22a1-41dc-4a75-8999-53df29d2fb5d\") " pod="openstack/rabbitmq-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.053855 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.056583 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.062299 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.062779 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.063401 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.063601 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.063869 4802 scope.go:117] "RemoveContainer" containerID="b9e568af7c79aacbf4797181e11d497576752abfab9922004b5880d7204e93c7" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.064343 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.064529 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"7b4f22a1-41dc-4a75-8999-53df29d2fb5d\") " pod="openstack/rabbitmq-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.064723 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.065010 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-lf7lb" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.092666 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.101799 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-server-0\" (UID: \"7b4f22a1-41dc-4a75-8999-53df29d2fb5d\") " pod="openstack/rabbitmq-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.145778 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-n5xnj"] Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.153225 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-n5xnj"] Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.164106 4802 scope.go:117] "RemoveContainer" containerID="22449ccc96a78c21ea51a33ed973838df22bdd5e3a3ca6334f607f011421ea11" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.190915 4802 scope.go:117] "RemoveContainer" containerID="446b95da2a728878e0bbb5342f503f35d9921dc555f3c506697957799edeb012" Dec 06 03:56:32 crc kubenswrapper[4802]: E1206 03:56:32.191349 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"446b95da2a728878e0bbb5342f503f35d9921dc555f3c506697957799edeb012\": container with ID starting with 446b95da2a728878e0bbb5342f503f35d9921dc555f3c506697957799edeb012 not found: ID does not exist" containerID="446b95da2a728878e0bbb5342f503f35d9921dc555f3c506697957799edeb012" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.191395 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"446b95da2a728878e0bbb5342f503f35d9921dc555f3c506697957799edeb012"} err="failed to get container status \"446b95da2a728878e0bbb5342f503f35d9921dc555f3c506697957799edeb012\": rpc error: code = NotFound desc = could not find container \"446b95da2a728878e0bbb5342f503f35d9921dc555f3c506697957799edeb012\": container with ID starting with 446b95da2a728878e0bbb5342f503f35d9921dc555f3c506697957799edeb012 not found: ID does not exist" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.191423 4802 scope.go:117] "RemoveContainer" containerID="b9e568af7c79aacbf4797181e11d497576752abfab9922004b5880d7204e93c7" Dec 06 03:56:32 crc kubenswrapper[4802]: E1206 03:56:32.191786 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9e568af7c79aacbf4797181e11d497576752abfab9922004b5880d7204e93c7\": container with ID starting with b9e568af7c79aacbf4797181e11d497576752abfab9922004b5880d7204e93c7 not found: ID does not exist" containerID="b9e568af7c79aacbf4797181e11d497576752abfab9922004b5880d7204e93c7" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.191848 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9e568af7c79aacbf4797181e11d497576752abfab9922004b5880d7204e93c7"} err="failed to get container status \"b9e568af7c79aacbf4797181e11d497576752abfab9922004b5880d7204e93c7\": rpc error: code = NotFound desc = could not find container \"b9e568af7c79aacbf4797181e11d497576752abfab9922004b5880d7204e93c7\": container with ID starting with b9e568af7c79aacbf4797181e11d497576752abfab9922004b5880d7204e93c7 not found: ID does not exist" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.191869 4802 scope.go:117] "RemoveContainer" containerID="22449ccc96a78c21ea51a33ed973838df22bdd5e3a3ca6334f607f011421ea11" Dec 06 03:56:32 crc kubenswrapper[4802]: E1206 03:56:32.192138 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22449ccc96a78c21ea51a33ed973838df22bdd5e3a3ca6334f607f011421ea11\": container with ID starting with 22449ccc96a78c21ea51a33ed973838df22bdd5e3a3ca6334f607f011421ea11 not found: ID does not exist" containerID="22449ccc96a78c21ea51a33ed973838df22bdd5e3a3ca6334f607f011421ea11" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.192161 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22449ccc96a78c21ea51a33ed973838df22bdd5e3a3ca6334f607f011421ea11"} err="failed to get container status \"22449ccc96a78c21ea51a33ed973838df22bdd5e3a3ca6334f607f011421ea11\": rpc error: code = NotFound desc = could not find container \"22449ccc96a78c21ea51a33ed973838df22bdd5e3a3ca6334f607f011421ea11\": container with ID starting with 22449ccc96a78c21ea51a33ed973838df22bdd5e3a3ca6334f607f011421ea11 not found: ID does not exist" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.241961 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8c8b6c49-6f71-403d-b639-4a17ae3bc802-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"8c8b6c49-6f71-403d-b639-4a17ae3bc802\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.242009 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8c8b6c49-6f71-403d-b639-4a17ae3bc802-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8c8b6c49-6f71-403d-b639-4a17ae3bc802\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.242035 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8c8b6c49-6f71-403d-b639-4a17ae3bc802-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"8c8b6c49-6f71-403d-b639-4a17ae3bc802\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.242054 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9lzq4\" (UniqueName: \"kubernetes.io/projected/8c8b6c49-6f71-403d-b639-4a17ae3bc802-kube-api-access-9lzq4\") pod \"rabbitmq-cell1-server-0\" (UID: \"8c8b6c49-6f71-403d-b639-4a17ae3bc802\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.242071 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8c8b6c49-6f71-403d-b639-4a17ae3bc802-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"8c8b6c49-6f71-403d-b639-4a17ae3bc802\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.242090 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8c8b6c49-6f71-403d-b639-4a17ae3bc802-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8c8b6c49-6f71-403d-b639-4a17ae3bc802\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.242111 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"8c8b6c49-6f71-403d-b639-4a17ae3bc802\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.242200 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8c8b6c49-6f71-403d-b639-4a17ae3bc802-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"8c8b6c49-6f71-403d-b639-4a17ae3bc802\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.242242 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8c8b6c49-6f71-403d-b639-4a17ae3bc802-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"8c8b6c49-6f71-403d-b639-4a17ae3bc802\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.242261 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8c8b6c49-6f71-403d-b639-4a17ae3bc802-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"8c8b6c49-6f71-403d-b639-4a17ae3bc802\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.242275 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8c8b6c49-6f71-403d-b639-4a17ae3bc802-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"8c8b6c49-6f71-403d-b639-4a17ae3bc802\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.343471 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9lzq4\" (UniqueName: \"kubernetes.io/projected/8c8b6c49-6f71-403d-b639-4a17ae3bc802-kube-api-access-9lzq4\") pod \"rabbitmq-cell1-server-0\" (UID: \"8c8b6c49-6f71-403d-b639-4a17ae3bc802\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.343515 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8c8b6c49-6f71-403d-b639-4a17ae3bc802-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"8c8b6c49-6f71-403d-b639-4a17ae3bc802\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.343542 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8c8b6c49-6f71-403d-b639-4a17ae3bc802-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8c8b6c49-6f71-403d-b639-4a17ae3bc802\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.343570 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"8c8b6c49-6f71-403d-b639-4a17ae3bc802\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.343600 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8c8b6c49-6f71-403d-b639-4a17ae3bc802-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"8c8b6c49-6f71-403d-b639-4a17ae3bc802\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.343619 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8c8b6c49-6f71-403d-b639-4a17ae3bc802-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"8c8b6c49-6f71-403d-b639-4a17ae3bc802\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.343634 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8c8b6c49-6f71-403d-b639-4a17ae3bc802-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"8c8b6c49-6f71-403d-b639-4a17ae3bc802\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.343648 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8c8b6c49-6f71-403d-b639-4a17ae3bc802-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"8c8b6c49-6f71-403d-b639-4a17ae3bc802\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.343709 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8c8b6c49-6f71-403d-b639-4a17ae3bc802-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"8c8b6c49-6f71-403d-b639-4a17ae3bc802\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.343730 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8c8b6c49-6f71-403d-b639-4a17ae3bc802-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8c8b6c49-6f71-403d-b639-4a17ae3bc802\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.343769 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8c8b6c49-6f71-403d-b639-4a17ae3bc802-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"8c8b6c49-6f71-403d-b639-4a17ae3bc802\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.344665 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8c8b6c49-6f71-403d-b639-4a17ae3bc802-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"8c8b6c49-6f71-403d-b639-4a17ae3bc802\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.344720 4802 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"8c8b6c49-6f71-403d-b639-4a17ae3bc802\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/rabbitmq-cell1-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.345099 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8c8b6c49-6f71-403d-b639-4a17ae3bc802-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"8c8b6c49-6f71-403d-b639-4a17ae3bc802\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.345250 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8c8b6c49-6f71-403d-b639-4a17ae3bc802-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"8c8b6c49-6f71-403d-b639-4a17ae3bc802\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.345307 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8c8b6c49-6f71-403d-b639-4a17ae3bc802-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8c8b6c49-6f71-403d-b639-4a17ae3bc802\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.345403 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8c8b6c49-6f71-403d-b639-4a17ae3bc802-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8c8b6c49-6f71-403d-b639-4a17ae3bc802\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.349093 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8c8b6c49-6f71-403d-b639-4a17ae3bc802-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"8c8b6c49-6f71-403d-b639-4a17ae3bc802\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.352307 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8c8b6c49-6f71-403d-b639-4a17ae3bc802-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"8c8b6c49-6f71-403d-b639-4a17ae3bc802\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.355360 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8c8b6c49-6f71-403d-b639-4a17ae3bc802-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"8c8b6c49-6f71-403d-b639-4a17ae3bc802\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.361429 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9lzq4\" (UniqueName: \"kubernetes.io/projected/8c8b6c49-6f71-403d-b639-4a17ae3bc802-kube-api-access-9lzq4\") pod \"rabbitmq-cell1-server-0\" (UID: \"8c8b6c49-6f71-403d-b639-4a17ae3bc802\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.362210 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8c8b6c49-6f71-403d-b639-4a17ae3bc802-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"8c8b6c49-6f71-403d-b639-4a17ae3bc802\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.370250 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"8c8b6c49-6f71-403d-b639-4a17ae3bc802\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.389785 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 06 03:56:32 crc kubenswrapper[4802]: I1206 03:56:32.448958 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 06 03:56:33 crc kubenswrapper[4802]: I1206 03:56:33.004462 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 03:56:33 crc kubenswrapper[4802]: W1206 03:56:33.017714 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7b4f22a1_41dc_4a75_8999_53df29d2fb5d.slice/crio-73c1739eed3a5ff0dcfc6eb179c448a0489e80c5312852db4d5873a762d3b055 WatchSource:0}: Error finding container 73c1739eed3a5ff0dcfc6eb179c448a0489e80c5312852db4d5873a762d3b055: Status 404 returned error can't find the container with id 73c1739eed3a5ff0dcfc6eb179c448a0489e80c5312852db4d5873a762d3b055 Dec 06 03:56:33 crc kubenswrapper[4802]: I1206 03:56:33.159243 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 03:56:33 crc kubenswrapper[4802]: W1206 03:56:33.250877 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8c8b6c49_6f71_403d_b639_4a17ae3bc802.slice/crio-9560d5c1751b3a124c55a3571e5ffd1725570c0666099243625f058a51958f87 WatchSource:0}: Error finding container 9560d5c1751b3a124c55a3571e5ffd1725570c0666099243625f058a51958f87: Status 404 returned error can't find the container with id 9560d5c1751b3a124c55a3571e5ffd1725570c0666099243625f058a51958f87 Dec 06 03:56:33 crc kubenswrapper[4802]: I1206 03:56:33.458879 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa" path="/var/lib/kubelet/pods/66ba9f3c-2e4d-4d46-860c-9ef14fdae3aa/volumes" Dec 06 03:56:33 crc kubenswrapper[4802]: I1206 03:56:33.504720 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 06 03:56:33 crc kubenswrapper[4802]: I1206 03:56:33.507043 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 06 03:56:33 crc kubenswrapper[4802]: I1206 03:56:33.514383 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 06 03:56:33 crc kubenswrapper[4802]: I1206 03:56:33.519725 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 06 03:56:33 crc kubenswrapper[4802]: I1206 03:56:33.520781 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-p8ww8" Dec 06 03:56:33 crc kubenswrapper[4802]: I1206 03:56:33.521643 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 06 03:56:33 crc kubenswrapper[4802]: I1206 03:56:33.523015 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 06 03:56:33 crc kubenswrapper[4802]: I1206 03:56:33.523408 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 06 03:56:33 crc kubenswrapper[4802]: I1206 03:56:33.676089 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2m5mb\" (UniqueName: \"kubernetes.io/projected/04b6f630-c38a-411e-b48c-99690653972a-kube-api-access-2m5mb\") pod \"openstack-galera-0\" (UID: \"04b6f630-c38a-411e-b48c-99690653972a\") " pod="openstack/openstack-galera-0" Dec 06 03:56:33 crc kubenswrapper[4802]: I1206 03:56:33.676430 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/04b6f630-c38a-411e-b48c-99690653972a-kolla-config\") pod \"openstack-galera-0\" (UID: \"04b6f630-c38a-411e-b48c-99690653972a\") " pod="openstack/openstack-galera-0" Dec 06 03:56:33 crc kubenswrapper[4802]: I1206 03:56:33.676455 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-0\" (UID: \"04b6f630-c38a-411e-b48c-99690653972a\") " pod="openstack/openstack-galera-0" Dec 06 03:56:33 crc kubenswrapper[4802]: I1206 03:56:33.676479 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/04b6f630-c38a-411e-b48c-99690653972a-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"04b6f630-c38a-411e-b48c-99690653972a\") " pod="openstack/openstack-galera-0" Dec 06 03:56:33 crc kubenswrapper[4802]: I1206 03:56:33.676595 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04b6f630-c38a-411e-b48c-99690653972a-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"04b6f630-c38a-411e-b48c-99690653972a\") " pod="openstack/openstack-galera-0" Dec 06 03:56:33 crc kubenswrapper[4802]: I1206 03:56:33.676632 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/04b6f630-c38a-411e-b48c-99690653972a-operator-scripts\") pod \"openstack-galera-0\" (UID: \"04b6f630-c38a-411e-b48c-99690653972a\") " pod="openstack/openstack-galera-0" Dec 06 03:56:33 crc kubenswrapper[4802]: I1206 03:56:33.676691 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/04b6f630-c38a-411e-b48c-99690653972a-config-data-generated\") pod \"openstack-galera-0\" (UID: \"04b6f630-c38a-411e-b48c-99690653972a\") " pod="openstack/openstack-galera-0" Dec 06 03:56:33 crc kubenswrapper[4802]: I1206 03:56:33.676829 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/04b6f630-c38a-411e-b48c-99690653972a-config-data-default\") pod \"openstack-galera-0\" (UID: \"04b6f630-c38a-411e-b48c-99690653972a\") " pod="openstack/openstack-galera-0" Dec 06 03:56:33 crc kubenswrapper[4802]: I1206 03:56:33.779863 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2m5mb\" (UniqueName: \"kubernetes.io/projected/04b6f630-c38a-411e-b48c-99690653972a-kube-api-access-2m5mb\") pod \"openstack-galera-0\" (UID: \"04b6f630-c38a-411e-b48c-99690653972a\") " pod="openstack/openstack-galera-0" Dec 06 03:56:33 crc kubenswrapper[4802]: I1206 03:56:33.779982 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/04b6f630-c38a-411e-b48c-99690653972a-kolla-config\") pod \"openstack-galera-0\" (UID: \"04b6f630-c38a-411e-b48c-99690653972a\") " pod="openstack/openstack-galera-0" Dec 06 03:56:33 crc kubenswrapper[4802]: I1206 03:56:33.780042 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-0\" (UID: \"04b6f630-c38a-411e-b48c-99690653972a\") " pod="openstack/openstack-galera-0" Dec 06 03:56:33 crc kubenswrapper[4802]: I1206 03:56:33.780080 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/04b6f630-c38a-411e-b48c-99690653972a-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"04b6f630-c38a-411e-b48c-99690653972a\") " pod="openstack/openstack-galera-0" Dec 06 03:56:33 crc kubenswrapper[4802]: I1206 03:56:33.780149 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04b6f630-c38a-411e-b48c-99690653972a-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"04b6f630-c38a-411e-b48c-99690653972a\") " pod="openstack/openstack-galera-0" Dec 06 03:56:33 crc kubenswrapper[4802]: I1206 03:56:33.780223 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/04b6f630-c38a-411e-b48c-99690653972a-operator-scripts\") pod \"openstack-galera-0\" (UID: \"04b6f630-c38a-411e-b48c-99690653972a\") " pod="openstack/openstack-galera-0" Dec 06 03:56:33 crc kubenswrapper[4802]: I1206 03:56:33.780688 4802 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-0\" (UID: \"04b6f630-c38a-411e-b48c-99690653972a\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/openstack-galera-0" Dec 06 03:56:33 crc kubenswrapper[4802]: I1206 03:56:33.783993 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/04b6f630-c38a-411e-b48c-99690653972a-kolla-config\") pod \"openstack-galera-0\" (UID: \"04b6f630-c38a-411e-b48c-99690653972a\") " pod="openstack/openstack-galera-0" Dec 06 03:56:33 crc kubenswrapper[4802]: I1206 03:56:33.784977 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/04b6f630-c38a-411e-b48c-99690653972a-operator-scripts\") pod \"openstack-galera-0\" (UID: \"04b6f630-c38a-411e-b48c-99690653972a\") " pod="openstack/openstack-galera-0" Dec 06 03:56:33 crc kubenswrapper[4802]: I1206 03:56:33.790060 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/04b6f630-c38a-411e-b48c-99690653972a-config-data-generated\") pod \"openstack-galera-0\" (UID: \"04b6f630-c38a-411e-b48c-99690653972a\") " pod="openstack/openstack-galera-0" Dec 06 03:56:33 crc kubenswrapper[4802]: I1206 03:56:33.790195 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/04b6f630-c38a-411e-b48c-99690653972a-config-data-default\") pod \"openstack-galera-0\" (UID: \"04b6f630-c38a-411e-b48c-99690653972a\") " pod="openstack/openstack-galera-0" Dec 06 03:56:33 crc kubenswrapper[4802]: I1206 03:56:33.790981 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/04b6f630-c38a-411e-b48c-99690653972a-config-data-generated\") pod \"openstack-galera-0\" (UID: \"04b6f630-c38a-411e-b48c-99690653972a\") " pod="openstack/openstack-galera-0" Dec 06 03:56:33 crc kubenswrapper[4802]: I1206 03:56:33.792116 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/04b6f630-c38a-411e-b48c-99690653972a-config-data-default\") pod \"openstack-galera-0\" (UID: \"04b6f630-c38a-411e-b48c-99690653972a\") " pod="openstack/openstack-galera-0" Dec 06 03:56:33 crc kubenswrapper[4802]: I1206 03:56:33.807791 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04b6f630-c38a-411e-b48c-99690653972a-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"04b6f630-c38a-411e-b48c-99690653972a\") " pod="openstack/openstack-galera-0" Dec 06 03:56:33 crc kubenswrapper[4802]: I1206 03:56:33.812606 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/04b6f630-c38a-411e-b48c-99690653972a-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"04b6f630-c38a-411e-b48c-99690653972a\") " pod="openstack/openstack-galera-0" Dec 06 03:56:33 crc kubenswrapper[4802]: I1206 03:56:33.820885 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2m5mb\" (UniqueName: \"kubernetes.io/projected/04b6f630-c38a-411e-b48c-99690653972a-kube-api-access-2m5mb\") pod \"openstack-galera-0\" (UID: \"04b6f630-c38a-411e-b48c-99690653972a\") " pod="openstack/openstack-galera-0" Dec 06 03:56:33 crc kubenswrapper[4802]: I1206 03:56:33.868697 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-0\" (UID: \"04b6f630-c38a-411e-b48c-99690653972a\") " pod="openstack/openstack-galera-0" Dec 06 03:56:34 crc kubenswrapper[4802]: I1206 03:56:34.032699 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8c8b6c49-6f71-403d-b639-4a17ae3bc802","Type":"ContainerStarted","Data":"9560d5c1751b3a124c55a3571e5ffd1725570c0666099243625f058a51958f87"} Dec 06 03:56:34 crc kubenswrapper[4802]: I1206 03:56:34.040158 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7b4f22a1-41dc-4a75-8999-53df29d2fb5d","Type":"ContainerStarted","Data":"73c1739eed3a5ff0dcfc6eb179c448a0489e80c5312852db4d5873a762d3b055"} Dec 06 03:56:34 crc kubenswrapper[4802]: I1206 03:56:34.143255 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 06 03:56:34 crc kubenswrapper[4802]: I1206 03:56:34.661251 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 06 03:56:34 crc kubenswrapper[4802]: I1206 03:56:34.871855 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 06 03:56:34 crc kubenswrapper[4802]: I1206 03:56:34.875195 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 06 03:56:34 crc kubenswrapper[4802]: I1206 03:56:34.882103 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-kzkkc" Dec 06 03:56:34 crc kubenswrapper[4802]: I1206 03:56:34.882976 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 06 03:56:34 crc kubenswrapper[4802]: I1206 03:56:34.883103 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 06 03:56:34 crc kubenswrapper[4802]: I1206 03:56:34.883825 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 06 03:56:34 crc kubenswrapper[4802]: I1206 03:56:34.884954 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 06 03:56:35 crc kubenswrapper[4802]: I1206 03:56:35.032354 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/db5b3080-8b4d-4232-a7da-66e544c09c60-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"db5b3080-8b4d-4232-a7da-66e544c09c60\") " pod="openstack/openstack-cell1-galera-0" Dec 06 03:56:35 crc kubenswrapper[4802]: I1206 03:56:35.032395 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/db5b3080-8b4d-4232-a7da-66e544c09c60-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"db5b3080-8b4d-4232-a7da-66e544c09c60\") " pod="openstack/openstack-cell1-galera-0" Dec 06 03:56:35 crc kubenswrapper[4802]: I1206 03:56:35.032433 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/db5b3080-8b4d-4232-a7da-66e544c09c60-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"db5b3080-8b4d-4232-a7da-66e544c09c60\") " pod="openstack/openstack-cell1-galera-0" Dec 06 03:56:35 crc kubenswrapper[4802]: I1206 03:56:35.032454 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/db5b3080-8b4d-4232-a7da-66e544c09c60-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"db5b3080-8b4d-4232-a7da-66e544c09c60\") " pod="openstack/openstack-cell1-galera-0" Dec 06 03:56:35 crc kubenswrapper[4802]: I1206 03:56:35.032480 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db5b3080-8b4d-4232-a7da-66e544c09c60-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"db5b3080-8b4d-4232-a7da-66e544c09c60\") " pod="openstack/openstack-cell1-galera-0" Dec 06 03:56:35 crc kubenswrapper[4802]: I1206 03:56:35.032508 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwplr\" (UniqueName: \"kubernetes.io/projected/db5b3080-8b4d-4232-a7da-66e544c09c60-kube-api-access-gwplr\") pod \"openstack-cell1-galera-0\" (UID: \"db5b3080-8b4d-4232-a7da-66e544c09c60\") " pod="openstack/openstack-cell1-galera-0" Dec 06 03:56:35 crc kubenswrapper[4802]: I1206 03:56:35.032531 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"db5b3080-8b4d-4232-a7da-66e544c09c60\") " pod="openstack/openstack-cell1-galera-0" Dec 06 03:56:35 crc kubenswrapper[4802]: I1206 03:56:35.032571 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/db5b3080-8b4d-4232-a7da-66e544c09c60-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"db5b3080-8b4d-4232-a7da-66e544c09c60\") " pod="openstack/openstack-cell1-galera-0" Dec 06 03:56:35 crc kubenswrapper[4802]: I1206 03:56:35.072559 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"04b6f630-c38a-411e-b48c-99690653972a","Type":"ContainerStarted","Data":"4edf816991a4cf539840af4e6852f35586e5febdfe6e7804d1a95fda2bfedfd3"} Dec 06 03:56:35 crc kubenswrapper[4802]: I1206 03:56:35.142258 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/db5b3080-8b4d-4232-a7da-66e544c09c60-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"db5b3080-8b4d-4232-a7da-66e544c09c60\") " pod="openstack/openstack-cell1-galera-0" Dec 06 03:56:35 crc kubenswrapper[4802]: I1206 03:56:35.142310 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/db5b3080-8b4d-4232-a7da-66e544c09c60-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"db5b3080-8b4d-4232-a7da-66e544c09c60\") " pod="openstack/openstack-cell1-galera-0" Dec 06 03:56:35 crc kubenswrapper[4802]: I1206 03:56:35.142403 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db5b3080-8b4d-4232-a7da-66e544c09c60-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"db5b3080-8b4d-4232-a7da-66e544c09c60\") " pod="openstack/openstack-cell1-galera-0" Dec 06 03:56:35 crc kubenswrapper[4802]: I1206 03:56:35.142461 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwplr\" (UniqueName: \"kubernetes.io/projected/db5b3080-8b4d-4232-a7da-66e544c09c60-kube-api-access-gwplr\") pod \"openstack-cell1-galera-0\" (UID: \"db5b3080-8b4d-4232-a7da-66e544c09c60\") " pod="openstack/openstack-cell1-galera-0" Dec 06 03:56:35 crc kubenswrapper[4802]: I1206 03:56:35.142492 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"db5b3080-8b4d-4232-a7da-66e544c09c60\") " pod="openstack/openstack-cell1-galera-0" Dec 06 03:56:35 crc kubenswrapper[4802]: I1206 03:56:35.142558 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/db5b3080-8b4d-4232-a7da-66e544c09c60-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"db5b3080-8b4d-4232-a7da-66e544c09c60\") " pod="openstack/openstack-cell1-galera-0" Dec 06 03:56:35 crc kubenswrapper[4802]: I1206 03:56:35.142618 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/db5b3080-8b4d-4232-a7da-66e544c09c60-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"db5b3080-8b4d-4232-a7da-66e544c09c60\") " pod="openstack/openstack-cell1-galera-0" Dec 06 03:56:35 crc kubenswrapper[4802]: I1206 03:56:35.142640 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/db5b3080-8b4d-4232-a7da-66e544c09c60-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"db5b3080-8b4d-4232-a7da-66e544c09c60\") " pod="openstack/openstack-cell1-galera-0" Dec 06 03:56:35 crc kubenswrapper[4802]: I1206 03:56:35.145131 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/db5b3080-8b4d-4232-a7da-66e544c09c60-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"db5b3080-8b4d-4232-a7da-66e544c09c60\") " pod="openstack/openstack-cell1-galera-0" Dec 06 03:56:35 crc kubenswrapper[4802]: I1206 03:56:35.145356 4802 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"db5b3080-8b4d-4232-a7da-66e544c09c60\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/openstack-cell1-galera-0" Dec 06 03:56:35 crc kubenswrapper[4802]: I1206 03:56:35.147540 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/db5b3080-8b4d-4232-a7da-66e544c09c60-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"db5b3080-8b4d-4232-a7da-66e544c09c60\") " pod="openstack/openstack-cell1-galera-0" Dec 06 03:56:35 crc kubenswrapper[4802]: I1206 03:56:35.147646 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/db5b3080-8b4d-4232-a7da-66e544c09c60-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"db5b3080-8b4d-4232-a7da-66e544c09c60\") " pod="openstack/openstack-cell1-galera-0" Dec 06 03:56:35 crc kubenswrapper[4802]: I1206 03:56:35.148287 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/db5b3080-8b4d-4232-a7da-66e544c09c60-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"db5b3080-8b4d-4232-a7da-66e544c09c60\") " pod="openstack/openstack-cell1-galera-0" Dec 06 03:56:35 crc kubenswrapper[4802]: I1206 03:56:35.181675 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/db5b3080-8b4d-4232-a7da-66e544c09c60-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"db5b3080-8b4d-4232-a7da-66e544c09c60\") " pod="openstack/openstack-cell1-galera-0" Dec 06 03:56:35 crc kubenswrapper[4802]: I1206 03:56:35.186177 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db5b3080-8b4d-4232-a7da-66e544c09c60-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"db5b3080-8b4d-4232-a7da-66e544c09c60\") " pod="openstack/openstack-cell1-galera-0" Dec 06 03:56:35 crc kubenswrapper[4802]: I1206 03:56:35.194706 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwplr\" (UniqueName: \"kubernetes.io/projected/db5b3080-8b4d-4232-a7da-66e544c09c60-kube-api-access-gwplr\") pod \"openstack-cell1-galera-0\" (UID: \"db5b3080-8b4d-4232-a7da-66e544c09c60\") " pod="openstack/openstack-cell1-galera-0" Dec 06 03:56:35 crc kubenswrapper[4802]: I1206 03:56:35.239372 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 06 03:56:35 crc kubenswrapper[4802]: I1206 03:56:35.240645 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 06 03:56:35 crc kubenswrapper[4802]: I1206 03:56:35.244671 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-lt5p6" Dec 06 03:56:35 crc kubenswrapper[4802]: I1206 03:56:35.244875 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 06 03:56:35 crc kubenswrapper[4802]: I1206 03:56:35.245000 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 06 03:56:35 crc kubenswrapper[4802]: I1206 03:56:35.250875 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 06 03:56:35 crc kubenswrapper[4802]: I1206 03:56:35.261077 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"db5b3080-8b4d-4232-a7da-66e544c09c60\") " pod="openstack/openstack-cell1-galera-0" Dec 06 03:56:35 crc kubenswrapper[4802]: I1206 03:56:35.352017 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/654f214b-8467-41d2-93f4-3c18377653b9-config-data\") pod \"memcached-0\" (UID: \"654f214b-8467-41d2-93f4-3c18377653b9\") " pod="openstack/memcached-0" Dec 06 03:56:35 crc kubenswrapper[4802]: I1206 03:56:35.352057 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/654f214b-8467-41d2-93f4-3c18377653b9-kolla-config\") pod \"memcached-0\" (UID: \"654f214b-8467-41d2-93f4-3c18377653b9\") " pod="openstack/memcached-0" Dec 06 03:56:35 crc kubenswrapper[4802]: I1206 03:56:35.352085 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/654f214b-8467-41d2-93f4-3c18377653b9-memcached-tls-certs\") pod \"memcached-0\" (UID: \"654f214b-8467-41d2-93f4-3c18377653b9\") " pod="openstack/memcached-0" Dec 06 03:56:35 crc kubenswrapper[4802]: I1206 03:56:35.352104 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/654f214b-8467-41d2-93f4-3c18377653b9-combined-ca-bundle\") pod \"memcached-0\" (UID: \"654f214b-8467-41d2-93f4-3c18377653b9\") " pod="openstack/memcached-0" Dec 06 03:56:35 crc kubenswrapper[4802]: I1206 03:56:35.352131 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t48tg\" (UniqueName: \"kubernetes.io/projected/654f214b-8467-41d2-93f4-3c18377653b9-kube-api-access-t48tg\") pod \"memcached-0\" (UID: \"654f214b-8467-41d2-93f4-3c18377653b9\") " pod="openstack/memcached-0" Dec 06 03:56:35 crc kubenswrapper[4802]: I1206 03:56:35.453707 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/654f214b-8467-41d2-93f4-3c18377653b9-config-data\") pod \"memcached-0\" (UID: \"654f214b-8467-41d2-93f4-3c18377653b9\") " pod="openstack/memcached-0" Dec 06 03:56:35 crc kubenswrapper[4802]: I1206 03:56:35.453739 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/654f214b-8467-41d2-93f4-3c18377653b9-kolla-config\") pod \"memcached-0\" (UID: \"654f214b-8467-41d2-93f4-3c18377653b9\") " pod="openstack/memcached-0" Dec 06 03:56:35 crc kubenswrapper[4802]: I1206 03:56:35.453782 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/654f214b-8467-41d2-93f4-3c18377653b9-memcached-tls-certs\") pod \"memcached-0\" (UID: \"654f214b-8467-41d2-93f4-3c18377653b9\") " pod="openstack/memcached-0" Dec 06 03:56:35 crc kubenswrapper[4802]: I1206 03:56:35.453800 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/654f214b-8467-41d2-93f4-3c18377653b9-combined-ca-bundle\") pod \"memcached-0\" (UID: \"654f214b-8467-41d2-93f4-3c18377653b9\") " pod="openstack/memcached-0" Dec 06 03:56:35 crc kubenswrapper[4802]: I1206 03:56:35.453828 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t48tg\" (UniqueName: \"kubernetes.io/projected/654f214b-8467-41d2-93f4-3c18377653b9-kube-api-access-t48tg\") pod \"memcached-0\" (UID: \"654f214b-8467-41d2-93f4-3c18377653b9\") " pod="openstack/memcached-0" Dec 06 03:56:35 crc kubenswrapper[4802]: I1206 03:56:35.454736 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/654f214b-8467-41d2-93f4-3c18377653b9-config-data\") pod \"memcached-0\" (UID: \"654f214b-8467-41d2-93f4-3c18377653b9\") " pod="openstack/memcached-0" Dec 06 03:56:35 crc kubenswrapper[4802]: I1206 03:56:35.455165 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/654f214b-8467-41d2-93f4-3c18377653b9-kolla-config\") pod \"memcached-0\" (UID: \"654f214b-8467-41d2-93f4-3c18377653b9\") " pod="openstack/memcached-0" Dec 06 03:56:35 crc kubenswrapper[4802]: I1206 03:56:35.484259 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/654f214b-8467-41d2-93f4-3c18377653b9-memcached-tls-certs\") pod \"memcached-0\" (UID: \"654f214b-8467-41d2-93f4-3c18377653b9\") " pod="openstack/memcached-0" Dec 06 03:56:35 crc kubenswrapper[4802]: I1206 03:56:35.518575 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 06 03:56:35 crc kubenswrapper[4802]: I1206 03:56:35.523394 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t48tg\" (UniqueName: \"kubernetes.io/projected/654f214b-8467-41d2-93f4-3c18377653b9-kube-api-access-t48tg\") pod \"memcached-0\" (UID: \"654f214b-8467-41d2-93f4-3c18377653b9\") " pod="openstack/memcached-0" Dec 06 03:56:35 crc kubenswrapper[4802]: I1206 03:56:35.523990 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/654f214b-8467-41d2-93f4-3c18377653b9-combined-ca-bundle\") pod \"memcached-0\" (UID: \"654f214b-8467-41d2-93f4-3c18377653b9\") " pod="openstack/memcached-0" Dec 06 03:56:35 crc kubenswrapper[4802]: I1206 03:56:35.624912 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 06 03:56:36 crc kubenswrapper[4802]: I1206 03:56:36.927991 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 03:56:36 crc kubenswrapper[4802]: I1206 03:56:36.929253 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 06 03:56:36 crc kubenswrapper[4802]: I1206 03:56:36.931206 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-vs64n" Dec 06 03:56:36 crc kubenswrapper[4802]: I1206 03:56:36.936903 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 03:56:37 crc kubenswrapper[4802]: I1206 03:56:37.018381 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbzm2\" (UniqueName: \"kubernetes.io/projected/b96722fa-dc6a-44c3-9460-0ce247056613-kube-api-access-fbzm2\") pod \"kube-state-metrics-0\" (UID: \"b96722fa-dc6a-44c3-9460-0ce247056613\") " pod="openstack/kube-state-metrics-0" Dec 06 03:56:37 crc kubenswrapper[4802]: I1206 03:56:37.124731 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbzm2\" (UniqueName: \"kubernetes.io/projected/b96722fa-dc6a-44c3-9460-0ce247056613-kube-api-access-fbzm2\") pod \"kube-state-metrics-0\" (UID: \"b96722fa-dc6a-44c3-9460-0ce247056613\") " pod="openstack/kube-state-metrics-0" Dec 06 03:56:37 crc kubenswrapper[4802]: I1206 03:56:37.146215 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbzm2\" (UniqueName: \"kubernetes.io/projected/b96722fa-dc6a-44c3-9460-0ce247056613-kube-api-access-fbzm2\") pod \"kube-state-metrics-0\" (UID: \"b96722fa-dc6a-44c3-9460-0ce247056613\") " pod="openstack/kube-state-metrics-0" Dec 06 03:56:37 crc kubenswrapper[4802]: I1206 03:56:37.282353 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 06 03:56:40 crc kubenswrapper[4802]: I1206 03:56:40.793519 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-sbt8p"] Dec 06 03:56:40 crc kubenswrapper[4802]: I1206 03:56:40.795181 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-sbt8p" Dec 06 03:56:40 crc kubenswrapper[4802]: I1206 03:56:40.803054 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-49cxn"] Dec 06 03:56:40 crc kubenswrapper[4802]: I1206 03:56:40.803485 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 06 03:56:40 crc kubenswrapper[4802]: I1206 03:56:40.805167 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-49cxn" Dec 06 03:56:40 crc kubenswrapper[4802]: I1206 03:56:40.807182 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-8l2m5" Dec 06 03:56:40 crc kubenswrapper[4802]: I1206 03:56:40.808596 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 06 03:56:40 crc kubenswrapper[4802]: I1206 03:56:40.811060 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-sbt8p"] Dec 06 03:56:40 crc kubenswrapper[4802]: I1206 03:56:40.834039 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-49cxn"] Dec 06 03:56:40 crc kubenswrapper[4802]: I1206 03:56:40.914262 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/edb79ab5-ba10-4379-96e9-b2de42f6cbc8-var-log\") pod \"ovn-controller-ovs-49cxn\" (UID: \"edb79ab5-ba10-4379-96e9-b2de42f6cbc8\") " pod="openstack/ovn-controller-ovs-49cxn" Dec 06 03:56:40 crc kubenswrapper[4802]: I1206 03:56:40.914324 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/edb79ab5-ba10-4379-96e9-b2de42f6cbc8-var-run\") pod \"ovn-controller-ovs-49cxn\" (UID: \"edb79ab5-ba10-4379-96e9-b2de42f6cbc8\") " pod="openstack/ovn-controller-ovs-49cxn" Dec 06 03:56:40 crc kubenswrapper[4802]: I1206 03:56:40.914410 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/edb79ab5-ba10-4379-96e9-b2de42f6cbc8-scripts\") pod \"ovn-controller-ovs-49cxn\" (UID: \"edb79ab5-ba10-4379-96e9-b2de42f6cbc8\") " pod="openstack/ovn-controller-ovs-49cxn" Dec 06 03:56:40 crc kubenswrapper[4802]: I1206 03:56:40.914442 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e785e69-07cd-4c5c-8446-3d2a0d7e7b0b-combined-ca-bundle\") pod \"ovn-controller-sbt8p\" (UID: \"6e785e69-07cd-4c5c-8446-3d2a0d7e7b0b\") " pod="openstack/ovn-controller-sbt8p" Dec 06 03:56:40 crc kubenswrapper[4802]: I1206 03:56:40.914563 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6e785e69-07cd-4c5c-8446-3d2a0d7e7b0b-var-run\") pod \"ovn-controller-sbt8p\" (UID: \"6e785e69-07cd-4c5c-8446-3d2a0d7e7b0b\") " pod="openstack/ovn-controller-sbt8p" Dec 06 03:56:40 crc kubenswrapper[4802]: I1206 03:56:40.914634 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6e785e69-07cd-4c5c-8446-3d2a0d7e7b0b-var-log-ovn\") pod \"ovn-controller-sbt8p\" (UID: \"6e785e69-07cd-4c5c-8446-3d2a0d7e7b0b\") " pod="openstack/ovn-controller-sbt8p" Dec 06 03:56:40 crc kubenswrapper[4802]: I1206 03:56:40.914685 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6e785e69-07cd-4c5c-8446-3d2a0d7e7b0b-var-run-ovn\") pod \"ovn-controller-sbt8p\" (UID: \"6e785e69-07cd-4c5c-8446-3d2a0d7e7b0b\") " pod="openstack/ovn-controller-sbt8p" Dec 06 03:56:40 crc kubenswrapper[4802]: I1206 03:56:40.914732 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c69lh\" (UniqueName: \"kubernetes.io/projected/edb79ab5-ba10-4379-96e9-b2de42f6cbc8-kube-api-access-c69lh\") pod \"ovn-controller-ovs-49cxn\" (UID: \"edb79ab5-ba10-4379-96e9-b2de42f6cbc8\") " pod="openstack/ovn-controller-ovs-49cxn" Dec 06 03:56:40 crc kubenswrapper[4802]: I1206 03:56:40.914776 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6e785e69-07cd-4c5c-8446-3d2a0d7e7b0b-scripts\") pod \"ovn-controller-sbt8p\" (UID: \"6e785e69-07cd-4c5c-8446-3d2a0d7e7b0b\") " pod="openstack/ovn-controller-sbt8p" Dec 06 03:56:40 crc kubenswrapper[4802]: I1206 03:56:40.914810 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/edb79ab5-ba10-4379-96e9-b2de42f6cbc8-var-lib\") pod \"ovn-controller-ovs-49cxn\" (UID: \"edb79ab5-ba10-4379-96e9-b2de42f6cbc8\") " pod="openstack/ovn-controller-ovs-49cxn" Dec 06 03:56:40 crc kubenswrapper[4802]: I1206 03:56:40.914899 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e785e69-07cd-4c5c-8446-3d2a0d7e7b0b-ovn-controller-tls-certs\") pod \"ovn-controller-sbt8p\" (UID: \"6e785e69-07cd-4c5c-8446-3d2a0d7e7b0b\") " pod="openstack/ovn-controller-sbt8p" Dec 06 03:56:40 crc kubenswrapper[4802]: I1206 03:56:40.914969 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dm5d8\" (UniqueName: \"kubernetes.io/projected/6e785e69-07cd-4c5c-8446-3d2a0d7e7b0b-kube-api-access-dm5d8\") pod \"ovn-controller-sbt8p\" (UID: \"6e785e69-07cd-4c5c-8446-3d2a0d7e7b0b\") " pod="openstack/ovn-controller-sbt8p" Dec 06 03:56:40 crc kubenswrapper[4802]: I1206 03:56:40.915022 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/edb79ab5-ba10-4379-96e9-b2de42f6cbc8-etc-ovs\") pod \"ovn-controller-ovs-49cxn\" (UID: \"edb79ab5-ba10-4379-96e9-b2de42f6cbc8\") " pod="openstack/ovn-controller-ovs-49cxn" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.016154 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/edb79ab5-ba10-4379-96e9-b2de42f6cbc8-var-run\") pod \"ovn-controller-ovs-49cxn\" (UID: \"edb79ab5-ba10-4379-96e9-b2de42f6cbc8\") " pod="openstack/ovn-controller-ovs-49cxn" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.016278 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/edb79ab5-ba10-4379-96e9-b2de42f6cbc8-scripts\") pod \"ovn-controller-ovs-49cxn\" (UID: \"edb79ab5-ba10-4379-96e9-b2de42f6cbc8\") " pod="openstack/ovn-controller-ovs-49cxn" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.016314 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e785e69-07cd-4c5c-8446-3d2a0d7e7b0b-combined-ca-bundle\") pod \"ovn-controller-sbt8p\" (UID: \"6e785e69-07cd-4c5c-8446-3d2a0d7e7b0b\") " pod="openstack/ovn-controller-sbt8p" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.016365 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6e785e69-07cd-4c5c-8446-3d2a0d7e7b0b-var-run\") pod \"ovn-controller-sbt8p\" (UID: \"6e785e69-07cd-4c5c-8446-3d2a0d7e7b0b\") " pod="openstack/ovn-controller-sbt8p" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.016410 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6e785e69-07cd-4c5c-8446-3d2a0d7e7b0b-var-log-ovn\") pod \"ovn-controller-sbt8p\" (UID: \"6e785e69-07cd-4c5c-8446-3d2a0d7e7b0b\") " pod="openstack/ovn-controller-sbt8p" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.016454 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6e785e69-07cd-4c5c-8446-3d2a0d7e7b0b-var-run-ovn\") pod \"ovn-controller-sbt8p\" (UID: \"6e785e69-07cd-4c5c-8446-3d2a0d7e7b0b\") " pod="openstack/ovn-controller-sbt8p" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.016494 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c69lh\" (UniqueName: \"kubernetes.io/projected/edb79ab5-ba10-4379-96e9-b2de42f6cbc8-kube-api-access-c69lh\") pod \"ovn-controller-ovs-49cxn\" (UID: \"edb79ab5-ba10-4379-96e9-b2de42f6cbc8\") " pod="openstack/ovn-controller-ovs-49cxn" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.016525 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6e785e69-07cd-4c5c-8446-3d2a0d7e7b0b-scripts\") pod \"ovn-controller-sbt8p\" (UID: \"6e785e69-07cd-4c5c-8446-3d2a0d7e7b0b\") " pod="openstack/ovn-controller-sbt8p" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.016558 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/edb79ab5-ba10-4379-96e9-b2de42f6cbc8-var-lib\") pod \"ovn-controller-ovs-49cxn\" (UID: \"edb79ab5-ba10-4379-96e9-b2de42f6cbc8\") " pod="openstack/ovn-controller-ovs-49cxn" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.016598 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/edb79ab5-ba10-4379-96e9-b2de42f6cbc8-var-run\") pod \"ovn-controller-ovs-49cxn\" (UID: \"edb79ab5-ba10-4379-96e9-b2de42f6cbc8\") " pod="openstack/ovn-controller-ovs-49cxn" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.016606 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e785e69-07cd-4c5c-8446-3d2a0d7e7b0b-ovn-controller-tls-certs\") pod \"ovn-controller-sbt8p\" (UID: \"6e785e69-07cd-4c5c-8446-3d2a0d7e7b0b\") " pod="openstack/ovn-controller-sbt8p" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.016694 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dm5d8\" (UniqueName: \"kubernetes.io/projected/6e785e69-07cd-4c5c-8446-3d2a0d7e7b0b-kube-api-access-dm5d8\") pod \"ovn-controller-sbt8p\" (UID: \"6e785e69-07cd-4c5c-8446-3d2a0d7e7b0b\") " pod="openstack/ovn-controller-sbt8p" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.016727 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/edb79ab5-ba10-4379-96e9-b2de42f6cbc8-etc-ovs\") pod \"ovn-controller-ovs-49cxn\" (UID: \"edb79ab5-ba10-4379-96e9-b2de42f6cbc8\") " pod="openstack/ovn-controller-ovs-49cxn" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.016790 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6e785e69-07cd-4c5c-8446-3d2a0d7e7b0b-var-log-ovn\") pod \"ovn-controller-sbt8p\" (UID: \"6e785e69-07cd-4c5c-8446-3d2a0d7e7b0b\") " pod="openstack/ovn-controller-sbt8p" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.016818 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/edb79ab5-ba10-4379-96e9-b2de42f6cbc8-var-log\") pod \"ovn-controller-ovs-49cxn\" (UID: \"edb79ab5-ba10-4379-96e9-b2de42f6cbc8\") " pod="openstack/ovn-controller-ovs-49cxn" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.017030 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/edb79ab5-ba10-4379-96e9-b2de42f6cbc8-var-log\") pod \"ovn-controller-ovs-49cxn\" (UID: \"edb79ab5-ba10-4379-96e9-b2de42f6cbc8\") " pod="openstack/ovn-controller-ovs-49cxn" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.017198 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6e785e69-07cd-4c5c-8446-3d2a0d7e7b0b-var-run-ovn\") pod \"ovn-controller-sbt8p\" (UID: \"6e785e69-07cd-4c5c-8446-3d2a0d7e7b0b\") " pod="openstack/ovn-controller-sbt8p" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.018027 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/edb79ab5-ba10-4379-96e9-b2de42f6cbc8-var-lib\") pod \"ovn-controller-ovs-49cxn\" (UID: \"edb79ab5-ba10-4379-96e9-b2de42f6cbc8\") " pod="openstack/ovn-controller-ovs-49cxn" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.018670 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/edb79ab5-ba10-4379-96e9-b2de42f6cbc8-scripts\") pod \"ovn-controller-ovs-49cxn\" (UID: \"edb79ab5-ba10-4379-96e9-b2de42f6cbc8\") " pod="openstack/ovn-controller-ovs-49cxn" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.018728 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6e785e69-07cd-4c5c-8446-3d2a0d7e7b0b-var-run\") pod \"ovn-controller-sbt8p\" (UID: \"6e785e69-07cd-4c5c-8446-3d2a0d7e7b0b\") " pod="openstack/ovn-controller-sbt8p" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.019028 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/edb79ab5-ba10-4379-96e9-b2de42f6cbc8-etc-ovs\") pod \"ovn-controller-ovs-49cxn\" (UID: \"edb79ab5-ba10-4379-96e9-b2de42f6cbc8\") " pod="openstack/ovn-controller-ovs-49cxn" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.020672 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6e785e69-07cd-4c5c-8446-3d2a0d7e7b0b-scripts\") pod \"ovn-controller-sbt8p\" (UID: \"6e785e69-07cd-4c5c-8446-3d2a0d7e7b0b\") " pod="openstack/ovn-controller-sbt8p" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.025067 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e785e69-07cd-4c5c-8446-3d2a0d7e7b0b-ovn-controller-tls-certs\") pod \"ovn-controller-sbt8p\" (UID: \"6e785e69-07cd-4c5c-8446-3d2a0d7e7b0b\") " pod="openstack/ovn-controller-sbt8p" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.035646 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e785e69-07cd-4c5c-8446-3d2a0d7e7b0b-combined-ca-bundle\") pod \"ovn-controller-sbt8p\" (UID: \"6e785e69-07cd-4c5c-8446-3d2a0d7e7b0b\") " pod="openstack/ovn-controller-sbt8p" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.037095 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dm5d8\" (UniqueName: \"kubernetes.io/projected/6e785e69-07cd-4c5c-8446-3d2a0d7e7b0b-kube-api-access-dm5d8\") pod \"ovn-controller-sbt8p\" (UID: \"6e785e69-07cd-4c5c-8446-3d2a0d7e7b0b\") " pod="openstack/ovn-controller-sbt8p" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.041863 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c69lh\" (UniqueName: \"kubernetes.io/projected/edb79ab5-ba10-4379-96e9-b2de42f6cbc8-kube-api-access-c69lh\") pod \"ovn-controller-ovs-49cxn\" (UID: \"edb79ab5-ba10-4379-96e9-b2de42f6cbc8\") " pod="openstack/ovn-controller-ovs-49cxn" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.122913 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-sbt8p" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.134048 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-49cxn" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.667457 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.668610 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.716583 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.716944 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.716735 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-g92vf" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.716855 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.717080 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.731319 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.828370 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56f9d065-29e4-424d-bc84-2657eba9929e-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"56f9d065-29e4-424d-bc84-2657eba9929e\") " pod="openstack/ovsdbserver-nb-0" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.828446 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/56f9d065-29e4-424d-bc84-2657eba9929e-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"56f9d065-29e4-424d-bc84-2657eba9929e\") " pod="openstack/ovsdbserver-nb-0" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.828474 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-px8lj\" (UniqueName: \"kubernetes.io/projected/56f9d065-29e4-424d-bc84-2657eba9929e-kube-api-access-px8lj\") pod \"ovsdbserver-nb-0\" (UID: \"56f9d065-29e4-424d-bc84-2657eba9929e\") " pod="openstack/ovsdbserver-nb-0" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.828512 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/56f9d065-29e4-424d-bc84-2657eba9929e-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"56f9d065-29e4-424d-bc84-2657eba9929e\") " pod="openstack/ovsdbserver-nb-0" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.828547 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/56f9d065-29e4-424d-bc84-2657eba9929e-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"56f9d065-29e4-424d-bc84-2657eba9929e\") " pod="openstack/ovsdbserver-nb-0" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.828572 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/56f9d065-29e4-424d-bc84-2657eba9929e-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"56f9d065-29e4-424d-bc84-2657eba9929e\") " pod="openstack/ovsdbserver-nb-0" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.828624 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"56f9d065-29e4-424d-bc84-2657eba9929e\") " pod="openstack/ovsdbserver-nb-0" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.828653 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/56f9d065-29e4-424d-bc84-2657eba9929e-config\") pod \"ovsdbserver-nb-0\" (UID: \"56f9d065-29e4-424d-bc84-2657eba9929e\") " pod="openstack/ovsdbserver-nb-0" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.930652 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56f9d065-29e4-424d-bc84-2657eba9929e-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"56f9d065-29e4-424d-bc84-2657eba9929e\") " pod="openstack/ovsdbserver-nb-0" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.930726 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/56f9d065-29e4-424d-bc84-2657eba9929e-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"56f9d065-29e4-424d-bc84-2657eba9929e\") " pod="openstack/ovsdbserver-nb-0" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.930750 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-px8lj\" (UniqueName: \"kubernetes.io/projected/56f9d065-29e4-424d-bc84-2657eba9929e-kube-api-access-px8lj\") pod \"ovsdbserver-nb-0\" (UID: \"56f9d065-29e4-424d-bc84-2657eba9929e\") " pod="openstack/ovsdbserver-nb-0" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.931154 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/56f9d065-29e4-424d-bc84-2657eba9929e-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"56f9d065-29e4-424d-bc84-2657eba9929e\") " pod="openstack/ovsdbserver-nb-0" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.931206 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/56f9d065-29e4-424d-bc84-2657eba9929e-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"56f9d065-29e4-424d-bc84-2657eba9929e\") " pod="openstack/ovsdbserver-nb-0" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.931225 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/56f9d065-29e4-424d-bc84-2657eba9929e-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"56f9d065-29e4-424d-bc84-2657eba9929e\") " pod="openstack/ovsdbserver-nb-0" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.931819 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/56f9d065-29e4-424d-bc84-2657eba9929e-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"56f9d065-29e4-424d-bc84-2657eba9929e\") " pod="openstack/ovsdbserver-nb-0" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.932646 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/56f9d065-29e4-424d-bc84-2657eba9929e-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"56f9d065-29e4-424d-bc84-2657eba9929e\") " pod="openstack/ovsdbserver-nb-0" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.932807 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"56f9d065-29e4-424d-bc84-2657eba9929e\") " pod="openstack/ovsdbserver-nb-0" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.933074 4802 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"56f9d065-29e4-424d-bc84-2657eba9929e\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/ovsdbserver-nb-0" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.933585 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/56f9d065-29e4-424d-bc84-2657eba9929e-config\") pod \"ovsdbserver-nb-0\" (UID: \"56f9d065-29e4-424d-bc84-2657eba9929e\") " pod="openstack/ovsdbserver-nb-0" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.933900 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/56f9d065-29e4-424d-bc84-2657eba9929e-config\") pod \"ovsdbserver-nb-0\" (UID: \"56f9d065-29e4-424d-bc84-2657eba9929e\") " pod="openstack/ovsdbserver-nb-0" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.936742 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/56f9d065-29e4-424d-bc84-2657eba9929e-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"56f9d065-29e4-424d-bc84-2657eba9929e\") " pod="openstack/ovsdbserver-nb-0" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.937812 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56f9d065-29e4-424d-bc84-2657eba9929e-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"56f9d065-29e4-424d-bc84-2657eba9929e\") " pod="openstack/ovsdbserver-nb-0" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.938383 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/56f9d065-29e4-424d-bc84-2657eba9929e-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"56f9d065-29e4-424d-bc84-2657eba9929e\") " pod="openstack/ovsdbserver-nb-0" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.948893 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-px8lj\" (UniqueName: \"kubernetes.io/projected/56f9d065-29e4-424d-bc84-2657eba9929e-kube-api-access-px8lj\") pod \"ovsdbserver-nb-0\" (UID: \"56f9d065-29e4-424d-bc84-2657eba9929e\") " pod="openstack/ovsdbserver-nb-0" Dec 06 03:56:41 crc kubenswrapper[4802]: I1206 03:56:41.959470 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"56f9d065-29e4-424d-bc84-2657eba9929e\") " pod="openstack/ovsdbserver-nb-0" Dec 06 03:56:42 crc kubenswrapper[4802]: I1206 03:56:42.043121 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 06 03:56:43 crc kubenswrapper[4802]: I1206 03:56:43.892968 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 06 03:56:43 crc kubenswrapper[4802]: I1206 03:56:43.900055 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 06 03:56:43 crc kubenswrapper[4802]: I1206 03:56:43.907015 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 06 03:56:43 crc kubenswrapper[4802]: I1206 03:56:43.907486 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 06 03:56:43 crc kubenswrapper[4802]: I1206 03:56:43.907717 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 06 03:56:43 crc kubenswrapper[4802]: I1206 03:56:43.908235 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-xcccg" Dec 06 03:56:43 crc kubenswrapper[4802]: I1206 03:56:43.912643 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 06 03:56:44 crc kubenswrapper[4802]: I1206 03:56:44.068402 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c45b4e0d-c52c-421b-b748-896ebc517784-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"c45b4e0d-c52c-421b-b748-896ebc517784\") " pod="openstack/ovsdbserver-sb-0" Dec 06 03:56:44 crc kubenswrapper[4802]: I1206 03:56:44.068469 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c45b4e0d-c52c-421b-b748-896ebc517784-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"c45b4e0d-c52c-421b-b748-896ebc517784\") " pod="openstack/ovsdbserver-sb-0" Dec 06 03:56:44 crc kubenswrapper[4802]: I1206 03:56:44.068493 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c45b4e0d-c52c-421b-b748-896ebc517784-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"c45b4e0d-c52c-421b-b748-896ebc517784\") " pod="openstack/ovsdbserver-sb-0" Dec 06 03:56:44 crc kubenswrapper[4802]: I1206 03:56:44.068687 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c45b4e0d-c52c-421b-b748-896ebc517784-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"c45b4e0d-c52c-421b-b748-896ebc517784\") " pod="openstack/ovsdbserver-sb-0" Dec 06 03:56:44 crc kubenswrapper[4802]: I1206 03:56:44.068888 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqs29\" (UniqueName: \"kubernetes.io/projected/c45b4e0d-c52c-421b-b748-896ebc517784-kube-api-access-zqs29\") pod \"ovsdbserver-sb-0\" (UID: \"c45b4e0d-c52c-421b-b748-896ebc517784\") " pod="openstack/ovsdbserver-sb-0" Dec 06 03:56:44 crc kubenswrapper[4802]: I1206 03:56:44.068931 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c45b4e0d-c52c-421b-b748-896ebc517784-config\") pod \"ovsdbserver-sb-0\" (UID: \"c45b4e0d-c52c-421b-b748-896ebc517784\") " pod="openstack/ovsdbserver-sb-0" Dec 06 03:56:44 crc kubenswrapper[4802]: I1206 03:56:44.068957 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"c45b4e0d-c52c-421b-b748-896ebc517784\") " pod="openstack/ovsdbserver-sb-0" Dec 06 03:56:44 crc kubenswrapper[4802]: I1206 03:56:44.069074 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c45b4e0d-c52c-421b-b748-896ebc517784-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"c45b4e0d-c52c-421b-b748-896ebc517784\") " pod="openstack/ovsdbserver-sb-0" Dec 06 03:56:44 crc kubenswrapper[4802]: I1206 03:56:44.170431 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"c45b4e0d-c52c-421b-b748-896ebc517784\") " pod="openstack/ovsdbserver-sb-0" Dec 06 03:56:44 crc kubenswrapper[4802]: I1206 03:56:44.170489 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c45b4e0d-c52c-421b-b748-896ebc517784-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"c45b4e0d-c52c-421b-b748-896ebc517784\") " pod="openstack/ovsdbserver-sb-0" Dec 06 03:56:44 crc kubenswrapper[4802]: I1206 03:56:44.170527 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c45b4e0d-c52c-421b-b748-896ebc517784-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"c45b4e0d-c52c-421b-b748-896ebc517784\") " pod="openstack/ovsdbserver-sb-0" Dec 06 03:56:44 crc kubenswrapper[4802]: I1206 03:56:44.170592 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c45b4e0d-c52c-421b-b748-896ebc517784-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"c45b4e0d-c52c-421b-b748-896ebc517784\") " pod="openstack/ovsdbserver-sb-0" Dec 06 03:56:44 crc kubenswrapper[4802]: I1206 03:56:44.170603 4802 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"c45b4e0d-c52c-421b-b748-896ebc517784\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/ovsdbserver-sb-0" Dec 06 03:56:44 crc kubenswrapper[4802]: I1206 03:56:44.171089 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c45b4e0d-c52c-421b-b748-896ebc517784-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"c45b4e0d-c52c-421b-b748-896ebc517784\") " pod="openstack/ovsdbserver-sb-0" Dec 06 03:56:44 crc kubenswrapper[4802]: I1206 03:56:44.170617 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c45b4e0d-c52c-421b-b748-896ebc517784-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"c45b4e0d-c52c-421b-b748-896ebc517784\") " pod="openstack/ovsdbserver-sb-0" Dec 06 03:56:44 crc kubenswrapper[4802]: I1206 03:56:44.171245 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c45b4e0d-c52c-421b-b748-896ebc517784-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"c45b4e0d-c52c-421b-b748-896ebc517784\") " pod="openstack/ovsdbserver-sb-0" Dec 06 03:56:44 crc kubenswrapper[4802]: I1206 03:56:44.171318 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqs29\" (UniqueName: \"kubernetes.io/projected/c45b4e0d-c52c-421b-b748-896ebc517784-kube-api-access-zqs29\") pod \"ovsdbserver-sb-0\" (UID: \"c45b4e0d-c52c-421b-b748-896ebc517784\") " pod="openstack/ovsdbserver-sb-0" Dec 06 03:56:44 crc kubenswrapper[4802]: I1206 03:56:44.171354 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c45b4e0d-c52c-421b-b748-896ebc517784-config\") pod \"ovsdbserver-sb-0\" (UID: \"c45b4e0d-c52c-421b-b748-896ebc517784\") " pod="openstack/ovsdbserver-sb-0" Dec 06 03:56:44 crc kubenswrapper[4802]: I1206 03:56:44.172186 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c45b4e0d-c52c-421b-b748-896ebc517784-config\") pod \"ovsdbserver-sb-0\" (UID: \"c45b4e0d-c52c-421b-b748-896ebc517784\") " pod="openstack/ovsdbserver-sb-0" Dec 06 03:56:44 crc kubenswrapper[4802]: I1206 03:56:44.172446 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c45b4e0d-c52c-421b-b748-896ebc517784-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"c45b4e0d-c52c-421b-b748-896ebc517784\") " pod="openstack/ovsdbserver-sb-0" Dec 06 03:56:44 crc kubenswrapper[4802]: I1206 03:56:44.178419 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c45b4e0d-c52c-421b-b748-896ebc517784-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"c45b4e0d-c52c-421b-b748-896ebc517784\") " pod="openstack/ovsdbserver-sb-0" Dec 06 03:56:44 crc kubenswrapper[4802]: I1206 03:56:44.178912 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c45b4e0d-c52c-421b-b748-896ebc517784-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"c45b4e0d-c52c-421b-b748-896ebc517784\") " pod="openstack/ovsdbserver-sb-0" Dec 06 03:56:44 crc kubenswrapper[4802]: I1206 03:56:44.182443 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c45b4e0d-c52c-421b-b748-896ebc517784-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"c45b4e0d-c52c-421b-b748-896ebc517784\") " pod="openstack/ovsdbserver-sb-0" Dec 06 03:56:44 crc kubenswrapper[4802]: I1206 03:56:44.193303 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"c45b4e0d-c52c-421b-b748-896ebc517784\") " pod="openstack/ovsdbserver-sb-0" Dec 06 03:56:44 crc kubenswrapper[4802]: I1206 03:56:44.202656 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqs29\" (UniqueName: \"kubernetes.io/projected/c45b4e0d-c52c-421b-b748-896ebc517784-kube-api-access-zqs29\") pod \"ovsdbserver-sb-0\" (UID: \"c45b4e0d-c52c-421b-b748-896ebc517784\") " pod="openstack/ovsdbserver-sb-0" Dec 06 03:56:44 crc kubenswrapper[4802]: I1206 03:56:44.236710 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 06 03:56:58 crc kubenswrapper[4802]: E1206 03:56:58.747918 4802 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Dec 06 03:56:58 crc kubenswrapper[4802]: E1206 03:56:58.748489 4802 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9lzq4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(8c8b6c49-6f71-403d-b639-4a17ae3bc802): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 03:56:58 crc kubenswrapper[4802]: E1206 03:56:58.749684 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="8c8b6c49-6f71-403d-b639-4a17ae3bc802" Dec 06 03:56:58 crc kubenswrapper[4802]: E1206 03:56:58.751038 4802 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Dec 06 03:56:58 crc kubenswrapper[4802]: E1206 03:56:58.751374 4802 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cjz5x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(7b4f22a1-41dc-4a75-8999-53df29d2fb5d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 03:56:58 crc kubenswrapper[4802]: E1206 03:56:58.752955 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="7b4f22a1-41dc-4a75-8999-53df29d2fb5d" Dec 06 03:56:59 crc kubenswrapper[4802]: E1206 03:56:59.282330 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-server-0" podUID="7b4f22a1-41dc-4a75-8999-53df29d2fb5d" Dec 06 03:56:59 crc kubenswrapper[4802]: E1206 03:56:59.282496 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="8c8b6c49-6f71-403d-b639-4a17ae3bc802" Dec 06 03:57:05 crc kubenswrapper[4802]: I1206 03:57:05.377237 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 03:57:05 crc kubenswrapper[4802]: E1206 03:57:05.724919 4802 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 06 03:57:05 crc kubenswrapper[4802]: E1206 03:57:05.725106 4802 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-m6f5m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-wtbxc_openstack(e78545cd-e888-4838-9b9b-6008a2c60a40): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 03:57:05 crc kubenswrapper[4802]: E1206 03:57:05.728355 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-wtbxc" podUID="e78545cd-e888-4838-9b9b-6008a2c60a40" Dec 06 03:57:05 crc kubenswrapper[4802]: W1206 03:57:05.728434 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb96722fa_dc6a_44c3_9460_0ce247056613.slice/crio-5224c4b5e7ebf432c7dfc363b364cfd34bf6890a993f567c73e1b1a4d4701034 WatchSource:0}: Error finding container 5224c4b5e7ebf432c7dfc363b364cfd34bf6890a993f567c73e1b1a4d4701034: Status 404 returned error can't find the container with id 5224c4b5e7ebf432c7dfc363b364cfd34bf6890a993f567c73e1b1a4d4701034 Dec 06 03:57:05 crc kubenswrapper[4802]: E1206 03:57:05.766508 4802 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 06 03:57:05 crc kubenswrapper[4802]: E1206 03:57:05.766679 4802 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-756r8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-c5jjr_openstack(68de0f89-2871-4ae7-aa90-e002255f92b8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 03:57:05 crc kubenswrapper[4802]: E1206 03:57:05.768672 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-c5jjr" podUID="68de0f89-2871-4ae7-aa90-e002255f92b8" Dec 06 03:57:05 crc kubenswrapper[4802]: E1206 03:57:05.771878 4802 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 06 03:57:05 crc kubenswrapper[4802]: E1206 03:57:05.772001 4802 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-r5rch,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-jg5xj_openstack(cf7fcfab-7fcb-4e45-87fd-207e25fb80b1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 03:57:05 crc kubenswrapper[4802]: E1206 03:57:05.773326 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-jg5xj" podUID="cf7fcfab-7fcb-4e45-87fd-207e25fb80b1" Dec 06 03:57:05 crc kubenswrapper[4802]: E1206 03:57:05.793661 4802 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 06 03:57:05 crc kubenswrapper[4802]: E1206 03:57:05.793811 4802 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5zdgr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-pg5kn_openstack(b463579f-6ba7-499f-b1e3-65a87f74026e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 03:57:05 crc kubenswrapper[4802]: E1206 03:57:05.801569 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-pg5kn" podUID="b463579f-6ba7-499f-b1e3-65a87f74026e" Dec 06 03:57:06 crc kubenswrapper[4802]: I1206 03:57:06.280768 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-49cxn"] Dec 06 03:57:06 crc kubenswrapper[4802]: W1206 03:57:06.297605 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podedb79ab5_ba10_4379_96e9_b2de42f6cbc8.slice/crio-9aa6542fefae696cc23207b6d68b0b61094c434aae2c1c3805e88a41b957efa4 WatchSource:0}: Error finding container 9aa6542fefae696cc23207b6d68b0b61094c434aae2c1c3805e88a41b957efa4: Status 404 returned error can't find the container with id 9aa6542fefae696cc23207b6d68b0b61094c434aae2c1c3805e88a41b957efa4 Dec 06 03:57:06 crc kubenswrapper[4802]: I1206 03:57:06.333276 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"b96722fa-dc6a-44c3-9460-0ce247056613","Type":"ContainerStarted","Data":"5224c4b5e7ebf432c7dfc363b364cfd34bf6890a993f567c73e1b1a4d4701034"} Dec 06 03:57:06 crc kubenswrapper[4802]: I1206 03:57:06.335283 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"04b6f630-c38a-411e-b48c-99690653972a","Type":"ContainerStarted","Data":"d3a5feb5c21e0f991929986d6b7cac29a11ad2c564dcfd01f301ca670e80e5c5"} Dec 06 03:57:06 crc kubenswrapper[4802]: I1206 03:57:06.338221 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 06 03:57:06 crc kubenswrapper[4802]: I1206 03:57:06.339502 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-49cxn" event={"ID":"edb79ab5-ba10-4379-96e9-b2de42f6cbc8","Type":"ContainerStarted","Data":"9aa6542fefae696cc23207b6d68b0b61094c434aae2c1c3805e88a41b957efa4"} Dec 06 03:57:06 crc kubenswrapper[4802]: E1206 03:57:06.341441 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-666b6646f7-wtbxc" podUID="e78545cd-e888-4838-9b9b-6008a2c60a40" Dec 06 03:57:06 crc kubenswrapper[4802]: E1206 03:57:06.341474 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-c5jjr" podUID="68de0f89-2871-4ae7-aa90-e002255f92b8" Dec 06 03:57:06 crc kubenswrapper[4802]: I1206 03:57:06.350571 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-sbt8p"] Dec 06 03:57:06 crc kubenswrapper[4802]: I1206 03:57:06.356965 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 06 03:57:06 crc kubenswrapper[4802]: I1206 03:57:06.408205 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 06 03:57:06 crc kubenswrapper[4802]: W1206 03:57:06.433910 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc45b4e0d_c52c_421b_b748_896ebc517784.slice/crio-38d7eccb38559732e07e736b74a58c3be3e5458a9f27d214667fa91dcd7852c6 WatchSource:0}: Error finding container 38d7eccb38559732e07e736b74a58c3be3e5458a9f27d214667fa91dcd7852c6: Status 404 returned error can't find the container with id 38d7eccb38559732e07e736b74a58c3be3e5458a9f27d214667fa91dcd7852c6 Dec 06 03:57:06 crc kubenswrapper[4802]: I1206 03:57:06.958968 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-jg5xj" Dec 06 03:57:07 crc kubenswrapper[4802]: I1206 03:57:07.023189 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-pg5kn" Dec 06 03:57:07 crc kubenswrapper[4802]: I1206 03:57:07.039847 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 06 03:57:07 crc kubenswrapper[4802]: I1206 03:57:07.074634 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r5rch\" (UniqueName: \"kubernetes.io/projected/cf7fcfab-7fcb-4e45-87fd-207e25fb80b1-kube-api-access-r5rch\") pod \"cf7fcfab-7fcb-4e45-87fd-207e25fb80b1\" (UID: \"cf7fcfab-7fcb-4e45-87fd-207e25fb80b1\") " Dec 06 03:57:07 crc kubenswrapper[4802]: I1206 03:57:07.074724 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf7fcfab-7fcb-4e45-87fd-207e25fb80b1-config\") pod \"cf7fcfab-7fcb-4e45-87fd-207e25fb80b1\" (UID: \"cf7fcfab-7fcb-4e45-87fd-207e25fb80b1\") " Dec 06 03:57:07 crc kubenswrapper[4802]: I1206 03:57:07.075738 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf7fcfab-7fcb-4e45-87fd-207e25fb80b1-config" (OuterVolumeSpecName: "config") pod "cf7fcfab-7fcb-4e45-87fd-207e25fb80b1" (UID: "cf7fcfab-7fcb-4e45-87fd-207e25fb80b1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:57:07 crc kubenswrapper[4802]: I1206 03:57:07.082584 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf7fcfab-7fcb-4e45-87fd-207e25fb80b1-kube-api-access-r5rch" (OuterVolumeSpecName: "kube-api-access-r5rch") pod "cf7fcfab-7fcb-4e45-87fd-207e25fb80b1" (UID: "cf7fcfab-7fcb-4e45-87fd-207e25fb80b1"). InnerVolumeSpecName "kube-api-access-r5rch". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:57:07 crc kubenswrapper[4802]: I1206 03:57:07.176235 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b463579f-6ba7-499f-b1e3-65a87f74026e-dns-svc\") pod \"b463579f-6ba7-499f-b1e3-65a87f74026e\" (UID: \"b463579f-6ba7-499f-b1e3-65a87f74026e\") " Dec 06 03:57:07 crc kubenswrapper[4802]: I1206 03:57:07.176300 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5zdgr\" (UniqueName: \"kubernetes.io/projected/b463579f-6ba7-499f-b1e3-65a87f74026e-kube-api-access-5zdgr\") pod \"b463579f-6ba7-499f-b1e3-65a87f74026e\" (UID: \"b463579f-6ba7-499f-b1e3-65a87f74026e\") " Dec 06 03:57:07 crc kubenswrapper[4802]: I1206 03:57:07.176334 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b463579f-6ba7-499f-b1e3-65a87f74026e-config\") pod \"b463579f-6ba7-499f-b1e3-65a87f74026e\" (UID: \"b463579f-6ba7-499f-b1e3-65a87f74026e\") " Dec 06 03:57:07 crc kubenswrapper[4802]: I1206 03:57:07.176677 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r5rch\" (UniqueName: \"kubernetes.io/projected/cf7fcfab-7fcb-4e45-87fd-207e25fb80b1-kube-api-access-r5rch\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:07 crc kubenswrapper[4802]: I1206 03:57:07.176688 4802 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf7fcfab-7fcb-4e45-87fd-207e25fb80b1-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:07 crc kubenswrapper[4802]: I1206 03:57:07.176883 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b463579f-6ba7-499f-b1e3-65a87f74026e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b463579f-6ba7-499f-b1e3-65a87f74026e" (UID: "b463579f-6ba7-499f-b1e3-65a87f74026e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:57:07 crc kubenswrapper[4802]: I1206 03:57:07.177165 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b463579f-6ba7-499f-b1e3-65a87f74026e-config" (OuterVolumeSpecName: "config") pod "b463579f-6ba7-499f-b1e3-65a87f74026e" (UID: "b463579f-6ba7-499f-b1e3-65a87f74026e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:57:07 crc kubenswrapper[4802]: I1206 03:57:07.179741 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b463579f-6ba7-499f-b1e3-65a87f74026e-kube-api-access-5zdgr" (OuterVolumeSpecName: "kube-api-access-5zdgr") pod "b463579f-6ba7-499f-b1e3-65a87f74026e" (UID: "b463579f-6ba7-499f-b1e3-65a87f74026e"). InnerVolumeSpecName "kube-api-access-5zdgr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:57:07 crc kubenswrapper[4802]: W1206 03:57:07.198821 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod56f9d065_29e4_424d_bc84_2657eba9929e.slice/crio-6f393a546a4a9b431dc9f1c6d1c55c2985aff5ef4476b2a5be2854d8186f96b9 WatchSource:0}: Error finding container 6f393a546a4a9b431dc9f1c6d1c55c2985aff5ef4476b2a5be2854d8186f96b9: Status 404 returned error can't find the container with id 6f393a546a4a9b431dc9f1c6d1c55c2985aff5ef4476b2a5be2854d8186f96b9 Dec 06 03:57:07 crc kubenswrapper[4802]: I1206 03:57:07.278154 4802 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b463579f-6ba7-499f-b1e3-65a87f74026e-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:07 crc kubenswrapper[4802]: I1206 03:57:07.278185 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5zdgr\" (UniqueName: \"kubernetes.io/projected/b463579f-6ba7-499f-b1e3-65a87f74026e-kube-api-access-5zdgr\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:07 crc kubenswrapper[4802]: I1206 03:57:07.278196 4802 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b463579f-6ba7-499f-b1e3-65a87f74026e-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:07 crc kubenswrapper[4802]: I1206 03:57:07.346939 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"db5b3080-8b4d-4232-a7da-66e544c09c60","Type":"ContainerStarted","Data":"78f0dec7ab3001067e560781f3ee07ad795ba770d4a25b94868258becb4b1b7d"} Dec 06 03:57:07 crc kubenswrapper[4802]: I1206 03:57:07.346979 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"db5b3080-8b4d-4232-a7da-66e544c09c60","Type":"ContainerStarted","Data":"05d1b0ce481fda2f162c9206edb43ae4171d2f7cf621f6ee2430dd509f13e14c"} Dec 06 03:57:07 crc kubenswrapper[4802]: I1206 03:57:07.349596 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-pg5kn" event={"ID":"b463579f-6ba7-499f-b1e3-65a87f74026e","Type":"ContainerDied","Data":"34afa2baa744284f63890657da95d6001a9024dc15edca350eab649edc97ebce"} Dec 06 03:57:07 crc kubenswrapper[4802]: I1206 03:57:07.349644 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-pg5kn" Dec 06 03:57:07 crc kubenswrapper[4802]: I1206 03:57:07.362305 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"654f214b-8467-41d2-93f4-3c18377653b9","Type":"ContainerStarted","Data":"289d40e93efc3583251e8232ee572bff94c45ac9bd8f5f0bb0fc5b45c0399011"} Dec 06 03:57:07 crc kubenswrapper[4802]: I1206 03:57:07.371165 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"56f9d065-29e4-424d-bc84-2657eba9929e","Type":"ContainerStarted","Data":"6f393a546a4a9b431dc9f1c6d1c55c2985aff5ef4476b2a5be2854d8186f96b9"} Dec 06 03:57:07 crc kubenswrapper[4802]: I1206 03:57:07.383352 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-jg5xj" Dec 06 03:57:07 crc kubenswrapper[4802]: I1206 03:57:07.383343 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-jg5xj" event={"ID":"cf7fcfab-7fcb-4e45-87fd-207e25fb80b1","Type":"ContainerDied","Data":"d2c9a022b9615bf0435a6bc22c9ecfcb642ef14a8fb22374da7eb0d4dc2e968d"} Dec 06 03:57:07 crc kubenswrapper[4802]: I1206 03:57:07.388860 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"c45b4e0d-c52c-421b-b748-896ebc517784","Type":"ContainerStarted","Data":"38d7eccb38559732e07e736b74a58c3be3e5458a9f27d214667fa91dcd7852c6"} Dec 06 03:57:07 crc kubenswrapper[4802]: I1206 03:57:07.405997 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-sbt8p" event={"ID":"6e785e69-07cd-4c5c-8446-3d2a0d7e7b0b","Type":"ContainerStarted","Data":"90b29d1344ac68f52fa6e5c5f26ea5343d24eba9453ea111e695b4db8a10a6ea"} Dec 06 03:57:07 crc kubenswrapper[4802]: I1206 03:57:07.417389 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-pg5kn"] Dec 06 03:57:07 crc kubenswrapper[4802]: I1206 03:57:07.444097 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-pg5kn"] Dec 06 03:57:07 crc kubenswrapper[4802]: I1206 03:57:07.460964 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b463579f-6ba7-499f-b1e3-65a87f74026e" path="/var/lib/kubelet/pods/b463579f-6ba7-499f-b1e3-65a87f74026e/volumes" Dec 06 03:57:07 crc kubenswrapper[4802]: I1206 03:57:07.463775 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-jg5xj"] Dec 06 03:57:07 crc kubenswrapper[4802]: I1206 03:57:07.472364 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-jg5xj"] Dec 06 03:57:09 crc kubenswrapper[4802]: I1206 03:57:09.465724 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf7fcfab-7fcb-4e45-87fd-207e25fb80b1" path="/var/lib/kubelet/pods/cf7fcfab-7fcb-4e45-87fd-207e25fb80b1/volumes" Dec 06 03:57:10 crc kubenswrapper[4802]: I1206 03:57:10.429728 4802 generic.go:334] "Generic (PLEG): container finished" podID="04b6f630-c38a-411e-b48c-99690653972a" containerID="d3a5feb5c21e0f991929986d6b7cac29a11ad2c564dcfd01f301ca670e80e5c5" exitCode=0 Dec 06 03:57:10 crc kubenswrapper[4802]: I1206 03:57:10.429810 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"04b6f630-c38a-411e-b48c-99690653972a","Type":"ContainerDied","Data":"d3a5feb5c21e0f991929986d6b7cac29a11ad2c564dcfd01f301ca670e80e5c5"} Dec 06 03:57:15 crc kubenswrapper[4802]: I1206 03:57:15.473967 4802 generic.go:334] "Generic (PLEG): container finished" podID="db5b3080-8b4d-4232-a7da-66e544c09c60" containerID="78f0dec7ab3001067e560781f3ee07ad795ba770d4a25b94868258becb4b1b7d" exitCode=0 Dec 06 03:57:15 crc kubenswrapper[4802]: I1206 03:57:15.474026 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"db5b3080-8b4d-4232-a7da-66e544c09c60","Type":"ContainerDied","Data":"78f0dec7ab3001067e560781f3ee07ad795ba770d4a25b94868258becb4b1b7d"} Dec 06 03:57:16 crc kubenswrapper[4802]: I1206 03:57:16.482092 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"654f214b-8467-41d2-93f4-3c18377653b9","Type":"ContainerStarted","Data":"e69dd5abad2a3a267a25bd0d6a6712e91d5cb4a9b8268d3064d2ef9fc6ceb332"} Dec 06 03:57:16 crc kubenswrapper[4802]: I1206 03:57:16.482722 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 06 03:57:16 crc kubenswrapper[4802]: I1206 03:57:16.483897 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"c45b4e0d-c52c-421b-b748-896ebc517784","Type":"ContainerStarted","Data":"6bc38b1fdbe7c92baa5f5fa4bf00d335cf81c9bf442521f9e289f197429e45ba"} Dec 06 03:57:16 crc kubenswrapper[4802]: I1206 03:57:16.485348 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"56f9d065-29e4-424d-bc84-2657eba9929e","Type":"ContainerStarted","Data":"510aea7f6029504f46f3e04a46ce09fa99da1648105088081a78957143ea9229"} Dec 06 03:57:16 crc kubenswrapper[4802]: I1206 03:57:16.487118 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"b96722fa-dc6a-44c3-9460-0ce247056613","Type":"ContainerStarted","Data":"44f31f10d271b5bc0656d1123c02d51de86d32004e13b130eeb6e5eb02a544e8"} Dec 06 03:57:16 crc kubenswrapper[4802]: I1206 03:57:16.487296 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 06 03:57:16 crc kubenswrapper[4802]: I1206 03:57:16.488373 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-sbt8p" event={"ID":"6e785e69-07cd-4c5c-8446-3d2a0d7e7b0b","Type":"ContainerStarted","Data":"d1facf1d0987c642a9d347519160f04d4ecc1218cd87bfec69e9bf3e01cd7939"} Dec 06 03:57:16 crc kubenswrapper[4802]: I1206 03:57:16.488824 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-sbt8p" Dec 06 03:57:16 crc kubenswrapper[4802]: I1206 03:57:16.490466 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"db5b3080-8b4d-4232-a7da-66e544c09c60","Type":"ContainerStarted","Data":"fa659d4910b151795ff5d79855227666df7376486c9175630a211ca7ed0315f5"} Dec 06 03:57:16 crc kubenswrapper[4802]: I1206 03:57:16.491917 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"04b6f630-c38a-411e-b48c-99690653972a","Type":"ContainerStarted","Data":"d0e3ba3bc04458439af7c02006eb91e8a5e586c52f11402ab51871692b46991c"} Dec 06 03:57:16 crc kubenswrapper[4802]: I1206 03:57:16.493074 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-49cxn" event={"ID":"edb79ab5-ba10-4379-96e9-b2de42f6cbc8","Type":"ContainerStarted","Data":"c5f7e8f3076092f55df43d181c9bed07f78a01db96c335e42cf3f8a3104100ed"} Dec 06 03:57:16 crc kubenswrapper[4802]: I1206 03:57:16.506234 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=32.083854882 podStartE2EDuration="41.506214902s" podCreationTimestamp="2025-12-06 03:56:35 +0000 UTC" firstStartedPulling="2025-12-06 03:57:06.350988966 +0000 UTC m=+1019.222898118" lastFinishedPulling="2025-12-06 03:57:15.773348986 +0000 UTC m=+1028.645258138" observedRunningTime="2025-12-06 03:57:16.499134798 +0000 UTC m=+1029.371043950" watchObservedRunningTime="2025-12-06 03:57:16.506214902 +0000 UTC m=+1029.378124054" Dec 06 03:57:16 crc kubenswrapper[4802]: I1206 03:57:16.521716 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-sbt8p" podStartSLOduration=27.061202362 podStartE2EDuration="36.521684884s" podCreationTimestamp="2025-12-06 03:56:40 +0000 UTC" firstStartedPulling="2025-12-06 03:57:06.384070089 +0000 UTC m=+1019.255979241" lastFinishedPulling="2025-12-06 03:57:15.844552611 +0000 UTC m=+1028.716461763" observedRunningTime="2025-12-06 03:57:16.514247461 +0000 UTC m=+1029.386156623" watchObservedRunningTime="2025-12-06 03:57:16.521684884 +0000 UTC m=+1029.393594036" Dec 06 03:57:16 crc kubenswrapper[4802]: I1206 03:57:16.540004 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=13.459671942 podStartE2EDuration="44.539982095s" podCreationTimestamp="2025-12-06 03:56:32 +0000 UTC" firstStartedPulling="2025-12-06 03:56:34.686815548 +0000 UTC m=+987.558724700" lastFinishedPulling="2025-12-06 03:57:05.767125681 +0000 UTC m=+1018.639034853" observedRunningTime="2025-12-06 03:57:16.531492223 +0000 UTC m=+1029.403401375" watchObservedRunningTime="2025-12-06 03:57:16.539982095 +0000 UTC m=+1029.411891267" Dec 06 03:57:16 crc kubenswrapper[4802]: I1206 03:57:16.565831 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=30.452939101 podStartE2EDuration="40.56580406s" podCreationTimestamp="2025-12-06 03:56:36 +0000 UTC" firstStartedPulling="2025-12-06 03:57:05.730969983 +0000 UTC m=+1018.602879145" lastFinishedPulling="2025-12-06 03:57:15.843834952 +0000 UTC m=+1028.715744104" observedRunningTime="2025-12-06 03:57:16.561260026 +0000 UTC m=+1029.433169178" watchObservedRunningTime="2025-12-06 03:57:16.56580406 +0000 UTC m=+1029.437713242" Dec 06 03:57:16 crc kubenswrapper[4802]: I1206 03:57:16.611886 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=43.611855368 podStartE2EDuration="43.611855368s" podCreationTimestamp="2025-12-06 03:56:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:57:16.607912481 +0000 UTC m=+1029.479821713" watchObservedRunningTime="2025-12-06 03:57:16.611855368 +0000 UTC m=+1029.483764530" Dec 06 03:57:17 crc kubenswrapper[4802]: I1206 03:57:17.502180 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8c8b6c49-6f71-403d-b639-4a17ae3bc802","Type":"ContainerStarted","Data":"f470e3dda7735fd4bf057dbf62cde306583097f22e2d9c550167fc04c5e92f72"} Dec 06 03:57:17 crc kubenswrapper[4802]: I1206 03:57:17.503707 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7b4f22a1-41dc-4a75-8999-53df29d2fb5d","Type":"ContainerStarted","Data":"fd1082f2aaf9365d3d9655c6924fa59bdb4c429364af5f69505c71489894ed4f"} Dec 06 03:57:17 crc kubenswrapper[4802]: I1206 03:57:17.505865 4802 generic.go:334] "Generic (PLEG): container finished" podID="edb79ab5-ba10-4379-96e9-b2de42f6cbc8" containerID="c5f7e8f3076092f55df43d181c9bed07f78a01db96c335e42cf3f8a3104100ed" exitCode=0 Dec 06 03:57:17 crc kubenswrapper[4802]: I1206 03:57:17.505957 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-49cxn" event={"ID":"edb79ab5-ba10-4379-96e9-b2de42f6cbc8","Type":"ContainerDied","Data":"c5f7e8f3076092f55df43d181c9bed07f78a01db96c335e42cf3f8a3104100ed"} Dec 06 03:57:18 crc kubenswrapper[4802]: I1206 03:57:18.514061 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-49cxn" event={"ID":"edb79ab5-ba10-4379-96e9-b2de42f6cbc8","Type":"ContainerStarted","Data":"f9fff09a517238920be790c4952478cea0d4d9b0a3bc5234e6092014c867a053"} Dec 06 03:57:19 crc kubenswrapper[4802]: I1206 03:57:19.525286 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"56f9d065-29e4-424d-bc84-2657eba9929e","Type":"ContainerStarted","Data":"1b42276f89435f637443c9236eb9148bd35144595a89a676675045ed0952264c"} Dec 06 03:57:19 crc kubenswrapper[4802]: I1206 03:57:19.531244 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"c45b4e0d-c52c-421b-b748-896ebc517784","Type":"ContainerStarted","Data":"b13250cea2e63f986539e4c34ec01917cdec8b1d4a60280f33535e6105863702"} Dec 06 03:57:19 crc kubenswrapper[4802]: I1206 03:57:19.535007 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-49cxn" event={"ID":"edb79ab5-ba10-4379-96e9-b2de42f6cbc8","Type":"ContainerStarted","Data":"4304624eae3ced4b88ef38c3deda9e2ffb0bfa61d946a297c7043486e008a270"} Dec 06 03:57:19 crc kubenswrapper[4802]: I1206 03:57:19.535320 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-49cxn" Dec 06 03:57:19 crc kubenswrapper[4802]: I1206 03:57:19.535370 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-49cxn" Dec 06 03:57:19 crc kubenswrapper[4802]: I1206 03:57:19.579135 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=27.824574911 podStartE2EDuration="39.579117069s" podCreationTimestamp="2025-12-06 03:56:40 +0000 UTC" firstStartedPulling="2025-12-06 03:57:07.202597866 +0000 UTC m=+1020.074507018" lastFinishedPulling="2025-12-06 03:57:18.957140024 +0000 UTC m=+1031.829049176" observedRunningTime="2025-12-06 03:57:19.551961418 +0000 UTC m=+1032.423870600" watchObservedRunningTime="2025-12-06 03:57:19.579117069 +0000 UTC m=+1032.451026221" Dec 06 03:57:19 crc kubenswrapper[4802]: I1206 03:57:19.580340 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-49cxn" podStartSLOduration=30.130923204 podStartE2EDuration="39.580331993s" podCreationTimestamp="2025-12-06 03:56:40 +0000 UTC" firstStartedPulling="2025-12-06 03:57:06.302050119 +0000 UTC m=+1019.173959281" lastFinishedPulling="2025-12-06 03:57:15.751458888 +0000 UTC m=+1028.623368070" observedRunningTime="2025-12-06 03:57:19.576741725 +0000 UTC m=+1032.448650897" watchObservedRunningTime="2025-12-06 03:57:19.580331993 +0000 UTC m=+1032.452241155" Dec 06 03:57:19 crc kubenswrapper[4802]: I1206 03:57:19.612965 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=25.114002276 podStartE2EDuration="37.612940195s" podCreationTimestamp="2025-12-06 03:56:42 +0000 UTC" firstStartedPulling="2025-12-06 03:57:06.442373733 +0000 UTC m=+1019.314282885" lastFinishedPulling="2025-12-06 03:57:18.941311632 +0000 UTC m=+1031.813220804" observedRunningTime="2025-12-06 03:57:19.611132075 +0000 UTC m=+1032.483041247" watchObservedRunningTime="2025-12-06 03:57:19.612940195 +0000 UTC m=+1032.484849387" Dec 06 03:57:20 crc kubenswrapper[4802]: I1206 03:57:20.237797 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 06 03:57:20 crc kubenswrapper[4802]: I1206 03:57:20.278276 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 06 03:57:20 crc kubenswrapper[4802]: I1206 03:57:20.544479 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 06 03:57:21 crc kubenswrapper[4802]: I1206 03:57:21.043636 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 06 03:57:21 crc kubenswrapper[4802]: I1206 03:57:21.083067 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 06 03:57:21 crc kubenswrapper[4802]: I1206 03:57:21.553034 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 06 03:57:21 crc kubenswrapper[4802]: I1206 03:57:21.606898 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 06 03:57:21 crc kubenswrapper[4802]: I1206 03:57:21.612383 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 06 03:57:21 crc kubenswrapper[4802]: I1206 03:57:21.774588 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-c5jjr"] Dec 06 03:57:21 crc kubenswrapper[4802]: I1206 03:57:21.806949 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-6mzpl"] Dec 06 03:57:21 crc kubenswrapper[4802]: I1206 03:57:21.808346 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-6mzpl" Dec 06 03:57:21 crc kubenswrapper[4802]: I1206 03:57:21.810561 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 06 03:57:21 crc kubenswrapper[4802]: I1206 03:57:21.863911 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-6mzpl"] Dec 06 03:57:21 crc kubenswrapper[4802]: I1206 03:57:21.920850 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/742a77a2-cc74-4671-9247-18f27582c980-config\") pod \"dnsmasq-dns-6bc7876d45-6mzpl\" (UID: \"742a77a2-cc74-4671-9247-18f27582c980\") " pod="openstack/dnsmasq-dns-6bc7876d45-6mzpl" Dec 06 03:57:21 crc kubenswrapper[4802]: I1206 03:57:21.920904 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/742a77a2-cc74-4671-9247-18f27582c980-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc7876d45-6mzpl\" (UID: \"742a77a2-cc74-4671-9247-18f27582c980\") " pod="openstack/dnsmasq-dns-6bc7876d45-6mzpl" Dec 06 03:57:21 crc kubenswrapper[4802]: I1206 03:57:21.920935 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/742a77a2-cc74-4671-9247-18f27582c980-dns-svc\") pod \"dnsmasq-dns-6bc7876d45-6mzpl\" (UID: \"742a77a2-cc74-4671-9247-18f27582c980\") " pod="openstack/dnsmasq-dns-6bc7876d45-6mzpl" Dec 06 03:57:21 crc kubenswrapper[4802]: I1206 03:57:21.921006 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xw4pn\" (UniqueName: \"kubernetes.io/projected/742a77a2-cc74-4671-9247-18f27582c980-kube-api-access-xw4pn\") pod \"dnsmasq-dns-6bc7876d45-6mzpl\" (UID: \"742a77a2-cc74-4671-9247-18f27582c980\") " pod="openstack/dnsmasq-dns-6bc7876d45-6mzpl" Dec 06 03:57:21 crc kubenswrapper[4802]: I1206 03:57:21.930018 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-7sljg"] Dec 06 03:57:21 crc kubenswrapper[4802]: I1206 03:57:21.938864 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-7sljg" Dec 06 03:57:21 crc kubenswrapper[4802]: I1206 03:57:21.945875 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 06 03:57:21 crc kubenswrapper[4802]: I1206 03:57:21.958522 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-7sljg"] Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.003896 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-wtbxc"] Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.006827 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.008927 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.011806 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-nnvvk" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.012010 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.013310 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.013312 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.029541 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xw4pn\" (UniqueName: \"kubernetes.io/projected/742a77a2-cc74-4671-9247-18f27582c980-kube-api-access-xw4pn\") pod \"dnsmasq-dns-6bc7876d45-6mzpl\" (UID: \"742a77a2-cc74-4671-9247-18f27582c980\") " pod="openstack/dnsmasq-dns-6bc7876d45-6mzpl" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.029710 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11e2072e-0127-48c0-b417-ced9cfa9bf64-combined-ca-bundle\") pod \"ovn-controller-metrics-7sljg\" (UID: \"11e2072e-0127-48c0-b417-ced9cfa9bf64\") " pod="openstack/ovn-controller-metrics-7sljg" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.029872 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vbsls\" (UniqueName: \"kubernetes.io/projected/11e2072e-0127-48c0-b417-ced9cfa9bf64-kube-api-access-vbsls\") pod \"ovn-controller-metrics-7sljg\" (UID: \"11e2072e-0127-48c0-b417-ced9cfa9bf64\") " pod="openstack/ovn-controller-metrics-7sljg" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.033049 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/11e2072e-0127-48c0-b417-ced9cfa9bf64-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-7sljg\" (UID: \"11e2072e-0127-48c0-b417-ced9cfa9bf64\") " pod="openstack/ovn-controller-metrics-7sljg" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.033089 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/742a77a2-cc74-4671-9247-18f27582c980-config\") pod \"dnsmasq-dns-6bc7876d45-6mzpl\" (UID: \"742a77a2-cc74-4671-9247-18f27582c980\") " pod="openstack/dnsmasq-dns-6bc7876d45-6mzpl" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.033111 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/742a77a2-cc74-4671-9247-18f27582c980-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc7876d45-6mzpl\" (UID: \"742a77a2-cc74-4671-9247-18f27582c980\") " pod="openstack/dnsmasq-dns-6bc7876d45-6mzpl" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.033132 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/742a77a2-cc74-4671-9247-18f27582c980-dns-svc\") pod \"dnsmasq-dns-6bc7876d45-6mzpl\" (UID: \"742a77a2-cc74-4671-9247-18f27582c980\") " pod="openstack/dnsmasq-dns-6bc7876d45-6mzpl" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.033173 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/11e2072e-0127-48c0-b417-ced9cfa9bf64-ovs-rundir\") pod \"ovn-controller-metrics-7sljg\" (UID: \"11e2072e-0127-48c0-b417-ced9cfa9bf64\") " pod="openstack/ovn-controller-metrics-7sljg" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.033212 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/11e2072e-0127-48c0-b417-ced9cfa9bf64-ovn-rundir\") pod \"ovn-controller-metrics-7sljg\" (UID: \"11e2072e-0127-48c0-b417-ced9cfa9bf64\") " pod="openstack/ovn-controller-metrics-7sljg" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.033227 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11e2072e-0127-48c0-b417-ced9cfa9bf64-config\") pod \"ovn-controller-metrics-7sljg\" (UID: \"11e2072e-0127-48c0-b417-ced9cfa9bf64\") " pod="openstack/ovn-controller-metrics-7sljg" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.033591 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.034443 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/742a77a2-cc74-4671-9247-18f27582c980-config\") pod \"dnsmasq-dns-6bc7876d45-6mzpl\" (UID: \"742a77a2-cc74-4671-9247-18f27582c980\") " pod="openstack/dnsmasq-dns-6bc7876d45-6mzpl" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.034537 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/742a77a2-cc74-4671-9247-18f27582c980-dns-svc\") pod \"dnsmasq-dns-6bc7876d45-6mzpl\" (UID: \"742a77a2-cc74-4671-9247-18f27582c980\") " pod="openstack/dnsmasq-dns-6bc7876d45-6mzpl" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.035095 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/742a77a2-cc74-4671-9247-18f27582c980-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc7876d45-6mzpl\" (UID: \"742a77a2-cc74-4671-9247-18f27582c980\") " pod="openstack/dnsmasq-dns-6bc7876d45-6mzpl" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.048869 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8554648995-4dwc7"] Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.050875 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-4dwc7" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.060100 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.067735 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xw4pn\" (UniqueName: \"kubernetes.io/projected/742a77a2-cc74-4671-9247-18f27582c980-kube-api-access-xw4pn\") pod \"dnsmasq-dns-6bc7876d45-6mzpl\" (UID: \"742a77a2-cc74-4671-9247-18f27582c980\") " pod="openstack/dnsmasq-dns-6bc7876d45-6mzpl" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.116091 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-4dwc7"] Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.127182 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-6mzpl" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.137574 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/11e2072e-0127-48c0-b417-ced9cfa9bf64-ovs-rundir\") pod \"ovn-controller-metrics-7sljg\" (UID: \"11e2072e-0127-48c0-b417-ced9cfa9bf64\") " pod="openstack/ovn-controller-metrics-7sljg" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.137647 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11e2072e-0127-48c0-b417-ced9cfa9bf64-config\") pod \"ovn-controller-metrics-7sljg\" (UID: \"11e2072e-0127-48c0-b417-ced9cfa9bf64\") " pod="openstack/ovn-controller-metrics-7sljg" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.137674 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/11e2072e-0127-48c0-b417-ced9cfa9bf64-ovn-rundir\") pod \"ovn-controller-metrics-7sljg\" (UID: \"11e2072e-0127-48c0-b417-ced9cfa9bf64\") " pod="openstack/ovn-controller-metrics-7sljg" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.137715 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkcxt\" (UniqueName: \"kubernetes.io/projected/31be02e0-3b12-49e9-a900-3d069fa35ac4-kube-api-access-kkcxt\") pod \"dnsmasq-dns-8554648995-4dwc7\" (UID: \"31be02e0-3b12-49e9-a900-3d069fa35ac4\") " pod="openstack/dnsmasq-dns-8554648995-4dwc7" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.137745 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/dca5fd9f-3aaf-4fd7-96d0-7f78c38c12b6-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"dca5fd9f-3aaf-4fd7-96d0-7f78c38c12b6\") " pod="openstack/ovn-northd-0" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.137788 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11e2072e-0127-48c0-b417-ced9cfa9bf64-combined-ca-bundle\") pod \"ovn-controller-metrics-7sljg\" (UID: \"11e2072e-0127-48c0-b417-ced9cfa9bf64\") " pod="openstack/ovn-controller-metrics-7sljg" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.137815 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vbsls\" (UniqueName: \"kubernetes.io/projected/11e2072e-0127-48c0-b417-ced9cfa9bf64-kube-api-access-vbsls\") pod \"ovn-controller-metrics-7sljg\" (UID: \"11e2072e-0127-48c0-b417-ced9cfa9bf64\") " pod="openstack/ovn-controller-metrics-7sljg" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.137852 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/dca5fd9f-3aaf-4fd7-96d0-7f78c38c12b6-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"dca5fd9f-3aaf-4fd7-96d0-7f78c38c12b6\") " pod="openstack/ovn-northd-0" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.137891 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtcj2\" (UniqueName: \"kubernetes.io/projected/dca5fd9f-3aaf-4fd7-96d0-7f78c38c12b6-kube-api-access-rtcj2\") pod \"ovn-northd-0\" (UID: \"dca5fd9f-3aaf-4fd7-96d0-7f78c38c12b6\") " pod="openstack/ovn-northd-0" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.137922 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/31be02e0-3b12-49e9-a900-3d069fa35ac4-dns-svc\") pod \"dnsmasq-dns-8554648995-4dwc7\" (UID: \"31be02e0-3b12-49e9-a900-3d069fa35ac4\") " pod="openstack/dnsmasq-dns-8554648995-4dwc7" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.137948 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/dca5fd9f-3aaf-4fd7-96d0-7f78c38c12b6-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"dca5fd9f-3aaf-4fd7-96d0-7f78c38c12b6\") " pod="openstack/ovn-northd-0" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.137969 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/31be02e0-3b12-49e9-a900-3d069fa35ac4-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-4dwc7\" (UID: \"31be02e0-3b12-49e9-a900-3d069fa35ac4\") " pod="openstack/dnsmasq-dns-8554648995-4dwc7" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.137992 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dca5fd9f-3aaf-4fd7-96d0-7f78c38c12b6-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"dca5fd9f-3aaf-4fd7-96d0-7f78c38c12b6\") " pod="openstack/ovn-northd-0" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.138012 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/11e2072e-0127-48c0-b417-ced9cfa9bf64-ovs-rundir\") pod \"ovn-controller-metrics-7sljg\" (UID: \"11e2072e-0127-48c0-b417-ced9cfa9bf64\") " pod="openstack/ovn-controller-metrics-7sljg" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.138029 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/11e2072e-0127-48c0-b417-ced9cfa9bf64-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-7sljg\" (UID: \"11e2072e-0127-48c0-b417-ced9cfa9bf64\") " pod="openstack/ovn-controller-metrics-7sljg" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.138093 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dca5fd9f-3aaf-4fd7-96d0-7f78c38c12b6-config\") pod \"ovn-northd-0\" (UID: \"dca5fd9f-3aaf-4fd7-96d0-7f78c38c12b6\") " pod="openstack/ovn-northd-0" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.138131 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/31be02e0-3b12-49e9-a900-3d069fa35ac4-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-4dwc7\" (UID: \"31be02e0-3b12-49e9-a900-3d069fa35ac4\") " pod="openstack/dnsmasq-dns-8554648995-4dwc7" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.138162 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31be02e0-3b12-49e9-a900-3d069fa35ac4-config\") pod \"dnsmasq-dns-8554648995-4dwc7\" (UID: \"31be02e0-3b12-49e9-a900-3d069fa35ac4\") " pod="openstack/dnsmasq-dns-8554648995-4dwc7" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.138190 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dca5fd9f-3aaf-4fd7-96d0-7f78c38c12b6-scripts\") pod \"ovn-northd-0\" (UID: \"dca5fd9f-3aaf-4fd7-96d0-7f78c38c12b6\") " pod="openstack/ovn-northd-0" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.138344 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/11e2072e-0127-48c0-b417-ced9cfa9bf64-ovn-rundir\") pod \"ovn-controller-metrics-7sljg\" (UID: \"11e2072e-0127-48c0-b417-ced9cfa9bf64\") " pod="openstack/ovn-controller-metrics-7sljg" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.138544 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11e2072e-0127-48c0-b417-ced9cfa9bf64-config\") pod \"ovn-controller-metrics-7sljg\" (UID: \"11e2072e-0127-48c0-b417-ced9cfa9bf64\") " pod="openstack/ovn-controller-metrics-7sljg" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.143349 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/11e2072e-0127-48c0-b417-ced9cfa9bf64-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-7sljg\" (UID: \"11e2072e-0127-48c0-b417-ced9cfa9bf64\") " pod="openstack/ovn-controller-metrics-7sljg" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.147329 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11e2072e-0127-48c0-b417-ced9cfa9bf64-combined-ca-bundle\") pod \"ovn-controller-metrics-7sljg\" (UID: \"11e2072e-0127-48c0-b417-ced9cfa9bf64\") " pod="openstack/ovn-controller-metrics-7sljg" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.159198 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vbsls\" (UniqueName: \"kubernetes.io/projected/11e2072e-0127-48c0-b417-ced9cfa9bf64-kube-api-access-vbsls\") pod \"ovn-controller-metrics-7sljg\" (UID: \"11e2072e-0127-48c0-b417-ced9cfa9bf64\") " pod="openstack/ovn-controller-metrics-7sljg" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.240326 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkcxt\" (UniqueName: \"kubernetes.io/projected/31be02e0-3b12-49e9-a900-3d069fa35ac4-kube-api-access-kkcxt\") pod \"dnsmasq-dns-8554648995-4dwc7\" (UID: \"31be02e0-3b12-49e9-a900-3d069fa35ac4\") " pod="openstack/dnsmasq-dns-8554648995-4dwc7" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.240575 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/dca5fd9f-3aaf-4fd7-96d0-7f78c38c12b6-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"dca5fd9f-3aaf-4fd7-96d0-7f78c38c12b6\") " pod="openstack/ovn-northd-0" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.240616 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/dca5fd9f-3aaf-4fd7-96d0-7f78c38c12b6-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"dca5fd9f-3aaf-4fd7-96d0-7f78c38c12b6\") " pod="openstack/ovn-northd-0" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.240648 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtcj2\" (UniqueName: \"kubernetes.io/projected/dca5fd9f-3aaf-4fd7-96d0-7f78c38c12b6-kube-api-access-rtcj2\") pod \"ovn-northd-0\" (UID: \"dca5fd9f-3aaf-4fd7-96d0-7f78c38c12b6\") " pod="openstack/ovn-northd-0" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.240676 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/dca5fd9f-3aaf-4fd7-96d0-7f78c38c12b6-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"dca5fd9f-3aaf-4fd7-96d0-7f78c38c12b6\") " pod="openstack/ovn-northd-0" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.240693 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/31be02e0-3b12-49e9-a900-3d069fa35ac4-dns-svc\") pod \"dnsmasq-dns-8554648995-4dwc7\" (UID: \"31be02e0-3b12-49e9-a900-3d069fa35ac4\") " pod="openstack/dnsmasq-dns-8554648995-4dwc7" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.240706 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/31be02e0-3b12-49e9-a900-3d069fa35ac4-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-4dwc7\" (UID: \"31be02e0-3b12-49e9-a900-3d069fa35ac4\") " pod="openstack/dnsmasq-dns-8554648995-4dwc7" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.240724 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dca5fd9f-3aaf-4fd7-96d0-7f78c38c12b6-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"dca5fd9f-3aaf-4fd7-96d0-7f78c38c12b6\") " pod="openstack/ovn-northd-0" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.240769 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dca5fd9f-3aaf-4fd7-96d0-7f78c38c12b6-config\") pod \"ovn-northd-0\" (UID: \"dca5fd9f-3aaf-4fd7-96d0-7f78c38c12b6\") " pod="openstack/ovn-northd-0" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.240790 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/31be02e0-3b12-49e9-a900-3d069fa35ac4-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-4dwc7\" (UID: \"31be02e0-3b12-49e9-a900-3d069fa35ac4\") " pod="openstack/dnsmasq-dns-8554648995-4dwc7" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.240811 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31be02e0-3b12-49e9-a900-3d069fa35ac4-config\") pod \"dnsmasq-dns-8554648995-4dwc7\" (UID: \"31be02e0-3b12-49e9-a900-3d069fa35ac4\") " pod="openstack/dnsmasq-dns-8554648995-4dwc7" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.240835 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dca5fd9f-3aaf-4fd7-96d0-7f78c38c12b6-scripts\") pod \"ovn-northd-0\" (UID: \"dca5fd9f-3aaf-4fd7-96d0-7f78c38c12b6\") " pod="openstack/ovn-northd-0" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.242453 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dca5fd9f-3aaf-4fd7-96d0-7f78c38c12b6-scripts\") pod \"ovn-northd-0\" (UID: \"dca5fd9f-3aaf-4fd7-96d0-7f78c38c12b6\") " pod="openstack/ovn-northd-0" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.242527 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/31be02e0-3b12-49e9-a900-3d069fa35ac4-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-4dwc7\" (UID: \"31be02e0-3b12-49e9-a900-3d069fa35ac4\") " pod="openstack/dnsmasq-dns-8554648995-4dwc7" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.242677 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/31be02e0-3b12-49e9-a900-3d069fa35ac4-dns-svc\") pod \"dnsmasq-dns-8554648995-4dwc7\" (UID: \"31be02e0-3b12-49e9-a900-3d069fa35ac4\") " pod="openstack/dnsmasq-dns-8554648995-4dwc7" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.243119 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/dca5fd9f-3aaf-4fd7-96d0-7f78c38c12b6-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"dca5fd9f-3aaf-4fd7-96d0-7f78c38c12b6\") " pod="openstack/ovn-northd-0" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.243191 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/31be02e0-3b12-49e9-a900-3d069fa35ac4-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-4dwc7\" (UID: \"31be02e0-3b12-49e9-a900-3d069fa35ac4\") " pod="openstack/dnsmasq-dns-8554648995-4dwc7" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.244025 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31be02e0-3b12-49e9-a900-3d069fa35ac4-config\") pod \"dnsmasq-dns-8554648995-4dwc7\" (UID: \"31be02e0-3b12-49e9-a900-3d069fa35ac4\") " pod="openstack/dnsmasq-dns-8554648995-4dwc7" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.245571 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dca5fd9f-3aaf-4fd7-96d0-7f78c38c12b6-config\") pod \"ovn-northd-0\" (UID: \"dca5fd9f-3aaf-4fd7-96d0-7f78c38c12b6\") " pod="openstack/ovn-northd-0" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.248235 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/dca5fd9f-3aaf-4fd7-96d0-7f78c38c12b6-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"dca5fd9f-3aaf-4fd7-96d0-7f78c38c12b6\") " pod="openstack/ovn-northd-0" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.249703 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dca5fd9f-3aaf-4fd7-96d0-7f78c38c12b6-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"dca5fd9f-3aaf-4fd7-96d0-7f78c38c12b6\") " pod="openstack/ovn-northd-0" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.250739 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/dca5fd9f-3aaf-4fd7-96d0-7f78c38c12b6-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"dca5fd9f-3aaf-4fd7-96d0-7f78c38c12b6\") " pod="openstack/ovn-northd-0" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.269902 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtcj2\" (UniqueName: \"kubernetes.io/projected/dca5fd9f-3aaf-4fd7-96d0-7f78c38c12b6-kube-api-access-rtcj2\") pod \"ovn-northd-0\" (UID: \"dca5fd9f-3aaf-4fd7-96d0-7f78c38c12b6\") " pod="openstack/ovn-northd-0" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.271294 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkcxt\" (UniqueName: \"kubernetes.io/projected/31be02e0-3b12-49e9-a900-3d069fa35ac4-kube-api-access-kkcxt\") pod \"dnsmasq-dns-8554648995-4dwc7\" (UID: \"31be02e0-3b12-49e9-a900-3d069fa35ac4\") " pod="openstack/dnsmasq-dns-8554648995-4dwc7" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.361056 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-7sljg" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.364044 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.481248 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-4dwc7" Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.582327 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-6mzpl"] Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.588020 4802 generic.go:334] "Generic (PLEG): container finished" podID="e78545cd-e888-4838-9b9b-6008a2c60a40" containerID="b22eb47d20f2d570122c953f696638fbd017a257f6297886f48695e8918b76d4" exitCode=0 Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.588071 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-wtbxc" event={"ID":"e78545cd-e888-4838-9b9b-6008a2c60a40","Type":"ContainerDied","Data":"b22eb47d20f2d570122c953f696638fbd017a257f6297886f48695e8918b76d4"} Dec 06 03:57:22 crc kubenswrapper[4802]: W1206 03:57:22.590076 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod742a77a2_cc74_4671_9247_18f27582c980.slice/crio-32a5d5393050330e787068373b72d70ac85fe7e4147e7e8fa6ecbd69696f49d6 WatchSource:0}: Error finding container 32a5d5393050330e787068373b72d70ac85fe7e4147e7e8fa6ecbd69696f49d6: Status 404 returned error can't find the container with id 32a5d5393050330e787068373b72d70ac85fe7e4147e7e8fa6ecbd69696f49d6 Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.605106 4802 generic.go:334] "Generic (PLEG): container finished" podID="68de0f89-2871-4ae7-aa90-e002255f92b8" containerID="144afcadfc65fc6ec52957a3771d24631229e58af2b3c642ef547a37f9241be3" exitCode=0 Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.605700 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-c5jjr" event={"ID":"68de0f89-2871-4ae7-aa90-e002255f92b8","Type":"ContainerDied","Data":"144afcadfc65fc6ec52957a3771d24631229e58af2b3c642ef547a37f9241be3"} Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.804015 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-7sljg"] Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.894465 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 06 03:57:22 crc kubenswrapper[4802]: W1206 03:57:22.921225 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddca5fd9f_3aaf_4fd7_96d0_7f78c38c12b6.slice/crio-d6a52d8eae259eb5ab18b756734ca110493078f4c4ddab2535e62b1806eaf3e0 WatchSource:0}: Error finding container d6a52d8eae259eb5ab18b756734ca110493078f4c4ddab2535e62b1806eaf3e0: Status 404 returned error can't find the container with id d6a52d8eae259eb5ab18b756734ca110493078f4c4ddab2535e62b1806eaf3e0 Dec 06 03:57:22 crc kubenswrapper[4802]: I1206 03:57:22.921940 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-c5jjr" Dec 06 03:57:23 crc kubenswrapper[4802]: I1206 03:57:23.037822 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-4dwc7"] Dec 06 03:57:23 crc kubenswrapper[4802]: I1206 03:57:23.041095 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-wtbxc" Dec 06 03:57:23 crc kubenswrapper[4802]: W1206 03:57:23.043983 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod31be02e0_3b12_49e9_a900_3d069fa35ac4.slice/crio-4c6a8faef9e7ceb14acd83bdd99e061a85f513446217176326f96a79e5e01f0d WatchSource:0}: Error finding container 4c6a8faef9e7ceb14acd83bdd99e061a85f513446217176326f96a79e5e01f0d: Status 404 returned error can't find the container with id 4c6a8faef9e7ceb14acd83bdd99e061a85f513446217176326f96a79e5e01f0d Dec 06 03:57:23 crc kubenswrapper[4802]: I1206 03:57:23.053618 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68de0f89-2871-4ae7-aa90-e002255f92b8-config\") pod \"68de0f89-2871-4ae7-aa90-e002255f92b8\" (UID: \"68de0f89-2871-4ae7-aa90-e002255f92b8\") " Dec 06 03:57:23 crc kubenswrapper[4802]: I1206 03:57:23.053688 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68de0f89-2871-4ae7-aa90-e002255f92b8-dns-svc\") pod \"68de0f89-2871-4ae7-aa90-e002255f92b8\" (UID: \"68de0f89-2871-4ae7-aa90-e002255f92b8\") " Dec 06 03:57:23 crc kubenswrapper[4802]: I1206 03:57:23.053830 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-756r8\" (UniqueName: \"kubernetes.io/projected/68de0f89-2871-4ae7-aa90-e002255f92b8-kube-api-access-756r8\") pod \"68de0f89-2871-4ae7-aa90-e002255f92b8\" (UID: \"68de0f89-2871-4ae7-aa90-e002255f92b8\") " Dec 06 03:57:23 crc kubenswrapper[4802]: I1206 03:57:23.061843 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68de0f89-2871-4ae7-aa90-e002255f92b8-kube-api-access-756r8" (OuterVolumeSpecName: "kube-api-access-756r8") pod "68de0f89-2871-4ae7-aa90-e002255f92b8" (UID: "68de0f89-2871-4ae7-aa90-e002255f92b8"). InnerVolumeSpecName "kube-api-access-756r8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:57:23 crc kubenswrapper[4802]: I1206 03:57:23.074105 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68de0f89-2871-4ae7-aa90-e002255f92b8-config" (OuterVolumeSpecName: "config") pod "68de0f89-2871-4ae7-aa90-e002255f92b8" (UID: "68de0f89-2871-4ae7-aa90-e002255f92b8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:57:23 crc kubenswrapper[4802]: I1206 03:57:23.076701 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68de0f89-2871-4ae7-aa90-e002255f92b8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "68de0f89-2871-4ae7-aa90-e002255f92b8" (UID: "68de0f89-2871-4ae7-aa90-e002255f92b8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:57:23 crc kubenswrapper[4802]: I1206 03:57:23.156478 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e78545cd-e888-4838-9b9b-6008a2c60a40-dns-svc\") pod \"e78545cd-e888-4838-9b9b-6008a2c60a40\" (UID: \"e78545cd-e888-4838-9b9b-6008a2c60a40\") " Dec 06 03:57:23 crc kubenswrapper[4802]: I1206 03:57:23.156549 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m6f5m\" (UniqueName: \"kubernetes.io/projected/e78545cd-e888-4838-9b9b-6008a2c60a40-kube-api-access-m6f5m\") pod \"e78545cd-e888-4838-9b9b-6008a2c60a40\" (UID: \"e78545cd-e888-4838-9b9b-6008a2c60a40\") " Dec 06 03:57:23 crc kubenswrapper[4802]: I1206 03:57:23.156625 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e78545cd-e888-4838-9b9b-6008a2c60a40-config\") pod \"e78545cd-e888-4838-9b9b-6008a2c60a40\" (UID: \"e78545cd-e888-4838-9b9b-6008a2c60a40\") " Dec 06 03:57:23 crc kubenswrapper[4802]: I1206 03:57:23.157850 4802 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68de0f89-2871-4ae7-aa90-e002255f92b8-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:23 crc kubenswrapper[4802]: I1206 03:57:23.157871 4802 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68de0f89-2871-4ae7-aa90-e002255f92b8-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:23 crc kubenswrapper[4802]: I1206 03:57:23.157880 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-756r8\" (UniqueName: \"kubernetes.io/projected/68de0f89-2871-4ae7-aa90-e002255f92b8-kube-api-access-756r8\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:23 crc kubenswrapper[4802]: I1206 03:57:23.161644 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e78545cd-e888-4838-9b9b-6008a2c60a40-kube-api-access-m6f5m" (OuterVolumeSpecName: "kube-api-access-m6f5m") pod "e78545cd-e888-4838-9b9b-6008a2c60a40" (UID: "e78545cd-e888-4838-9b9b-6008a2c60a40"). InnerVolumeSpecName "kube-api-access-m6f5m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:57:23 crc kubenswrapper[4802]: I1206 03:57:23.184321 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e78545cd-e888-4838-9b9b-6008a2c60a40-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e78545cd-e888-4838-9b9b-6008a2c60a40" (UID: "e78545cd-e888-4838-9b9b-6008a2c60a40"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:57:23 crc kubenswrapper[4802]: I1206 03:57:23.184568 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e78545cd-e888-4838-9b9b-6008a2c60a40-config" (OuterVolumeSpecName: "config") pod "e78545cd-e888-4838-9b9b-6008a2c60a40" (UID: "e78545cd-e888-4838-9b9b-6008a2c60a40"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:57:23 crc kubenswrapper[4802]: I1206 03:57:23.259898 4802 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e78545cd-e888-4838-9b9b-6008a2c60a40-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:23 crc kubenswrapper[4802]: I1206 03:57:23.259942 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m6f5m\" (UniqueName: \"kubernetes.io/projected/e78545cd-e888-4838-9b9b-6008a2c60a40-kube-api-access-m6f5m\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:23 crc kubenswrapper[4802]: I1206 03:57:23.259957 4802 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e78545cd-e888-4838-9b9b-6008a2c60a40-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:23 crc kubenswrapper[4802]: I1206 03:57:23.622651 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-wtbxc" Dec 06 03:57:23 crc kubenswrapper[4802]: I1206 03:57:23.622679 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-wtbxc" event={"ID":"e78545cd-e888-4838-9b9b-6008a2c60a40","Type":"ContainerDied","Data":"144186cebdc968c870d90cf997757a5dbce30f6d108b3c222507de2edbeb34cf"} Dec 06 03:57:23 crc kubenswrapper[4802]: I1206 03:57:23.622741 4802 scope.go:117] "RemoveContainer" containerID="b22eb47d20f2d570122c953f696638fbd017a257f6297886f48695e8918b76d4" Dec 06 03:57:23 crc kubenswrapper[4802]: I1206 03:57:23.625140 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-7sljg" event={"ID":"11e2072e-0127-48c0-b417-ced9cfa9bf64","Type":"ContainerStarted","Data":"431318900d34a9ac1174cea734af0db8c4c4b57b681d0db8913af9456acc691e"} Dec 06 03:57:23 crc kubenswrapper[4802]: I1206 03:57:23.625178 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-7sljg" event={"ID":"11e2072e-0127-48c0-b417-ced9cfa9bf64","Type":"ContainerStarted","Data":"4886dd6a28b84410839f926d85bb8df3c62902743d54f078f3da9909d4cd6a69"} Dec 06 03:57:23 crc kubenswrapper[4802]: I1206 03:57:23.627796 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-c5jjr" event={"ID":"68de0f89-2871-4ae7-aa90-e002255f92b8","Type":"ContainerDied","Data":"7db130bbc8e83573ea51f3edf21192810e30f562805639dd8b72344a90c1ca41"} Dec 06 03:57:23 crc kubenswrapper[4802]: I1206 03:57:23.627878 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-c5jjr" Dec 06 03:57:23 crc kubenswrapper[4802]: I1206 03:57:23.630192 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"dca5fd9f-3aaf-4fd7-96d0-7f78c38c12b6","Type":"ContainerStarted","Data":"d6a52d8eae259eb5ab18b756734ca110493078f4c4ddab2535e62b1806eaf3e0"} Dec 06 03:57:23 crc kubenswrapper[4802]: I1206 03:57:23.632391 4802 generic.go:334] "Generic (PLEG): container finished" podID="742a77a2-cc74-4671-9247-18f27582c980" containerID="4ee6be7e3ab0768e135dca75dd0f7881bb7f564368ab8119ed82ba40babfebd2" exitCode=0 Dec 06 03:57:23 crc kubenswrapper[4802]: I1206 03:57:23.632503 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-6mzpl" event={"ID":"742a77a2-cc74-4671-9247-18f27582c980","Type":"ContainerDied","Data":"4ee6be7e3ab0768e135dca75dd0f7881bb7f564368ab8119ed82ba40babfebd2"} Dec 06 03:57:23 crc kubenswrapper[4802]: I1206 03:57:23.632536 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-6mzpl" event={"ID":"742a77a2-cc74-4671-9247-18f27582c980","Type":"ContainerStarted","Data":"32a5d5393050330e787068373b72d70ac85fe7e4147e7e8fa6ecbd69696f49d6"} Dec 06 03:57:23 crc kubenswrapper[4802]: I1206 03:57:23.637050 4802 generic.go:334] "Generic (PLEG): container finished" podID="31be02e0-3b12-49e9-a900-3d069fa35ac4" containerID="58e6ce4ec8424e87b10a3e1c0ccf3f5b0c841d67d4f4f675da676460244a9096" exitCode=0 Dec 06 03:57:23 crc kubenswrapper[4802]: I1206 03:57:23.637085 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-4dwc7" event={"ID":"31be02e0-3b12-49e9-a900-3d069fa35ac4","Type":"ContainerDied","Data":"58e6ce4ec8424e87b10a3e1c0ccf3f5b0c841d67d4f4f675da676460244a9096"} Dec 06 03:57:23 crc kubenswrapper[4802]: I1206 03:57:23.637562 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-4dwc7" event={"ID":"31be02e0-3b12-49e9-a900-3d069fa35ac4","Type":"ContainerStarted","Data":"4c6a8faef9e7ceb14acd83bdd99e061a85f513446217176326f96a79e5e01f0d"} Dec 06 03:57:23 crc kubenswrapper[4802]: I1206 03:57:23.660403 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-7sljg" podStartSLOduration=2.660384382 podStartE2EDuration="2.660384382s" podCreationTimestamp="2025-12-06 03:57:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:57:23.648876367 +0000 UTC m=+1036.520785529" watchObservedRunningTime="2025-12-06 03:57:23.660384382 +0000 UTC m=+1036.532293534" Dec 06 03:57:23 crc kubenswrapper[4802]: I1206 03:57:23.667696 4802 scope.go:117] "RemoveContainer" containerID="144afcadfc65fc6ec52957a3771d24631229e58af2b3c642ef547a37f9241be3" Dec 06 03:57:23 crc kubenswrapper[4802]: I1206 03:57:23.758099 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-wtbxc"] Dec 06 03:57:23 crc kubenswrapper[4802]: I1206 03:57:23.766916 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-wtbxc"] Dec 06 03:57:23 crc kubenswrapper[4802]: I1206 03:57:23.806262 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-c5jjr"] Dec 06 03:57:23 crc kubenswrapper[4802]: I1206 03:57:23.816822 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-c5jjr"] Dec 06 03:57:24 crc kubenswrapper[4802]: I1206 03:57:24.144229 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 06 03:57:24 crc kubenswrapper[4802]: I1206 03:57:24.144607 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 06 03:57:24 crc kubenswrapper[4802]: I1206 03:57:24.217541 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 06 03:57:24 crc kubenswrapper[4802]: I1206 03:57:24.646750 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"dca5fd9f-3aaf-4fd7-96d0-7f78c38c12b6","Type":"ContainerStarted","Data":"17e17aaa0a5127c817b551520e2b13d5595ea8accb89ed5567ccf7e71d629933"} Dec 06 03:57:24 crc kubenswrapper[4802]: I1206 03:57:24.646816 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"dca5fd9f-3aaf-4fd7-96d0-7f78c38c12b6","Type":"ContainerStarted","Data":"da38bd4ca8ab9894102ce03a27fe0a3496315573ae010fcbc4a81ff12d75a79a"} Dec 06 03:57:24 crc kubenswrapper[4802]: I1206 03:57:24.646841 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 06 03:57:24 crc kubenswrapper[4802]: I1206 03:57:24.650355 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-6mzpl" event={"ID":"742a77a2-cc74-4671-9247-18f27582c980","Type":"ContainerStarted","Data":"7d9e7538763552a56a625a1a2a0cd63ff1b289d4b121bb60255ccc8733bcca4d"} Dec 06 03:57:24 crc kubenswrapper[4802]: I1206 03:57:24.651213 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6bc7876d45-6mzpl" Dec 06 03:57:24 crc kubenswrapper[4802]: I1206 03:57:24.652307 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-4dwc7" event={"ID":"31be02e0-3b12-49e9-a900-3d069fa35ac4","Type":"ContainerStarted","Data":"53eb2ba35507f38304fc35d966b3530345562545e5fa5c41844dd46be2d6f21d"} Dec 06 03:57:24 crc kubenswrapper[4802]: I1206 03:57:24.652466 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8554648995-4dwc7" Dec 06 03:57:24 crc kubenswrapper[4802]: I1206 03:57:24.669804 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.408038278 podStartE2EDuration="3.669782255s" podCreationTimestamp="2025-12-06 03:57:21 +0000 UTC" firstStartedPulling="2025-12-06 03:57:22.924656048 +0000 UTC m=+1035.796565200" lastFinishedPulling="2025-12-06 03:57:24.186400015 +0000 UTC m=+1037.058309177" observedRunningTime="2025-12-06 03:57:24.666217896 +0000 UTC m=+1037.538127058" watchObservedRunningTime="2025-12-06 03:57:24.669782255 +0000 UTC m=+1037.541691407" Dec 06 03:57:24 crc kubenswrapper[4802]: I1206 03:57:24.703718 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8554648995-4dwc7" podStartSLOduration=3.7036975610000002 podStartE2EDuration="3.703697561s" podCreationTimestamp="2025-12-06 03:57:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:57:24.698135179 +0000 UTC m=+1037.570044341" watchObservedRunningTime="2025-12-06 03:57:24.703697561 +0000 UTC m=+1037.575606713" Dec 06 03:57:24 crc kubenswrapper[4802]: I1206 03:57:24.720169 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6bc7876d45-6mzpl" podStartSLOduration=3.720145251 podStartE2EDuration="3.720145251s" podCreationTimestamp="2025-12-06 03:57:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:57:24.713875989 +0000 UTC m=+1037.585785141" watchObservedRunningTime="2025-12-06 03:57:24.720145251 +0000 UTC m=+1037.592054413" Dec 06 03:57:24 crc kubenswrapper[4802]: I1206 03:57:24.733394 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 06 03:57:25 crc kubenswrapper[4802]: I1206 03:57:25.191875 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-e5f1-account-create-update-zxdsz"] Dec 06 03:57:25 crc kubenswrapper[4802]: E1206 03:57:25.192243 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e78545cd-e888-4838-9b9b-6008a2c60a40" containerName="init" Dec 06 03:57:25 crc kubenswrapper[4802]: I1206 03:57:25.192263 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="e78545cd-e888-4838-9b9b-6008a2c60a40" containerName="init" Dec 06 03:57:25 crc kubenswrapper[4802]: E1206 03:57:25.192301 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68de0f89-2871-4ae7-aa90-e002255f92b8" containerName="init" Dec 06 03:57:25 crc kubenswrapper[4802]: I1206 03:57:25.192309 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="68de0f89-2871-4ae7-aa90-e002255f92b8" containerName="init" Dec 06 03:57:25 crc kubenswrapper[4802]: I1206 03:57:25.192634 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="e78545cd-e888-4838-9b9b-6008a2c60a40" containerName="init" Dec 06 03:57:25 crc kubenswrapper[4802]: I1206 03:57:25.192664 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="68de0f89-2871-4ae7-aa90-e002255f92b8" containerName="init" Dec 06 03:57:25 crc kubenswrapper[4802]: I1206 03:57:25.193290 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-e5f1-account-create-update-zxdsz" Dec 06 03:57:25 crc kubenswrapper[4802]: I1206 03:57:25.196546 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 06 03:57:25 crc kubenswrapper[4802]: I1206 03:57:25.203815 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-e5f1-account-create-update-zxdsz"] Dec 06 03:57:25 crc kubenswrapper[4802]: I1206 03:57:25.232498 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-49mm8"] Dec 06 03:57:25 crc kubenswrapper[4802]: I1206 03:57:25.233661 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-49mm8" Dec 06 03:57:25 crc kubenswrapper[4802]: I1206 03:57:25.241381 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-49mm8"] Dec 06 03:57:25 crc kubenswrapper[4802]: I1206 03:57:25.308519 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58v7z\" (UniqueName: \"kubernetes.io/projected/fb7eaedf-a7ed-4c1f-8609-456afd84daa6-kube-api-access-58v7z\") pod \"keystone-e5f1-account-create-update-zxdsz\" (UID: \"fb7eaedf-a7ed-4c1f-8609-456afd84daa6\") " pod="openstack/keystone-e5f1-account-create-update-zxdsz" Dec 06 03:57:25 crc kubenswrapper[4802]: I1206 03:57:25.308577 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6w7pl\" (UniqueName: \"kubernetes.io/projected/19fbf090-979f-4d84-b774-c9cf98ea1501-kube-api-access-6w7pl\") pod \"keystone-db-create-49mm8\" (UID: \"19fbf090-979f-4d84-b774-c9cf98ea1501\") " pod="openstack/keystone-db-create-49mm8" Dec 06 03:57:25 crc kubenswrapper[4802]: I1206 03:57:25.308622 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/19fbf090-979f-4d84-b774-c9cf98ea1501-operator-scripts\") pod \"keystone-db-create-49mm8\" (UID: \"19fbf090-979f-4d84-b774-c9cf98ea1501\") " pod="openstack/keystone-db-create-49mm8" Dec 06 03:57:25 crc kubenswrapper[4802]: I1206 03:57:25.308833 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fb7eaedf-a7ed-4c1f-8609-456afd84daa6-operator-scripts\") pod \"keystone-e5f1-account-create-update-zxdsz\" (UID: \"fb7eaedf-a7ed-4c1f-8609-456afd84daa6\") " pod="openstack/keystone-e5f1-account-create-update-zxdsz" Dec 06 03:57:25 crc kubenswrapper[4802]: I1206 03:57:25.410318 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58v7z\" (UniqueName: \"kubernetes.io/projected/fb7eaedf-a7ed-4c1f-8609-456afd84daa6-kube-api-access-58v7z\") pod \"keystone-e5f1-account-create-update-zxdsz\" (UID: \"fb7eaedf-a7ed-4c1f-8609-456afd84daa6\") " pod="openstack/keystone-e5f1-account-create-update-zxdsz" Dec 06 03:57:25 crc kubenswrapper[4802]: I1206 03:57:25.410374 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6w7pl\" (UniqueName: \"kubernetes.io/projected/19fbf090-979f-4d84-b774-c9cf98ea1501-kube-api-access-6w7pl\") pod \"keystone-db-create-49mm8\" (UID: \"19fbf090-979f-4d84-b774-c9cf98ea1501\") " pod="openstack/keystone-db-create-49mm8" Dec 06 03:57:25 crc kubenswrapper[4802]: I1206 03:57:25.410427 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/19fbf090-979f-4d84-b774-c9cf98ea1501-operator-scripts\") pod \"keystone-db-create-49mm8\" (UID: \"19fbf090-979f-4d84-b774-c9cf98ea1501\") " pod="openstack/keystone-db-create-49mm8" Dec 06 03:57:25 crc kubenswrapper[4802]: I1206 03:57:25.410466 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fb7eaedf-a7ed-4c1f-8609-456afd84daa6-operator-scripts\") pod \"keystone-e5f1-account-create-update-zxdsz\" (UID: \"fb7eaedf-a7ed-4c1f-8609-456afd84daa6\") " pod="openstack/keystone-e5f1-account-create-update-zxdsz" Dec 06 03:57:25 crc kubenswrapper[4802]: I1206 03:57:25.411399 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/19fbf090-979f-4d84-b774-c9cf98ea1501-operator-scripts\") pod \"keystone-db-create-49mm8\" (UID: \"19fbf090-979f-4d84-b774-c9cf98ea1501\") " pod="openstack/keystone-db-create-49mm8" Dec 06 03:57:25 crc kubenswrapper[4802]: I1206 03:57:25.411459 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fb7eaedf-a7ed-4c1f-8609-456afd84daa6-operator-scripts\") pod \"keystone-e5f1-account-create-update-zxdsz\" (UID: \"fb7eaedf-a7ed-4c1f-8609-456afd84daa6\") " pod="openstack/keystone-e5f1-account-create-update-zxdsz" Dec 06 03:57:25 crc kubenswrapper[4802]: I1206 03:57:25.425791 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6w7pl\" (UniqueName: \"kubernetes.io/projected/19fbf090-979f-4d84-b774-c9cf98ea1501-kube-api-access-6w7pl\") pod \"keystone-db-create-49mm8\" (UID: \"19fbf090-979f-4d84-b774-c9cf98ea1501\") " pod="openstack/keystone-db-create-49mm8" Dec 06 03:57:25 crc kubenswrapper[4802]: I1206 03:57:25.426685 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58v7z\" (UniqueName: \"kubernetes.io/projected/fb7eaedf-a7ed-4c1f-8609-456afd84daa6-kube-api-access-58v7z\") pod \"keystone-e5f1-account-create-update-zxdsz\" (UID: \"fb7eaedf-a7ed-4c1f-8609-456afd84daa6\") " pod="openstack/keystone-e5f1-account-create-update-zxdsz" Dec 06 03:57:25 crc kubenswrapper[4802]: I1206 03:57:25.460508 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68de0f89-2871-4ae7-aa90-e002255f92b8" path="/var/lib/kubelet/pods/68de0f89-2871-4ae7-aa90-e002255f92b8/volumes" Dec 06 03:57:25 crc kubenswrapper[4802]: I1206 03:57:25.461213 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e78545cd-e888-4838-9b9b-6008a2c60a40" path="/var/lib/kubelet/pods/e78545cd-e888-4838-9b9b-6008a2c60a40/volumes" Dec 06 03:57:25 crc kubenswrapper[4802]: I1206 03:57:25.472225 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-l2tpj"] Dec 06 03:57:25 crc kubenswrapper[4802]: I1206 03:57:25.473342 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-l2tpj" Dec 06 03:57:25 crc kubenswrapper[4802]: I1206 03:57:25.509187 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-l2tpj"] Dec 06 03:57:25 crc kubenswrapper[4802]: I1206 03:57:25.512909 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-e5f1-account-create-update-zxdsz" Dec 06 03:57:25 crc kubenswrapper[4802]: I1206 03:57:25.519677 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 06 03:57:25 crc kubenswrapper[4802]: I1206 03:57:25.520502 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 06 03:57:25 crc kubenswrapper[4802]: I1206 03:57:25.558293 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-49mm8" Dec 06 03:57:25 crc kubenswrapper[4802]: I1206 03:57:25.611962 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-785f-account-create-update-4x7n2"] Dec 06 03:57:25 crc kubenswrapper[4802]: I1206 03:57:25.613313 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-785f-account-create-update-4x7n2" Dec 06 03:57:25 crc kubenswrapper[4802]: I1206 03:57:25.651880 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 06 03:57:25 crc kubenswrapper[4802]: I1206 03:57:25.652339 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phxnv\" (UniqueName: \"kubernetes.io/projected/76accf7a-8aeb-48a1-88c5-3b64e8a0d25e-kube-api-access-phxnv\") pod \"placement-db-create-l2tpj\" (UID: \"76accf7a-8aeb-48a1-88c5-3b64e8a0d25e\") " pod="openstack/placement-db-create-l2tpj" Dec 06 03:57:25 crc kubenswrapper[4802]: I1206 03:57:25.652374 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/76accf7a-8aeb-48a1-88c5-3b64e8a0d25e-operator-scripts\") pod \"placement-db-create-l2tpj\" (UID: \"76accf7a-8aeb-48a1-88c5-3b64e8a0d25e\") " pod="openstack/placement-db-create-l2tpj" Dec 06 03:57:25 crc kubenswrapper[4802]: I1206 03:57:25.652401 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 06 03:57:25 crc kubenswrapper[4802]: I1206 03:57:25.659222 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-785f-account-create-update-4x7n2"] Dec 06 03:57:25 crc kubenswrapper[4802]: I1206 03:57:25.660240 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 06 03:57:25 crc kubenswrapper[4802]: I1206 03:57:25.754711 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8flqx\" (UniqueName: \"kubernetes.io/projected/b4efbdc9-de14-4774-9c69-66b4cb0f79b9-kube-api-access-8flqx\") pod \"placement-785f-account-create-update-4x7n2\" (UID: \"b4efbdc9-de14-4774-9c69-66b4cb0f79b9\") " pod="openstack/placement-785f-account-create-update-4x7n2" Dec 06 03:57:25 crc kubenswrapper[4802]: I1206 03:57:25.755002 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phxnv\" (UniqueName: \"kubernetes.io/projected/76accf7a-8aeb-48a1-88c5-3b64e8a0d25e-kube-api-access-phxnv\") pod \"placement-db-create-l2tpj\" (UID: \"76accf7a-8aeb-48a1-88c5-3b64e8a0d25e\") " pod="openstack/placement-db-create-l2tpj" Dec 06 03:57:25 crc kubenswrapper[4802]: I1206 03:57:25.755031 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b4efbdc9-de14-4774-9c69-66b4cb0f79b9-operator-scripts\") pod \"placement-785f-account-create-update-4x7n2\" (UID: \"b4efbdc9-de14-4774-9c69-66b4cb0f79b9\") " pod="openstack/placement-785f-account-create-update-4x7n2" Dec 06 03:57:25 crc kubenswrapper[4802]: I1206 03:57:25.755064 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/76accf7a-8aeb-48a1-88c5-3b64e8a0d25e-operator-scripts\") pod \"placement-db-create-l2tpj\" (UID: \"76accf7a-8aeb-48a1-88c5-3b64e8a0d25e\") " pod="openstack/placement-db-create-l2tpj" Dec 06 03:57:25 crc kubenswrapper[4802]: I1206 03:57:25.761028 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/76accf7a-8aeb-48a1-88c5-3b64e8a0d25e-operator-scripts\") pod \"placement-db-create-l2tpj\" (UID: \"76accf7a-8aeb-48a1-88c5-3b64e8a0d25e\") " pod="openstack/placement-db-create-l2tpj" Dec 06 03:57:25 crc kubenswrapper[4802]: I1206 03:57:25.805376 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phxnv\" (UniqueName: \"kubernetes.io/projected/76accf7a-8aeb-48a1-88c5-3b64e8a0d25e-kube-api-access-phxnv\") pod \"placement-db-create-l2tpj\" (UID: \"76accf7a-8aeb-48a1-88c5-3b64e8a0d25e\") " pod="openstack/placement-db-create-l2tpj" Dec 06 03:57:25 crc kubenswrapper[4802]: I1206 03:57:25.810883 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-l2tpj" Dec 06 03:57:25 crc kubenswrapper[4802]: I1206 03:57:25.859152 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8flqx\" (UniqueName: \"kubernetes.io/projected/b4efbdc9-de14-4774-9c69-66b4cb0f79b9-kube-api-access-8flqx\") pod \"placement-785f-account-create-update-4x7n2\" (UID: \"b4efbdc9-de14-4774-9c69-66b4cb0f79b9\") " pod="openstack/placement-785f-account-create-update-4x7n2" Dec 06 03:57:25 crc kubenswrapper[4802]: I1206 03:57:25.859210 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b4efbdc9-de14-4774-9c69-66b4cb0f79b9-operator-scripts\") pod \"placement-785f-account-create-update-4x7n2\" (UID: \"b4efbdc9-de14-4774-9c69-66b4cb0f79b9\") " pod="openstack/placement-785f-account-create-update-4x7n2" Dec 06 03:57:25 crc kubenswrapper[4802]: I1206 03:57:25.859864 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b4efbdc9-de14-4774-9c69-66b4cb0f79b9-operator-scripts\") pod \"placement-785f-account-create-update-4x7n2\" (UID: \"b4efbdc9-de14-4774-9c69-66b4cb0f79b9\") " pod="openstack/placement-785f-account-create-update-4x7n2" Dec 06 03:57:25 crc kubenswrapper[4802]: I1206 03:57:25.894584 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8flqx\" (UniqueName: \"kubernetes.io/projected/b4efbdc9-de14-4774-9c69-66b4cb0f79b9-kube-api-access-8flqx\") pod \"placement-785f-account-create-update-4x7n2\" (UID: \"b4efbdc9-de14-4774-9c69-66b4cb0f79b9\") " pod="openstack/placement-785f-account-create-update-4x7n2" Dec 06 03:57:25 crc kubenswrapper[4802]: I1206 03:57:25.939369 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-ckg82"] Dec 06 03:57:25 crc kubenswrapper[4802]: I1206 03:57:25.940468 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-ckg82" Dec 06 03:57:25 crc kubenswrapper[4802]: I1206 03:57:25.945183 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-ckg82"] Dec 06 03:57:25 crc kubenswrapper[4802]: I1206 03:57:25.971891 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-785f-account-create-update-4x7n2" Dec 06 03:57:26 crc kubenswrapper[4802]: I1206 03:57:26.022687 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 06 03:57:26 crc kubenswrapper[4802]: I1206 03:57:26.034087 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-633c-account-create-update-ld6ht"] Dec 06 03:57:26 crc kubenswrapper[4802]: I1206 03:57:26.035075 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-633c-account-create-update-ld6ht" Dec 06 03:57:26 crc kubenswrapper[4802]: I1206 03:57:26.042462 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-633c-account-create-update-ld6ht"] Dec 06 03:57:26 crc kubenswrapper[4802]: I1206 03:57:26.047220 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 06 03:57:26 crc kubenswrapper[4802]: I1206 03:57:26.070306 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ssgkv\" (UniqueName: \"kubernetes.io/projected/a2ab2ae8-12da-43f3-bcd8-1de34185d285-kube-api-access-ssgkv\") pod \"glance-db-create-ckg82\" (UID: \"a2ab2ae8-12da-43f3-bcd8-1de34185d285\") " pod="openstack/glance-db-create-ckg82" Dec 06 03:57:26 crc kubenswrapper[4802]: I1206 03:57:26.070366 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a2ab2ae8-12da-43f3-bcd8-1de34185d285-operator-scripts\") pod \"glance-db-create-ckg82\" (UID: \"a2ab2ae8-12da-43f3-bcd8-1de34185d285\") " pod="openstack/glance-db-create-ckg82" Dec 06 03:57:26 crc kubenswrapper[4802]: I1206 03:57:26.156093 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-e5f1-account-create-update-zxdsz"] Dec 06 03:57:26 crc kubenswrapper[4802]: W1206 03:57:26.161058 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfb7eaedf_a7ed_4c1f_8609_456afd84daa6.slice/crio-62eddb98424ae1cd38cccca3d04e8251ba4d4dd758c8365d1ea46bed4f588fd8 WatchSource:0}: Error finding container 62eddb98424ae1cd38cccca3d04e8251ba4d4dd758c8365d1ea46bed4f588fd8: Status 404 returned error can't find the container with id 62eddb98424ae1cd38cccca3d04e8251ba4d4dd758c8365d1ea46bed4f588fd8 Dec 06 03:57:26 crc kubenswrapper[4802]: I1206 03:57:26.181593 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdcqq\" (UniqueName: \"kubernetes.io/projected/7a4a22d0-ccb1-40ec-9872-05b41d788fd4-kube-api-access-qdcqq\") pod \"glance-633c-account-create-update-ld6ht\" (UID: \"7a4a22d0-ccb1-40ec-9872-05b41d788fd4\") " pod="openstack/glance-633c-account-create-update-ld6ht" Dec 06 03:57:26 crc kubenswrapper[4802]: I1206 03:57:26.181663 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7a4a22d0-ccb1-40ec-9872-05b41d788fd4-operator-scripts\") pod \"glance-633c-account-create-update-ld6ht\" (UID: \"7a4a22d0-ccb1-40ec-9872-05b41d788fd4\") " pod="openstack/glance-633c-account-create-update-ld6ht" Dec 06 03:57:26 crc kubenswrapper[4802]: I1206 03:57:26.182013 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ssgkv\" (UniqueName: \"kubernetes.io/projected/a2ab2ae8-12da-43f3-bcd8-1de34185d285-kube-api-access-ssgkv\") pod \"glance-db-create-ckg82\" (UID: \"a2ab2ae8-12da-43f3-bcd8-1de34185d285\") " pod="openstack/glance-db-create-ckg82" Dec 06 03:57:26 crc kubenswrapper[4802]: I1206 03:57:26.182051 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a2ab2ae8-12da-43f3-bcd8-1de34185d285-operator-scripts\") pod \"glance-db-create-ckg82\" (UID: \"a2ab2ae8-12da-43f3-bcd8-1de34185d285\") " pod="openstack/glance-db-create-ckg82" Dec 06 03:57:26 crc kubenswrapper[4802]: I1206 03:57:26.183053 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a2ab2ae8-12da-43f3-bcd8-1de34185d285-operator-scripts\") pod \"glance-db-create-ckg82\" (UID: \"a2ab2ae8-12da-43f3-bcd8-1de34185d285\") " pod="openstack/glance-db-create-ckg82" Dec 06 03:57:26 crc kubenswrapper[4802]: I1206 03:57:26.199042 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ssgkv\" (UniqueName: \"kubernetes.io/projected/a2ab2ae8-12da-43f3-bcd8-1de34185d285-kube-api-access-ssgkv\") pod \"glance-db-create-ckg82\" (UID: \"a2ab2ae8-12da-43f3-bcd8-1de34185d285\") " pod="openstack/glance-db-create-ckg82" Dec 06 03:57:26 crc kubenswrapper[4802]: I1206 03:57:26.283330 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdcqq\" (UniqueName: \"kubernetes.io/projected/7a4a22d0-ccb1-40ec-9872-05b41d788fd4-kube-api-access-qdcqq\") pod \"glance-633c-account-create-update-ld6ht\" (UID: \"7a4a22d0-ccb1-40ec-9872-05b41d788fd4\") " pod="openstack/glance-633c-account-create-update-ld6ht" Dec 06 03:57:26 crc kubenswrapper[4802]: I1206 03:57:26.283383 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7a4a22d0-ccb1-40ec-9872-05b41d788fd4-operator-scripts\") pod \"glance-633c-account-create-update-ld6ht\" (UID: \"7a4a22d0-ccb1-40ec-9872-05b41d788fd4\") " pod="openstack/glance-633c-account-create-update-ld6ht" Dec 06 03:57:26 crc kubenswrapper[4802]: I1206 03:57:26.284217 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7a4a22d0-ccb1-40ec-9872-05b41d788fd4-operator-scripts\") pod \"glance-633c-account-create-update-ld6ht\" (UID: \"7a4a22d0-ccb1-40ec-9872-05b41d788fd4\") " pod="openstack/glance-633c-account-create-update-ld6ht" Dec 06 03:57:26 crc kubenswrapper[4802]: I1206 03:57:26.306246 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdcqq\" (UniqueName: \"kubernetes.io/projected/7a4a22d0-ccb1-40ec-9872-05b41d788fd4-kube-api-access-qdcqq\") pod \"glance-633c-account-create-update-ld6ht\" (UID: \"7a4a22d0-ccb1-40ec-9872-05b41d788fd4\") " pod="openstack/glance-633c-account-create-update-ld6ht" Dec 06 03:57:26 crc kubenswrapper[4802]: I1206 03:57:26.356684 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-49mm8"] Dec 06 03:57:26 crc kubenswrapper[4802]: W1206 03:57:26.364958 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod19fbf090_979f_4d84_b774_c9cf98ea1501.slice/crio-df9edf97881a0983b315e67b60ab2f5b48014cf579be17d6f4393eb6fa0f1f7b WatchSource:0}: Error finding container df9edf97881a0983b315e67b60ab2f5b48014cf579be17d6f4393eb6fa0f1f7b: Status 404 returned error can't find the container with id df9edf97881a0983b315e67b60ab2f5b48014cf579be17d6f4393eb6fa0f1f7b Dec 06 03:57:26 crc kubenswrapper[4802]: I1206 03:57:26.388835 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-ckg82" Dec 06 03:57:26 crc kubenswrapper[4802]: I1206 03:57:26.402237 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-633c-account-create-update-ld6ht" Dec 06 03:57:26 crc kubenswrapper[4802]: I1206 03:57:26.489414 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-l2tpj"] Dec 06 03:57:26 crc kubenswrapper[4802]: I1206 03:57:26.506549 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-785f-account-create-update-4x7n2"] Dec 06 03:57:26 crc kubenswrapper[4802]: W1206 03:57:26.513535 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod76accf7a_8aeb_48a1_88c5_3b64e8a0d25e.slice/crio-c07816566bd8038b13d768ab8a6e9508a4ffa8e8b72173563ef7217967b27527 WatchSource:0}: Error finding container c07816566bd8038b13d768ab8a6e9508a4ffa8e8b72173563ef7217967b27527: Status 404 returned error can't find the container with id c07816566bd8038b13d768ab8a6e9508a4ffa8e8b72173563ef7217967b27527 Dec 06 03:57:26 crc kubenswrapper[4802]: I1206 03:57:26.680373 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-l2tpj" event={"ID":"76accf7a-8aeb-48a1-88c5-3b64e8a0d25e","Type":"ContainerStarted","Data":"c07816566bd8038b13d768ab8a6e9508a4ffa8e8b72173563ef7217967b27527"} Dec 06 03:57:26 crc kubenswrapper[4802]: I1206 03:57:26.681949 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-785f-account-create-update-4x7n2" event={"ID":"b4efbdc9-de14-4774-9c69-66b4cb0f79b9","Type":"ContainerStarted","Data":"4a7734f9c96b23d0e3b83711be5b5aa82c3921c002fa3d97844e773dc6496514"} Dec 06 03:57:26 crc kubenswrapper[4802]: I1206 03:57:26.685498 4802 generic.go:334] "Generic (PLEG): container finished" podID="fb7eaedf-a7ed-4c1f-8609-456afd84daa6" containerID="1fd6bec64ef8692b5a1e61e69e2d0eed99d6c95b54133c5bdf95b50381d88e38" exitCode=0 Dec 06 03:57:26 crc kubenswrapper[4802]: I1206 03:57:26.685623 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-e5f1-account-create-update-zxdsz" event={"ID":"fb7eaedf-a7ed-4c1f-8609-456afd84daa6","Type":"ContainerDied","Data":"1fd6bec64ef8692b5a1e61e69e2d0eed99d6c95b54133c5bdf95b50381d88e38"} Dec 06 03:57:26 crc kubenswrapper[4802]: I1206 03:57:26.685654 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-e5f1-account-create-update-zxdsz" event={"ID":"fb7eaedf-a7ed-4c1f-8609-456afd84daa6","Type":"ContainerStarted","Data":"62eddb98424ae1cd38cccca3d04e8251ba4d4dd758c8365d1ea46bed4f588fd8"} Dec 06 03:57:26 crc kubenswrapper[4802]: I1206 03:57:26.687850 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-49mm8" event={"ID":"19fbf090-979f-4d84-b774-c9cf98ea1501","Type":"ContainerStarted","Data":"b59967445aad7babb9789c7d2853dc467dec4269e482d759935beb41a5ee1358"} Dec 06 03:57:26 crc kubenswrapper[4802]: I1206 03:57:26.687924 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-49mm8" event={"ID":"19fbf090-979f-4d84-b774-c9cf98ea1501","Type":"ContainerStarted","Data":"df9edf97881a0983b315e67b60ab2f5b48014cf579be17d6f4393eb6fa0f1f7b"} Dec 06 03:57:26 crc kubenswrapper[4802]: I1206 03:57:26.726114 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-49mm8" podStartSLOduration=1.726095704 podStartE2EDuration="1.726095704s" podCreationTimestamp="2025-12-06 03:57:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:57:26.722526486 +0000 UTC m=+1039.594435658" watchObservedRunningTime="2025-12-06 03:57:26.726095704 +0000 UTC m=+1039.598004856" Dec 06 03:57:26 crc kubenswrapper[4802]: I1206 03:57:26.889111 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-ckg82"] Dec 06 03:57:26 crc kubenswrapper[4802]: W1206 03:57:26.894477 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda2ab2ae8_12da_43f3_bcd8_1de34185d285.slice/crio-a48ccb94072e670c18ea0bc2b2c6a87df91b5eeb68c456137269cfd1a1cb8cc2 WatchSource:0}: Error finding container a48ccb94072e670c18ea0bc2b2c6a87df91b5eeb68c456137269cfd1a1cb8cc2: Status 404 returned error can't find the container with id a48ccb94072e670c18ea0bc2b2c6a87df91b5eeb68c456137269cfd1a1cb8cc2 Dec 06 03:57:26 crc kubenswrapper[4802]: I1206 03:57:26.947002 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-633c-account-create-update-ld6ht"] Dec 06 03:57:27 crc kubenswrapper[4802]: I1206 03:57:27.251244 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-6mzpl"] Dec 06 03:57:27 crc kubenswrapper[4802]: I1206 03:57:27.251784 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6bc7876d45-6mzpl" podUID="742a77a2-cc74-4671-9247-18f27582c980" containerName="dnsmasq-dns" containerID="cri-o://7d9e7538763552a56a625a1a2a0cd63ff1b289d4b121bb60255ccc8733bcca4d" gracePeriod=10 Dec 06 03:57:27 crc kubenswrapper[4802]: I1206 03:57:27.297118 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 06 03:57:27 crc kubenswrapper[4802]: I1206 03:57:27.298721 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-77frt"] Dec 06 03:57:27 crc kubenswrapper[4802]: I1206 03:57:27.300052 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-77frt" Dec 06 03:57:27 crc kubenswrapper[4802]: I1206 03:57:27.311148 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-77frt"] Dec 06 03:57:27 crc kubenswrapper[4802]: I1206 03:57:27.411849 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5cf08e89-5a37-4f63-8067-e8c161437deb-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-77frt\" (UID: \"5cf08e89-5a37-4f63-8067-e8c161437deb\") " pod="openstack/dnsmasq-dns-b8fbc5445-77frt" Dec 06 03:57:27 crc kubenswrapper[4802]: I1206 03:57:27.412228 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5cf08e89-5a37-4f63-8067-e8c161437deb-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-77frt\" (UID: \"5cf08e89-5a37-4f63-8067-e8c161437deb\") " pod="openstack/dnsmasq-dns-b8fbc5445-77frt" Dec 06 03:57:27 crc kubenswrapper[4802]: I1206 03:57:27.412351 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5cf08e89-5a37-4f63-8067-e8c161437deb-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-77frt\" (UID: \"5cf08e89-5a37-4f63-8067-e8c161437deb\") " pod="openstack/dnsmasq-dns-b8fbc5445-77frt" Dec 06 03:57:27 crc kubenswrapper[4802]: I1206 03:57:27.412374 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44thn\" (UniqueName: \"kubernetes.io/projected/5cf08e89-5a37-4f63-8067-e8c161437deb-kube-api-access-44thn\") pod \"dnsmasq-dns-b8fbc5445-77frt\" (UID: \"5cf08e89-5a37-4f63-8067-e8c161437deb\") " pod="openstack/dnsmasq-dns-b8fbc5445-77frt" Dec 06 03:57:27 crc kubenswrapper[4802]: I1206 03:57:27.412408 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5cf08e89-5a37-4f63-8067-e8c161437deb-config\") pod \"dnsmasq-dns-b8fbc5445-77frt\" (UID: \"5cf08e89-5a37-4f63-8067-e8c161437deb\") " pod="openstack/dnsmasq-dns-b8fbc5445-77frt" Dec 06 03:57:27 crc kubenswrapper[4802]: I1206 03:57:27.513338 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5cf08e89-5a37-4f63-8067-e8c161437deb-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-77frt\" (UID: \"5cf08e89-5a37-4f63-8067-e8c161437deb\") " pod="openstack/dnsmasq-dns-b8fbc5445-77frt" Dec 06 03:57:27 crc kubenswrapper[4802]: I1206 03:57:27.513387 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44thn\" (UniqueName: \"kubernetes.io/projected/5cf08e89-5a37-4f63-8067-e8c161437deb-kube-api-access-44thn\") pod \"dnsmasq-dns-b8fbc5445-77frt\" (UID: \"5cf08e89-5a37-4f63-8067-e8c161437deb\") " pod="openstack/dnsmasq-dns-b8fbc5445-77frt" Dec 06 03:57:27 crc kubenswrapper[4802]: I1206 03:57:27.513418 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5cf08e89-5a37-4f63-8067-e8c161437deb-config\") pod \"dnsmasq-dns-b8fbc5445-77frt\" (UID: \"5cf08e89-5a37-4f63-8067-e8c161437deb\") " pod="openstack/dnsmasq-dns-b8fbc5445-77frt" Dec 06 03:57:27 crc kubenswrapper[4802]: I1206 03:57:27.513633 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5cf08e89-5a37-4f63-8067-e8c161437deb-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-77frt\" (UID: \"5cf08e89-5a37-4f63-8067-e8c161437deb\") " pod="openstack/dnsmasq-dns-b8fbc5445-77frt" Dec 06 03:57:27 crc kubenswrapper[4802]: I1206 03:57:27.513688 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5cf08e89-5a37-4f63-8067-e8c161437deb-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-77frt\" (UID: \"5cf08e89-5a37-4f63-8067-e8c161437deb\") " pod="openstack/dnsmasq-dns-b8fbc5445-77frt" Dec 06 03:57:27 crc kubenswrapper[4802]: I1206 03:57:27.515950 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5cf08e89-5a37-4f63-8067-e8c161437deb-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-77frt\" (UID: \"5cf08e89-5a37-4f63-8067-e8c161437deb\") " pod="openstack/dnsmasq-dns-b8fbc5445-77frt" Dec 06 03:57:27 crc kubenswrapper[4802]: I1206 03:57:27.516186 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5cf08e89-5a37-4f63-8067-e8c161437deb-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-77frt\" (UID: \"5cf08e89-5a37-4f63-8067-e8c161437deb\") " pod="openstack/dnsmasq-dns-b8fbc5445-77frt" Dec 06 03:57:27 crc kubenswrapper[4802]: I1206 03:57:27.516688 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5cf08e89-5a37-4f63-8067-e8c161437deb-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-77frt\" (UID: \"5cf08e89-5a37-4f63-8067-e8c161437deb\") " pod="openstack/dnsmasq-dns-b8fbc5445-77frt" Dec 06 03:57:27 crc kubenswrapper[4802]: I1206 03:57:27.519324 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5cf08e89-5a37-4f63-8067-e8c161437deb-config\") pod \"dnsmasq-dns-b8fbc5445-77frt\" (UID: \"5cf08e89-5a37-4f63-8067-e8c161437deb\") " pod="openstack/dnsmasq-dns-b8fbc5445-77frt" Dec 06 03:57:27 crc kubenswrapper[4802]: I1206 03:57:27.536070 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44thn\" (UniqueName: \"kubernetes.io/projected/5cf08e89-5a37-4f63-8067-e8c161437deb-kube-api-access-44thn\") pod \"dnsmasq-dns-b8fbc5445-77frt\" (UID: \"5cf08e89-5a37-4f63-8067-e8c161437deb\") " pod="openstack/dnsmasq-dns-b8fbc5445-77frt" Dec 06 03:57:27 crc kubenswrapper[4802]: I1206 03:57:27.658290 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-77frt" Dec 06 03:57:27 crc kubenswrapper[4802]: I1206 03:57:27.698576 4802 generic.go:334] "Generic (PLEG): container finished" podID="a2ab2ae8-12da-43f3-bcd8-1de34185d285" containerID="f67ec7c76f73d0c62d661119f023a38670154a30378ef57061af626c7fd1edef" exitCode=0 Dec 06 03:57:27 crc kubenswrapper[4802]: I1206 03:57:27.698637 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-ckg82" event={"ID":"a2ab2ae8-12da-43f3-bcd8-1de34185d285","Type":"ContainerDied","Data":"f67ec7c76f73d0c62d661119f023a38670154a30378ef57061af626c7fd1edef"} Dec 06 03:57:27 crc kubenswrapper[4802]: I1206 03:57:27.698664 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-ckg82" event={"ID":"a2ab2ae8-12da-43f3-bcd8-1de34185d285","Type":"ContainerStarted","Data":"a48ccb94072e670c18ea0bc2b2c6a87df91b5eeb68c456137269cfd1a1cb8cc2"} Dec 06 03:57:27 crc kubenswrapper[4802]: I1206 03:57:27.700076 4802 generic.go:334] "Generic (PLEG): container finished" podID="b4efbdc9-de14-4774-9c69-66b4cb0f79b9" containerID="41aae8e61a7c611287056fd7264028019934d014219736f29d2f668182695dce" exitCode=0 Dec 06 03:57:27 crc kubenswrapper[4802]: I1206 03:57:27.700144 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-785f-account-create-update-4x7n2" event={"ID":"b4efbdc9-de14-4774-9c69-66b4cb0f79b9","Type":"ContainerDied","Data":"41aae8e61a7c611287056fd7264028019934d014219736f29d2f668182695dce"} Dec 06 03:57:27 crc kubenswrapper[4802]: I1206 03:57:27.702226 4802 generic.go:334] "Generic (PLEG): container finished" podID="19fbf090-979f-4d84-b774-c9cf98ea1501" containerID="b59967445aad7babb9789c7d2853dc467dec4269e482d759935beb41a5ee1358" exitCode=0 Dec 06 03:57:27 crc kubenswrapper[4802]: I1206 03:57:27.702329 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-49mm8" event={"ID":"19fbf090-979f-4d84-b774-c9cf98ea1501","Type":"ContainerDied","Data":"b59967445aad7babb9789c7d2853dc467dec4269e482d759935beb41a5ee1358"} Dec 06 03:57:27 crc kubenswrapper[4802]: I1206 03:57:27.703803 4802 generic.go:334] "Generic (PLEG): container finished" podID="7a4a22d0-ccb1-40ec-9872-05b41d788fd4" containerID="067fef1bf0135ad1b23802007a5d041456aea47d26cc06bd9c2d118a45dee413" exitCode=0 Dec 06 03:57:27 crc kubenswrapper[4802]: I1206 03:57:27.703881 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-633c-account-create-update-ld6ht" event={"ID":"7a4a22d0-ccb1-40ec-9872-05b41d788fd4","Type":"ContainerDied","Data":"067fef1bf0135ad1b23802007a5d041456aea47d26cc06bd9c2d118a45dee413"} Dec 06 03:57:27 crc kubenswrapper[4802]: I1206 03:57:27.703907 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-633c-account-create-update-ld6ht" event={"ID":"7a4a22d0-ccb1-40ec-9872-05b41d788fd4","Type":"ContainerStarted","Data":"15b03d33dae569129109dd769578dd6c2fede2e52547ffd65feca2930833ea29"} Dec 06 03:57:27 crc kubenswrapper[4802]: I1206 03:57:27.706282 4802 generic.go:334] "Generic (PLEG): container finished" podID="76accf7a-8aeb-48a1-88c5-3b64e8a0d25e" containerID="2770947b36cbd725b7ffe56bbc6f88a033ccb57bda0f1825d49e4634b4aac59c" exitCode=0 Dec 06 03:57:27 crc kubenswrapper[4802]: I1206 03:57:27.706289 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-l2tpj" event={"ID":"76accf7a-8aeb-48a1-88c5-3b64e8a0d25e","Type":"ContainerDied","Data":"2770947b36cbd725b7ffe56bbc6f88a033ccb57bda0f1825d49e4634b4aac59c"} Dec 06 03:57:27 crc kubenswrapper[4802]: I1206 03:57:27.715033 4802 generic.go:334] "Generic (PLEG): container finished" podID="742a77a2-cc74-4671-9247-18f27582c980" containerID="7d9e7538763552a56a625a1a2a0cd63ff1b289d4b121bb60255ccc8733bcca4d" exitCode=0 Dec 06 03:57:27 crc kubenswrapper[4802]: I1206 03:57:27.715428 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-6mzpl" event={"ID":"742a77a2-cc74-4671-9247-18f27582c980","Type":"ContainerDied","Data":"7d9e7538763552a56a625a1a2a0cd63ff1b289d4b121bb60255ccc8733bcca4d"} Dec 06 03:57:27 crc kubenswrapper[4802]: I1206 03:57:27.753086 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-6mzpl" Dec 06 03:57:27 crc kubenswrapper[4802]: I1206 03:57:27.818431 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/742a77a2-cc74-4671-9247-18f27582c980-ovsdbserver-sb\") pod \"742a77a2-cc74-4671-9247-18f27582c980\" (UID: \"742a77a2-cc74-4671-9247-18f27582c980\") " Dec 06 03:57:27 crc kubenswrapper[4802]: I1206 03:57:27.818684 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/742a77a2-cc74-4671-9247-18f27582c980-dns-svc\") pod \"742a77a2-cc74-4671-9247-18f27582c980\" (UID: \"742a77a2-cc74-4671-9247-18f27582c980\") " Dec 06 03:57:27 crc kubenswrapper[4802]: I1206 03:57:27.818722 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xw4pn\" (UniqueName: \"kubernetes.io/projected/742a77a2-cc74-4671-9247-18f27582c980-kube-api-access-xw4pn\") pod \"742a77a2-cc74-4671-9247-18f27582c980\" (UID: \"742a77a2-cc74-4671-9247-18f27582c980\") " Dec 06 03:57:27 crc kubenswrapper[4802]: I1206 03:57:27.818786 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/742a77a2-cc74-4671-9247-18f27582c980-config\") pod \"742a77a2-cc74-4671-9247-18f27582c980\" (UID: \"742a77a2-cc74-4671-9247-18f27582c980\") " Dec 06 03:57:27 crc kubenswrapper[4802]: I1206 03:57:27.841010 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/742a77a2-cc74-4671-9247-18f27582c980-kube-api-access-xw4pn" (OuterVolumeSpecName: "kube-api-access-xw4pn") pod "742a77a2-cc74-4671-9247-18f27582c980" (UID: "742a77a2-cc74-4671-9247-18f27582c980"). InnerVolumeSpecName "kube-api-access-xw4pn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:57:27 crc kubenswrapper[4802]: I1206 03:57:27.922192 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xw4pn\" (UniqueName: \"kubernetes.io/projected/742a77a2-cc74-4671-9247-18f27582c980-kube-api-access-xw4pn\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:27 crc kubenswrapper[4802]: I1206 03:57:27.923270 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/742a77a2-cc74-4671-9247-18f27582c980-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "742a77a2-cc74-4671-9247-18f27582c980" (UID: "742a77a2-cc74-4671-9247-18f27582c980"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:57:27 crc kubenswrapper[4802]: I1206 03:57:27.963198 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/742a77a2-cc74-4671-9247-18f27582c980-config" (OuterVolumeSpecName: "config") pod "742a77a2-cc74-4671-9247-18f27582c980" (UID: "742a77a2-cc74-4671-9247-18f27582c980"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:57:27 crc kubenswrapper[4802]: I1206 03:57:27.977028 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/742a77a2-cc74-4671-9247-18f27582c980-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "742a77a2-cc74-4671-9247-18f27582c980" (UID: "742a77a2-cc74-4671-9247-18f27582c980"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.023467 4802 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/742a77a2-cc74-4671-9247-18f27582c980-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.023511 4802 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/742a77a2-cc74-4671-9247-18f27582c980-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.023523 4802 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/742a77a2-cc74-4671-9247-18f27582c980-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.221434 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-e5f1-account-create-update-zxdsz" Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.326974 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fb7eaedf-a7ed-4c1f-8609-456afd84daa6-operator-scripts\") pod \"fb7eaedf-a7ed-4c1f-8609-456afd84daa6\" (UID: \"fb7eaedf-a7ed-4c1f-8609-456afd84daa6\") " Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.327141 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-58v7z\" (UniqueName: \"kubernetes.io/projected/fb7eaedf-a7ed-4c1f-8609-456afd84daa6-kube-api-access-58v7z\") pod \"fb7eaedf-a7ed-4c1f-8609-456afd84daa6\" (UID: \"fb7eaedf-a7ed-4c1f-8609-456afd84daa6\") " Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.327762 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb7eaedf-a7ed-4c1f-8609-456afd84daa6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fb7eaedf-a7ed-4c1f-8609-456afd84daa6" (UID: "fb7eaedf-a7ed-4c1f-8609-456afd84daa6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.333556 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Dec 06 03:57:28 crc kubenswrapper[4802]: E1206 03:57:28.334146 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb7eaedf-a7ed-4c1f-8609-456afd84daa6" containerName="mariadb-account-create-update" Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.334240 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb7eaedf-a7ed-4c1f-8609-456afd84daa6" containerName="mariadb-account-create-update" Dec 06 03:57:28 crc kubenswrapper[4802]: E1206 03:57:28.334369 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="742a77a2-cc74-4671-9247-18f27582c980" containerName="dnsmasq-dns" Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.334438 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="742a77a2-cc74-4671-9247-18f27582c980" containerName="dnsmasq-dns" Dec 06 03:57:28 crc kubenswrapper[4802]: E1206 03:57:28.334527 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="742a77a2-cc74-4671-9247-18f27582c980" containerName="init" Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.334595 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="742a77a2-cc74-4671-9247-18f27582c980" containerName="init" Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.334877 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb7eaedf-a7ed-4c1f-8609-456afd84daa6" containerName="mariadb-account-create-update" Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.334981 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="742a77a2-cc74-4671-9247-18f27582c980" containerName="dnsmasq-dns" Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.337531 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb7eaedf-a7ed-4c1f-8609-456afd84daa6-kube-api-access-58v7z" (OuterVolumeSpecName: "kube-api-access-58v7z") pod "fb7eaedf-a7ed-4c1f-8609-456afd84daa6" (UID: "fb7eaedf-a7ed-4c1f-8609-456afd84daa6"). InnerVolumeSpecName "kube-api-access-58v7z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.341843 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.343882 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.344093 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.344264 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-9t286" Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.345496 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.351939 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.398198 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-77frt"] Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.428366 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/508d9d64-dd16-4d21-b492-052d123f1a6a-etc-swift\") pod \"swift-storage-0\" (UID: \"508d9d64-dd16-4d21-b492-052d123f1a6a\") " pod="openstack/swift-storage-0" Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.428421 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"508d9d64-dd16-4d21-b492-052d123f1a6a\") " pod="openstack/swift-storage-0" Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.428558 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/508d9d64-dd16-4d21-b492-052d123f1a6a-cache\") pod \"swift-storage-0\" (UID: \"508d9d64-dd16-4d21-b492-052d123f1a6a\") " pod="openstack/swift-storage-0" Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.428651 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/508d9d64-dd16-4d21-b492-052d123f1a6a-lock\") pod \"swift-storage-0\" (UID: \"508d9d64-dd16-4d21-b492-052d123f1a6a\") " pod="openstack/swift-storage-0" Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.428678 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blb78\" (UniqueName: \"kubernetes.io/projected/508d9d64-dd16-4d21-b492-052d123f1a6a-kube-api-access-blb78\") pod \"swift-storage-0\" (UID: \"508d9d64-dd16-4d21-b492-052d123f1a6a\") " pod="openstack/swift-storage-0" Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.428763 4802 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fb7eaedf-a7ed-4c1f-8609-456afd84daa6-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.428782 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-58v7z\" (UniqueName: \"kubernetes.io/projected/fb7eaedf-a7ed-4c1f-8609-456afd84daa6-kube-api-access-58v7z\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.530013 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blb78\" (UniqueName: \"kubernetes.io/projected/508d9d64-dd16-4d21-b492-052d123f1a6a-kube-api-access-blb78\") pod \"swift-storage-0\" (UID: \"508d9d64-dd16-4d21-b492-052d123f1a6a\") " pod="openstack/swift-storage-0" Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.530268 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/508d9d64-dd16-4d21-b492-052d123f1a6a-lock\") pod \"swift-storage-0\" (UID: \"508d9d64-dd16-4d21-b492-052d123f1a6a\") " pod="openstack/swift-storage-0" Dec 06 03:57:28 crc kubenswrapper[4802]: E1206 03:57:28.530897 4802 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 06 03:57:28 crc kubenswrapper[4802]: E1206 03:57:28.530928 4802 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.531341 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/508d9d64-dd16-4d21-b492-052d123f1a6a-etc-swift\") pod \"swift-storage-0\" (UID: \"508d9d64-dd16-4d21-b492-052d123f1a6a\") " pod="openstack/swift-storage-0" Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.531389 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"508d9d64-dd16-4d21-b492-052d123f1a6a\") " pod="openstack/swift-storage-0" Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.531430 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/508d9d64-dd16-4d21-b492-052d123f1a6a-cache\") pod \"swift-storage-0\" (UID: \"508d9d64-dd16-4d21-b492-052d123f1a6a\") " pod="openstack/swift-storage-0" Dec 06 03:57:28 crc kubenswrapper[4802]: E1206 03:57:28.532177 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/508d9d64-dd16-4d21-b492-052d123f1a6a-etc-swift podName:508d9d64-dd16-4d21-b492-052d123f1a6a nodeName:}" failed. No retries permitted until 2025-12-06 03:57:29.032154355 +0000 UTC m=+1041.904063507 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/508d9d64-dd16-4d21-b492-052d123f1a6a-etc-swift") pod "swift-storage-0" (UID: "508d9d64-dd16-4d21-b492-052d123f1a6a") : configmap "swift-ring-files" not found Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.532184 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/508d9d64-dd16-4d21-b492-052d123f1a6a-lock\") pod \"swift-storage-0\" (UID: \"508d9d64-dd16-4d21-b492-052d123f1a6a\") " pod="openstack/swift-storage-0" Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.532447 4802 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"508d9d64-dd16-4d21-b492-052d123f1a6a\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/swift-storage-0" Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.532587 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/508d9d64-dd16-4d21-b492-052d123f1a6a-cache\") pod \"swift-storage-0\" (UID: \"508d9d64-dd16-4d21-b492-052d123f1a6a\") " pod="openstack/swift-storage-0" Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.547580 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blb78\" (UniqueName: \"kubernetes.io/projected/508d9d64-dd16-4d21-b492-052d123f1a6a-kube-api-access-blb78\") pod \"swift-storage-0\" (UID: \"508d9d64-dd16-4d21-b492-052d123f1a6a\") " pod="openstack/swift-storage-0" Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.556131 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"508d9d64-dd16-4d21-b492-052d123f1a6a\") " pod="openstack/swift-storage-0" Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.723638 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-6mzpl" event={"ID":"742a77a2-cc74-4671-9247-18f27582c980","Type":"ContainerDied","Data":"32a5d5393050330e787068373b72d70ac85fe7e4147e7e8fa6ecbd69696f49d6"} Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.723682 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-6mzpl" Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.723693 4802 scope.go:117] "RemoveContainer" containerID="7d9e7538763552a56a625a1a2a0cd63ff1b289d4b121bb60255ccc8733bcca4d" Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.726014 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-77frt" event={"ID":"5cf08e89-5a37-4f63-8067-e8c161437deb","Type":"ContainerStarted","Data":"6b964fa18fa81f1dd9eea2df8466f8fb04be19ebac233f2616c463b64af9ddc9"} Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.730071 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-e5f1-account-create-update-zxdsz" event={"ID":"fb7eaedf-a7ed-4c1f-8609-456afd84daa6","Type":"ContainerDied","Data":"62eddb98424ae1cd38cccca3d04e8251ba4d4dd758c8365d1ea46bed4f588fd8"} Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.730111 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="62eddb98424ae1cd38cccca3d04e8251ba4d4dd758c8365d1ea46bed4f588fd8" Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.730215 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-e5f1-account-create-update-zxdsz" Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.759813 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-6mzpl"] Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.765473 4802 scope.go:117] "RemoveContainer" containerID="4ee6be7e3ab0768e135dca75dd0f7881bb7f564368ab8119ed82ba40babfebd2" Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.772633 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-6mzpl"] Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.861529 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-flhmv"] Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.863583 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-flhmv" Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.868994 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.869825 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.869976 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.878326 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-flhmv"] Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.938208 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fp2sw\" (UniqueName: \"kubernetes.io/projected/de4cab50-e92b-4282-bd9d-1b9439373aad-kube-api-access-fp2sw\") pod \"swift-ring-rebalance-flhmv\" (UID: \"de4cab50-e92b-4282-bd9d-1b9439373aad\") " pod="openstack/swift-ring-rebalance-flhmv" Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.938294 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/de4cab50-e92b-4282-bd9d-1b9439373aad-scripts\") pod \"swift-ring-rebalance-flhmv\" (UID: \"de4cab50-e92b-4282-bd9d-1b9439373aad\") " pod="openstack/swift-ring-rebalance-flhmv" Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.938325 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/de4cab50-e92b-4282-bd9d-1b9439373aad-ring-data-devices\") pod \"swift-ring-rebalance-flhmv\" (UID: \"de4cab50-e92b-4282-bd9d-1b9439373aad\") " pod="openstack/swift-ring-rebalance-flhmv" Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.938364 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/de4cab50-e92b-4282-bd9d-1b9439373aad-dispersionconf\") pod \"swift-ring-rebalance-flhmv\" (UID: \"de4cab50-e92b-4282-bd9d-1b9439373aad\") " pod="openstack/swift-ring-rebalance-flhmv" Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.938485 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/de4cab50-e92b-4282-bd9d-1b9439373aad-etc-swift\") pod \"swift-ring-rebalance-flhmv\" (UID: \"de4cab50-e92b-4282-bd9d-1b9439373aad\") " pod="openstack/swift-ring-rebalance-flhmv" Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.938520 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de4cab50-e92b-4282-bd9d-1b9439373aad-combined-ca-bundle\") pod \"swift-ring-rebalance-flhmv\" (UID: \"de4cab50-e92b-4282-bd9d-1b9439373aad\") " pod="openstack/swift-ring-rebalance-flhmv" Dec 06 03:57:28 crc kubenswrapper[4802]: I1206 03:57:28.938559 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/de4cab50-e92b-4282-bd9d-1b9439373aad-swiftconf\") pod \"swift-ring-rebalance-flhmv\" (UID: \"de4cab50-e92b-4282-bd9d-1b9439373aad\") " pod="openstack/swift-ring-rebalance-flhmv" Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.039603 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fp2sw\" (UniqueName: \"kubernetes.io/projected/de4cab50-e92b-4282-bd9d-1b9439373aad-kube-api-access-fp2sw\") pod \"swift-ring-rebalance-flhmv\" (UID: \"de4cab50-e92b-4282-bd9d-1b9439373aad\") " pod="openstack/swift-ring-rebalance-flhmv" Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.039891 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/de4cab50-e92b-4282-bd9d-1b9439373aad-scripts\") pod \"swift-ring-rebalance-flhmv\" (UID: \"de4cab50-e92b-4282-bd9d-1b9439373aad\") " pod="openstack/swift-ring-rebalance-flhmv" Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.039920 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/de4cab50-e92b-4282-bd9d-1b9439373aad-ring-data-devices\") pod \"swift-ring-rebalance-flhmv\" (UID: \"de4cab50-e92b-4282-bd9d-1b9439373aad\") " pod="openstack/swift-ring-rebalance-flhmv" Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.039950 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/de4cab50-e92b-4282-bd9d-1b9439373aad-dispersionconf\") pod \"swift-ring-rebalance-flhmv\" (UID: \"de4cab50-e92b-4282-bd9d-1b9439373aad\") " pod="openstack/swift-ring-rebalance-flhmv" Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.039973 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/508d9d64-dd16-4d21-b492-052d123f1a6a-etc-swift\") pod \"swift-storage-0\" (UID: \"508d9d64-dd16-4d21-b492-052d123f1a6a\") " pod="openstack/swift-storage-0" Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.040001 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/de4cab50-e92b-4282-bd9d-1b9439373aad-etc-swift\") pod \"swift-ring-rebalance-flhmv\" (UID: \"de4cab50-e92b-4282-bd9d-1b9439373aad\") " pod="openstack/swift-ring-rebalance-flhmv" Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.040027 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de4cab50-e92b-4282-bd9d-1b9439373aad-combined-ca-bundle\") pod \"swift-ring-rebalance-flhmv\" (UID: \"de4cab50-e92b-4282-bd9d-1b9439373aad\") " pod="openstack/swift-ring-rebalance-flhmv" Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.040058 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/de4cab50-e92b-4282-bd9d-1b9439373aad-swiftconf\") pod \"swift-ring-rebalance-flhmv\" (UID: \"de4cab50-e92b-4282-bd9d-1b9439373aad\") " pod="openstack/swift-ring-rebalance-flhmv" Dec 06 03:57:29 crc kubenswrapper[4802]: E1206 03:57:29.040096 4802 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 06 03:57:29 crc kubenswrapper[4802]: E1206 03:57:29.040119 4802 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 06 03:57:29 crc kubenswrapper[4802]: E1206 03:57:29.040165 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/508d9d64-dd16-4d21-b492-052d123f1a6a-etc-swift podName:508d9d64-dd16-4d21-b492-052d123f1a6a nodeName:}" failed. No retries permitted until 2025-12-06 03:57:30.040147686 +0000 UTC m=+1042.912056898 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/508d9d64-dd16-4d21-b492-052d123f1a6a-etc-swift") pod "swift-storage-0" (UID: "508d9d64-dd16-4d21-b492-052d123f1a6a") : configmap "swift-ring-files" not found Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.041026 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/de4cab50-e92b-4282-bd9d-1b9439373aad-scripts\") pod \"swift-ring-rebalance-flhmv\" (UID: \"de4cab50-e92b-4282-bd9d-1b9439373aad\") " pod="openstack/swift-ring-rebalance-flhmv" Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.041040 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/de4cab50-e92b-4282-bd9d-1b9439373aad-ring-data-devices\") pod \"swift-ring-rebalance-flhmv\" (UID: \"de4cab50-e92b-4282-bd9d-1b9439373aad\") " pod="openstack/swift-ring-rebalance-flhmv" Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.041801 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/de4cab50-e92b-4282-bd9d-1b9439373aad-etc-swift\") pod \"swift-ring-rebalance-flhmv\" (UID: \"de4cab50-e92b-4282-bd9d-1b9439373aad\") " pod="openstack/swift-ring-rebalance-flhmv" Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.048273 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/de4cab50-e92b-4282-bd9d-1b9439373aad-dispersionconf\") pod \"swift-ring-rebalance-flhmv\" (UID: \"de4cab50-e92b-4282-bd9d-1b9439373aad\") " pod="openstack/swift-ring-rebalance-flhmv" Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.048357 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/de4cab50-e92b-4282-bd9d-1b9439373aad-swiftconf\") pod \"swift-ring-rebalance-flhmv\" (UID: \"de4cab50-e92b-4282-bd9d-1b9439373aad\") " pod="openstack/swift-ring-rebalance-flhmv" Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.052356 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de4cab50-e92b-4282-bd9d-1b9439373aad-combined-ca-bundle\") pod \"swift-ring-rebalance-flhmv\" (UID: \"de4cab50-e92b-4282-bd9d-1b9439373aad\") " pod="openstack/swift-ring-rebalance-flhmv" Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.062495 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fp2sw\" (UniqueName: \"kubernetes.io/projected/de4cab50-e92b-4282-bd9d-1b9439373aad-kube-api-access-fp2sw\") pod \"swift-ring-rebalance-flhmv\" (UID: \"de4cab50-e92b-4282-bd9d-1b9439373aad\") " pod="openstack/swift-ring-rebalance-flhmv" Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.122957 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-49mm8" Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.141180 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6w7pl\" (UniqueName: \"kubernetes.io/projected/19fbf090-979f-4d84-b774-c9cf98ea1501-kube-api-access-6w7pl\") pod \"19fbf090-979f-4d84-b774-c9cf98ea1501\" (UID: \"19fbf090-979f-4d84-b774-c9cf98ea1501\") " Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.141356 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/19fbf090-979f-4d84-b774-c9cf98ea1501-operator-scripts\") pod \"19fbf090-979f-4d84-b774-c9cf98ea1501\" (UID: \"19fbf090-979f-4d84-b774-c9cf98ea1501\") " Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.142434 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19fbf090-979f-4d84-b774-c9cf98ea1501-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "19fbf090-979f-4d84-b774-c9cf98ea1501" (UID: "19fbf090-979f-4d84-b774-c9cf98ea1501"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.156810 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19fbf090-979f-4d84-b774-c9cf98ea1501-kube-api-access-6w7pl" (OuterVolumeSpecName: "kube-api-access-6w7pl") pod "19fbf090-979f-4d84-b774-c9cf98ea1501" (UID: "19fbf090-979f-4d84-b774-c9cf98ea1501"). InnerVolumeSpecName "kube-api-access-6w7pl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.201305 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-flhmv" Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.246285 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-633c-account-create-update-ld6ht" Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.247471 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6w7pl\" (UniqueName: \"kubernetes.io/projected/19fbf090-979f-4d84-b774-c9cf98ea1501-kube-api-access-6w7pl\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.247571 4802 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/19fbf090-979f-4d84-b774-c9cf98ea1501-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.329202 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-785f-account-create-update-4x7n2" Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.335195 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-l2tpj" Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.340624 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-ckg82" Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.348393 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8flqx\" (UniqueName: \"kubernetes.io/projected/b4efbdc9-de14-4774-9c69-66b4cb0f79b9-kube-api-access-8flqx\") pod \"b4efbdc9-de14-4774-9c69-66b4cb0f79b9\" (UID: \"b4efbdc9-de14-4774-9c69-66b4cb0f79b9\") " Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.348441 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-phxnv\" (UniqueName: \"kubernetes.io/projected/76accf7a-8aeb-48a1-88c5-3b64e8a0d25e-kube-api-access-phxnv\") pod \"76accf7a-8aeb-48a1-88c5-3b64e8a0d25e\" (UID: \"76accf7a-8aeb-48a1-88c5-3b64e8a0d25e\") " Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.348532 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/76accf7a-8aeb-48a1-88c5-3b64e8a0d25e-operator-scripts\") pod \"76accf7a-8aeb-48a1-88c5-3b64e8a0d25e\" (UID: \"76accf7a-8aeb-48a1-88c5-3b64e8a0d25e\") " Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.348569 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7a4a22d0-ccb1-40ec-9872-05b41d788fd4-operator-scripts\") pod \"7a4a22d0-ccb1-40ec-9872-05b41d788fd4\" (UID: \"7a4a22d0-ccb1-40ec-9872-05b41d788fd4\") " Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.348646 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qdcqq\" (UniqueName: \"kubernetes.io/projected/7a4a22d0-ccb1-40ec-9872-05b41d788fd4-kube-api-access-qdcqq\") pod \"7a4a22d0-ccb1-40ec-9872-05b41d788fd4\" (UID: \"7a4a22d0-ccb1-40ec-9872-05b41d788fd4\") " Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.348674 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b4efbdc9-de14-4774-9c69-66b4cb0f79b9-operator-scripts\") pod \"b4efbdc9-de14-4774-9c69-66b4cb0f79b9\" (UID: \"b4efbdc9-de14-4774-9c69-66b4cb0f79b9\") " Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.349892 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4efbdc9-de14-4774-9c69-66b4cb0f79b9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b4efbdc9-de14-4774-9c69-66b4cb0f79b9" (UID: "b4efbdc9-de14-4774-9c69-66b4cb0f79b9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.350487 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a4a22d0-ccb1-40ec-9872-05b41d788fd4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7a4a22d0-ccb1-40ec-9872-05b41d788fd4" (UID: "7a4a22d0-ccb1-40ec-9872-05b41d788fd4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.356514 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4efbdc9-de14-4774-9c69-66b4cb0f79b9-kube-api-access-8flqx" (OuterVolumeSpecName: "kube-api-access-8flqx") pod "b4efbdc9-de14-4774-9c69-66b4cb0f79b9" (UID: "b4efbdc9-de14-4774-9c69-66b4cb0f79b9"). InnerVolumeSpecName "kube-api-access-8flqx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.356606 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76accf7a-8aeb-48a1-88c5-3b64e8a0d25e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "76accf7a-8aeb-48a1-88c5-3b64e8a0d25e" (UID: "76accf7a-8aeb-48a1-88c5-3b64e8a0d25e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.369323 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76accf7a-8aeb-48a1-88c5-3b64e8a0d25e-kube-api-access-phxnv" (OuterVolumeSpecName: "kube-api-access-phxnv") pod "76accf7a-8aeb-48a1-88c5-3b64e8a0d25e" (UID: "76accf7a-8aeb-48a1-88c5-3b64e8a0d25e"). InnerVolumeSpecName "kube-api-access-phxnv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.372422 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a4a22d0-ccb1-40ec-9872-05b41d788fd4-kube-api-access-qdcqq" (OuterVolumeSpecName: "kube-api-access-qdcqq") pod "7a4a22d0-ccb1-40ec-9872-05b41d788fd4" (UID: "7a4a22d0-ccb1-40ec-9872-05b41d788fd4"). InnerVolumeSpecName "kube-api-access-qdcqq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.450535 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a2ab2ae8-12da-43f3-bcd8-1de34185d285-operator-scripts\") pod \"a2ab2ae8-12da-43f3-bcd8-1de34185d285\" (UID: \"a2ab2ae8-12da-43f3-bcd8-1de34185d285\") " Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.450644 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ssgkv\" (UniqueName: \"kubernetes.io/projected/a2ab2ae8-12da-43f3-bcd8-1de34185d285-kube-api-access-ssgkv\") pod \"a2ab2ae8-12da-43f3-bcd8-1de34185d285\" (UID: \"a2ab2ae8-12da-43f3-bcd8-1de34185d285\") " Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.451138 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2ab2ae8-12da-43f3-bcd8-1de34185d285-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a2ab2ae8-12da-43f3-bcd8-1de34185d285" (UID: "a2ab2ae8-12da-43f3-bcd8-1de34185d285"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.451883 4802 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/76accf7a-8aeb-48a1-88c5-3b64e8a0d25e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.451922 4802 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7a4a22d0-ccb1-40ec-9872-05b41d788fd4-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.451937 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qdcqq\" (UniqueName: \"kubernetes.io/projected/7a4a22d0-ccb1-40ec-9872-05b41d788fd4-kube-api-access-qdcqq\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.451951 4802 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b4efbdc9-de14-4774-9c69-66b4cb0f79b9-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.451963 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8flqx\" (UniqueName: \"kubernetes.io/projected/b4efbdc9-de14-4774-9c69-66b4cb0f79b9-kube-api-access-8flqx\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.451976 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-phxnv\" (UniqueName: \"kubernetes.io/projected/76accf7a-8aeb-48a1-88c5-3b64e8a0d25e-kube-api-access-phxnv\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.453791 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2ab2ae8-12da-43f3-bcd8-1de34185d285-kube-api-access-ssgkv" (OuterVolumeSpecName: "kube-api-access-ssgkv") pod "a2ab2ae8-12da-43f3-bcd8-1de34185d285" (UID: "a2ab2ae8-12da-43f3-bcd8-1de34185d285"). InnerVolumeSpecName "kube-api-access-ssgkv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.461934 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="742a77a2-cc74-4671-9247-18f27582c980" path="/var/lib/kubelet/pods/742a77a2-cc74-4671-9247-18f27582c980/volumes" Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.553396 4802 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a2ab2ae8-12da-43f3-bcd8-1de34185d285-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.553431 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ssgkv\" (UniqueName: \"kubernetes.io/projected/a2ab2ae8-12da-43f3-bcd8-1de34185d285-kube-api-access-ssgkv\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:29 crc kubenswrapper[4802]: W1206 03:57:29.713327 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podde4cab50_e92b_4282_bd9d_1b9439373aad.slice/crio-fb63b9f045d40054ecdcfe61fb7fa7ab4384160fdc80f6eaf64870bffd41175e WatchSource:0}: Error finding container fb63b9f045d40054ecdcfe61fb7fa7ab4384160fdc80f6eaf64870bffd41175e: Status 404 returned error can't find the container with id fb63b9f045d40054ecdcfe61fb7fa7ab4384160fdc80f6eaf64870bffd41175e Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.717049 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-flhmv"] Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.740184 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-785f-account-create-update-4x7n2" event={"ID":"b4efbdc9-de14-4774-9c69-66b4cb0f79b9","Type":"ContainerDied","Data":"4a7734f9c96b23d0e3b83711be5b5aa82c3921c002fa3d97844e773dc6496514"} Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.740228 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a7734f9c96b23d0e3b83711be5b5aa82c3921c002fa3d97844e773dc6496514" Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.740298 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-785f-account-create-update-4x7n2" Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.742802 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-flhmv" event={"ID":"de4cab50-e92b-4282-bd9d-1b9439373aad","Type":"ContainerStarted","Data":"fb63b9f045d40054ecdcfe61fb7fa7ab4384160fdc80f6eaf64870bffd41175e"} Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.745096 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-49mm8" event={"ID":"19fbf090-979f-4d84-b774-c9cf98ea1501","Type":"ContainerDied","Data":"df9edf97881a0983b315e67b60ab2f5b48014cf579be17d6f4393eb6fa0f1f7b"} Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.745131 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df9edf97881a0983b315e67b60ab2f5b48014cf579be17d6f4393eb6fa0f1f7b" Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.745132 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-49mm8" Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.746695 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-633c-account-create-update-ld6ht" event={"ID":"7a4a22d0-ccb1-40ec-9872-05b41d788fd4","Type":"ContainerDied","Data":"15b03d33dae569129109dd769578dd6c2fede2e52547ffd65feca2930833ea29"} Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.746723 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="15b03d33dae569129109dd769578dd6c2fede2e52547ffd65feca2930833ea29" Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.746803 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-633c-account-create-update-ld6ht" Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.749444 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-l2tpj" event={"ID":"76accf7a-8aeb-48a1-88c5-3b64e8a0d25e","Type":"ContainerDied","Data":"c07816566bd8038b13d768ab8a6e9508a4ffa8e8b72173563ef7217967b27527"} Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.749488 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-l2tpj" Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.749498 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c07816566bd8038b13d768ab8a6e9508a4ffa8e8b72173563ef7217967b27527" Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.752268 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-ckg82" event={"ID":"a2ab2ae8-12da-43f3-bcd8-1de34185d285","Type":"ContainerDied","Data":"a48ccb94072e670c18ea0bc2b2c6a87df91b5eeb68c456137269cfd1a1cb8cc2"} Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.752305 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a48ccb94072e670c18ea0bc2b2c6a87df91b5eeb68c456137269cfd1a1cb8cc2" Dec 06 03:57:29 crc kubenswrapper[4802]: I1206 03:57:29.752314 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-ckg82" Dec 06 03:57:30 crc kubenswrapper[4802]: I1206 03:57:30.062084 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/508d9d64-dd16-4d21-b492-052d123f1a6a-etc-swift\") pod \"swift-storage-0\" (UID: \"508d9d64-dd16-4d21-b492-052d123f1a6a\") " pod="openstack/swift-storage-0" Dec 06 03:57:30 crc kubenswrapper[4802]: E1206 03:57:30.062637 4802 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 06 03:57:30 crc kubenswrapper[4802]: E1206 03:57:30.062659 4802 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 06 03:57:30 crc kubenswrapper[4802]: E1206 03:57:30.062709 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/508d9d64-dd16-4d21-b492-052d123f1a6a-etc-swift podName:508d9d64-dd16-4d21-b492-052d123f1a6a nodeName:}" failed. No retries permitted until 2025-12-06 03:57:32.062692288 +0000 UTC m=+1044.934601440 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/508d9d64-dd16-4d21-b492-052d123f1a6a-etc-swift") pod "swift-storage-0" (UID: "508d9d64-dd16-4d21-b492-052d123f1a6a") : configmap "swift-ring-files" not found Dec 06 03:57:31 crc kubenswrapper[4802]: I1206 03:57:31.179983 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-mlhfp"] Dec 06 03:57:31 crc kubenswrapper[4802]: E1206 03:57:31.180357 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2ab2ae8-12da-43f3-bcd8-1de34185d285" containerName="mariadb-database-create" Dec 06 03:57:31 crc kubenswrapper[4802]: I1206 03:57:31.180373 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2ab2ae8-12da-43f3-bcd8-1de34185d285" containerName="mariadb-database-create" Dec 06 03:57:31 crc kubenswrapper[4802]: E1206 03:57:31.180396 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a4a22d0-ccb1-40ec-9872-05b41d788fd4" containerName="mariadb-account-create-update" Dec 06 03:57:31 crc kubenswrapper[4802]: I1206 03:57:31.180405 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a4a22d0-ccb1-40ec-9872-05b41d788fd4" containerName="mariadb-account-create-update" Dec 06 03:57:31 crc kubenswrapper[4802]: E1206 03:57:31.180437 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4efbdc9-de14-4774-9c69-66b4cb0f79b9" containerName="mariadb-account-create-update" Dec 06 03:57:31 crc kubenswrapper[4802]: I1206 03:57:31.180447 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4efbdc9-de14-4774-9c69-66b4cb0f79b9" containerName="mariadb-account-create-update" Dec 06 03:57:31 crc kubenswrapper[4802]: E1206 03:57:31.180461 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76accf7a-8aeb-48a1-88c5-3b64e8a0d25e" containerName="mariadb-database-create" Dec 06 03:57:31 crc kubenswrapper[4802]: I1206 03:57:31.180470 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="76accf7a-8aeb-48a1-88c5-3b64e8a0d25e" containerName="mariadb-database-create" Dec 06 03:57:31 crc kubenswrapper[4802]: E1206 03:57:31.180481 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19fbf090-979f-4d84-b774-c9cf98ea1501" containerName="mariadb-database-create" Dec 06 03:57:31 crc kubenswrapper[4802]: I1206 03:57:31.180489 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="19fbf090-979f-4d84-b774-c9cf98ea1501" containerName="mariadb-database-create" Dec 06 03:57:31 crc kubenswrapper[4802]: I1206 03:57:31.180692 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4efbdc9-de14-4774-9c69-66b4cb0f79b9" containerName="mariadb-account-create-update" Dec 06 03:57:31 crc kubenswrapper[4802]: I1206 03:57:31.180710 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="19fbf090-979f-4d84-b774-c9cf98ea1501" containerName="mariadb-database-create" Dec 06 03:57:31 crc kubenswrapper[4802]: I1206 03:57:31.180729 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a4a22d0-ccb1-40ec-9872-05b41d788fd4" containerName="mariadb-account-create-update" Dec 06 03:57:31 crc kubenswrapper[4802]: I1206 03:57:31.180772 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="76accf7a-8aeb-48a1-88c5-3b64e8a0d25e" containerName="mariadb-database-create" Dec 06 03:57:31 crc kubenswrapper[4802]: I1206 03:57:31.180789 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2ab2ae8-12da-43f3-bcd8-1de34185d285" containerName="mariadb-database-create" Dec 06 03:57:31 crc kubenswrapper[4802]: I1206 03:57:31.181393 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-mlhfp" Dec 06 03:57:31 crc kubenswrapper[4802]: I1206 03:57:31.183261 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 06 03:57:31 crc kubenswrapper[4802]: I1206 03:57:31.185461 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-n458c" Dec 06 03:57:31 crc kubenswrapper[4802]: I1206 03:57:31.194891 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-mlhfp"] Dec 06 03:57:31 crc kubenswrapper[4802]: I1206 03:57:31.281665 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xt576\" (UniqueName: \"kubernetes.io/projected/5d947661-37cd-49c2-bd89-e927e4034ba4-kube-api-access-xt576\") pod \"glance-db-sync-mlhfp\" (UID: \"5d947661-37cd-49c2-bd89-e927e4034ba4\") " pod="openstack/glance-db-sync-mlhfp" Dec 06 03:57:31 crc kubenswrapper[4802]: I1206 03:57:31.281761 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5d947661-37cd-49c2-bd89-e927e4034ba4-db-sync-config-data\") pod \"glance-db-sync-mlhfp\" (UID: \"5d947661-37cd-49c2-bd89-e927e4034ba4\") " pod="openstack/glance-db-sync-mlhfp" Dec 06 03:57:31 crc kubenswrapper[4802]: I1206 03:57:31.281918 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d947661-37cd-49c2-bd89-e927e4034ba4-combined-ca-bundle\") pod \"glance-db-sync-mlhfp\" (UID: \"5d947661-37cd-49c2-bd89-e927e4034ba4\") " pod="openstack/glance-db-sync-mlhfp" Dec 06 03:57:31 crc kubenswrapper[4802]: I1206 03:57:31.281954 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d947661-37cd-49c2-bd89-e927e4034ba4-config-data\") pod \"glance-db-sync-mlhfp\" (UID: \"5d947661-37cd-49c2-bd89-e927e4034ba4\") " pod="openstack/glance-db-sync-mlhfp" Dec 06 03:57:31 crc kubenswrapper[4802]: I1206 03:57:31.384336 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5d947661-37cd-49c2-bd89-e927e4034ba4-db-sync-config-data\") pod \"glance-db-sync-mlhfp\" (UID: \"5d947661-37cd-49c2-bd89-e927e4034ba4\") " pod="openstack/glance-db-sync-mlhfp" Dec 06 03:57:31 crc kubenswrapper[4802]: I1206 03:57:31.384497 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d947661-37cd-49c2-bd89-e927e4034ba4-combined-ca-bundle\") pod \"glance-db-sync-mlhfp\" (UID: \"5d947661-37cd-49c2-bd89-e927e4034ba4\") " pod="openstack/glance-db-sync-mlhfp" Dec 06 03:57:31 crc kubenswrapper[4802]: I1206 03:57:31.384531 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d947661-37cd-49c2-bd89-e927e4034ba4-config-data\") pod \"glance-db-sync-mlhfp\" (UID: \"5d947661-37cd-49c2-bd89-e927e4034ba4\") " pod="openstack/glance-db-sync-mlhfp" Dec 06 03:57:31 crc kubenswrapper[4802]: I1206 03:57:31.384566 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xt576\" (UniqueName: \"kubernetes.io/projected/5d947661-37cd-49c2-bd89-e927e4034ba4-kube-api-access-xt576\") pod \"glance-db-sync-mlhfp\" (UID: \"5d947661-37cd-49c2-bd89-e927e4034ba4\") " pod="openstack/glance-db-sync-mlhfp" Dec 06 03:57:31 crc kubenswrapper[4802]: I1206 03:57:31.393771 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d947661-37cd-49c2-bd89-e927e4034ba4-config-data\") pod \"glance-db-sync-mlhfp\" (UID: \"5d947661-37cd-49c2-bd89-e927e4034ba4\") " pod="openstack/glance-db-sync-mlhfp" Dec 06 03:57:31 crc kubenswrapper[4802]: I1206 03:57:31.397435 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d947661-37cd-49c2-bd89-e927e4034ba4-combined-ca-bundle\") pod \"glance-db-sync-mlhfp\" (UID: \"5d947661-37cd-49c2-bd89-e927e4034ba4\") " pod="openstack/glance-db-sync-mlhfp" Dec 06 03:57:31 crc kubenswrapper[4802]: I1206 03:57:31.401952 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5d947661-37cd-49c2-bd89-e927e4034ba4-db-sync-config-data\") pod \"glance-db-sync-mlhfp\" (UID: \"5d947661-37cd-49c2-bd89-e927e4034ba4\") " pod="openstack/glance-db-sync-mlhfp" Dec 06 03:57:31 crc kubenswrapper[4802]: I1206 03:57:31.402265 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xt576\" (UniqueName: \"kubernetes.io/projected/5d947661-37cd-49c2-bd89-e927e4034ba4-kube-api-access-xt576\") pod \"glance-db-sync-mlhfp\" (UID: \"5d947661-37cd-49c2-bd89-e927e4034ba4\") " pod="openstack/glance-db-sync-mlhfp" Dec 06 03:57:31 crc kubenswrapper[4802]: I1206 03:57:31.561908 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-mlhfp" Dec 06 03:57:31 crc kubenswrapper[4802]: I1206 03:57:31.773073 4802 generic.go:334] "Generic (PLEG): container finished" podID="5cf08e89-5a37-4f63-8067-e8c161437deb" containerID="850ac4aa9d4757681aa9d6b1c01963349575b45bbbfc69dbd6dd97dae052e093" exitCode=0 Dec 06 03:57:31 crc kubenswrapper[4802]: I1206 03:57:31.773134 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-77frt" event={"ID":"5cf08e89-5a37-4f63-8067-e8c161437deb","Type":"ContainerDied","Data":"850ac4aa9d4757681aa9d6b1c01963349575b45bbbfc69dbd6dd97dae052e093"} Dec 06 03:57:32 crc kubenswrapper[4802]: I1206 03:57:32.070668 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-mlhfp"] Dec 06 03:57:32 crc kubenswrapper[4802]: W1206 03:57:32.072004 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5d947661_37cd_49c2_bd89_e927e4034ba4.slice/crio-ef95b6034d02153fe5279d159ad7e5567b7b3122f29307525c7760694c91ff4f WatchSource:0}: Error finding container ef95b6034d02153fe5279d159ad7e5567b7b3122f29307525c7760694c91ff4f: Status 404 returned error can't find the container with id ef95b6034d02153fe5279d159ad7e5567b7b3122f29307525c7760694c91ff4f Dec 06 03:57:32 crc kubenswrapper[4802]: I1206 03:57:32.098271 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/508d9d64-dd16-4d21-b492-052d123f1a6a-etc-swift\") pod \"swift-storage-0\" (UID: \"508d9d64-dd16-4d21-b492-052d123f1a6a\") " pod="openstack/swift-storage-0" Dec 06 03:57:32 crc kubenswrapper[4802]: E1206 03:57:32.098412 4802 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 06 03:57:32 crc kubenswrapper[4802]: E1206 03:57:32.098504 4802 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 06 03:57:32 crc kubenswrapper[4802]: E1206 03:57:32.098555 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/508d9d64-dd16-4d21-b492-052d123f1a6a-etc-swift podName:508d9d64-dd16-4d21-b492-052d123f1a6a nodeName:}" failed. No retries permitted until 2025-12-06 03:57:36.098538328 +0000 UTC m=+1048.970447480 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/508d9d64-dd16-4d21-b492-052d123f1a6a-etc-swift") pod "swift-storage-0" (UID: "508d9d64-dd16-4d21-b492-052d123f1a6a") : configmap "swift-ring-files" not found Dec 06 03:57:32 crc kubenswrapper[4802]: I1206 03:57:32.483942 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8554648995-4dwc7" Dec 06 03:57:32 crc kubenswrapper[4802]: I1206 03:57:32.785399 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-mlhfp" event={"ID":"5d947661-37cd-49c2-bd89-e927e4034ba4","Type":"ContainerStarted","Data":"ef95b6034d02153fe5279d159ad7e5567b7b3122f29307525c7760694c91ff4f"} Dec 06 03:57:32 crc kubenswrapper[4802]: I1206 03:57:32.794148 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-77frt" event={"ID":"5cf08e89-5a37-4f63-8067-e8c161437deb","Type":"ContainerStarted","Data":"c83ceab3315891cab0cb020370c47d360ceda11dd02b35e3f9db75790ec648da"} Dec 06 03:57:32 crc kubenswrapper[4802]: I1206 03:57:32.794366 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b8fbc5445-77frt" Dec 06 03:57:32 crc kubenswrapper[4802]: I1206 03:57:32.809899 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b8fbc5445-77frt" podStartSLOduration=5.809877935 podStartE2EDuration="5.809877935s" podCreationTimestamp="2025-12-06 03:57:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:57:32.809862795 +0000 UTC m=+1045.681771947" watchObservedRunningTime="2025-12-06 03:57:32.809877935 +0000 UTC m=+1045.681787097" Dec 06 03:57:36 crc kubenswrapper[4802]: I1206 03:57:36.176416 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/508d9d64-dd16-4d21-b492-052d123f1a6a-etc-swift\") pod \"swift-storage-0\" (UID: \"508d9d64-dd16-4d21-b492-052d123f1a6a\") " pod="openstack/swift-storage-0" Dec 06 03:57:36 crc kubenswrapper[4802]: E1206 03:57:36.177132 4802 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 06 03:57:36 crc kubenswrapper[4802]: E1206 03:57:36.177296 4802 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 06 03:57:36 crc kubenswrapper[4802]: E1206 03:57:36.177353 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/508d9d64-dd16-4d21-b492-052d123f1a6a-etc-swift podName:508d9d64-dd16-4d21-b492-052d123f1a6a nodeName:}" failed. No retries permitted until 2025-12-06 03:57:44.177335104 +0000 UTC m=+1057.049244256 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/508d9d64-dd16-4d21-b492-052d123f1a6a-etc-swift") pod "swift-storage-0" (UID: "508d9d64-dd16-4d21-b492-052d123f1a6a") : configmap "swift-ring-files" not found Dec 06 03:57:37 crc kubenswrapper[4802]: I1206 03:57:37.428658 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 06 03:57:37 crc kubenswrapper[4802]: I1206 03:57:37.660903 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b8fbc5445-77frt" Dec 06 03:57:37 crc kubenswrapper[4802]: I1206 03:57:37.717260 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-4dwc7"] Dec 06 03:57:37 crc kubenswrapper[4802]: I1206 03:57:37.717469 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8554648995-4dwc7" podUID="31be02e0-3b12-49e9-a900-3d069fa35ac4" containerName="dnsmasq-dns" containerID="cri-o://53eb2ba35507f38304fc35d966b3530345562545e5fa5c41844dd46be2d6f21d" gracePeriod=10 Dec 06 03:57:40 crc kubenswrapper[4802]: I1206 03:57:40.862282 4802 generic.go:334] "Generic (PLEG): container finished" podID="31be02e0-3b12-49e9-a900-3d069fa35ac4" containerID="53eb2ba35507f38304fc35d966b3530345562545e5fa5c41844dd46be2d6f21d" exitCode=0 Dec 06 03:57:40 crc kubenswrapper[4802]: I1206 03:57:40.862914 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-4dwc7" event={"ID":"31be02e0-3b12-49e9-a900-3d069fa35ac4","Type":"ContainerDied","Data":"53eb2ba35507f38304fc35d966b3530345562545e5fa5c41844dd46be2d6f21d"} Dec 06 03:57:44 crc kubenswrapper[4802]: I1206 03:57:44.144110 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-4dwc7" Dec 06 03:57:44 crc kubenswrapper[4802]: I1206 03:57:44.225246 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/508d9d64-dd16-4d21-b492-052d123f1a6a-etc-swift\") pod \"swift-storage-0\" (UID: \"508d9d64-dd16-4d21-b492-052d123f1a6a\") " pod="openstack/swift-storage-0" Dec 06 03:57:44 crc kubenswrapper[4802]: E1206 03:57:44.225664 4802 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 06 03:57:44 crc kubenswrapper[4802]: E1206 03:57:44.225708 4802 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 06 03:57:44 crc kubenswrapper[4802]: E1206 03:57:44.225880 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/508d9d64-dd16-4d21-b492-052d123f1a6a-etc-swift podName:508d9d64-dd16-4d21-b492-052d123f1a6a nodeName:}" failed. No retries permitted until 2025-12-06 03:58:00.225863654 +0000 UTC m=+1073.097772806 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/508d9d64-dd16-4d21-b492-052d123f1a6a-etc-swift") pod "swift-storage-0" (UID: "508d9d64-dd16-4d21-b492-052d123f1a6a") : configmap "swift-ring-files" not found Dec 06 03:57:44 crc kubenswrapper[4802]: I1206 03:57:44.326661 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kkcxt\" (UniqueName: \"kubernetes.io/projected/31be02e0-3b12-49e9-a900-3d069fa35ac4-kube-api-access-kkcxt\") pod \"31be02e0-3b12-49e9-a900-3d069fa35ac4\" (UID: \"31be02e0-3b12-49e9-a900-3d069fa35ac4\") " Dec 06 03:57:44 crc kubenswrapper[4802]: I1206 03:57:44.326885 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/31be02e0-3b12-49e9-a900-3d069fa35ac4-ovsdbserver-nb\") pod \"31be02e0-3b12-49e9-a900-3d069fa35ac4\" (UID: \"31be02e0-3b12-49e9-a900-3d069fa35ac4\") " Dec 06 03:57:44 crc kubenswrapper[4802]: I1206 03:57:44.326970 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/31be02e0-3b12-49e9-a900-3d069fa35ac4-dns-svc\") pod \"31be02e0-3b12-49e9-a900-3d069fa35ac4\" (UID: \"31be02e0-3b12-49e9-a900-3d069fa35ac4\") " Dec 06 03:57:44 crc kubenswrapper[4802]: I1206 03:57:44.327025 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/31be02e0-3b12-49e9-a900-3d069fa35ac4-ovsdbserver-sb\") pod \"31be02e0-3b12-49e9-a900-3d069fa35ac4\" (UID: \"31be02e0-3b12-49e9-a900-3d069fa35ac4\") " Dec 06 03:57:44 crc kubenswrapper[4802]: I1206 03:57:44.327058 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31be02e0-3b12-49e9-a900-3d069fa35ac4-config\") pod \"31be02e0-3b12-49e9-a900-3d069fa35ac4\" (UID: \"31be02e0-3b12-49e9-a900-3d069fa35ac4\") " Dec 06 03:57:44 crc kubenswrapper[4802]: I1206 03:57:44.344551 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31be02e0-3b12-49e9-a900-3d069fa35ac4-kube-api-access-kkcxt" (OuterVolumeSpecName: "kube-api-access-kkcxt") pod "31be02e0-3b12-49e9-a900-3d069fa35ac4" (UID: "31be02e0-3b12-49e9-a900-3d069fa35ac4"). InnerVolumeSpecName "kube-api-access-kkcxt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:57:44 crc kubenswrapper[4802]: I1206 03:57:44.367272 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31be02e0-3b12-49e9-a900-3d069fa35ac4-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "31be02e0-3b12-49e9-a900-3d069fa35ac4" (UID: "31be02e0-3b12-49e9-a900-3d069fa35ac4"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:57:44 crc kubenswrapper[4802]: I1206 03:57:44.371221 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31be02e0-3b12-49e9-a900-3d069fa35ac4-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "31be02e0-3b12-49e9-a900-3d069fa35ac4" (UID: "31be02e0-3b12-49e9-a900-3d069fa35ac4"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:57:44 crc kubenswrapper[4802]: I1206 03:57:44.372170 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31be02e0-3b12-49e9-a900-3d069fa35ac4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "31be02e0-3b12-49e9-a900-3d069fa35ac4" (UID: "31be02e0-3b12-49e9-a900-3d069fa35ac4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:57:44 crc kubenswrapper[4802]: I1206 03:57:44.376587 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31be02e0-3b12-49e9-a900-3d069fa35ac4-config" (OuterVolumeSpecName: "config") pod "31be02e0-3b12-49e9-a900-3d069fa35ac4" (UID: "31be02e0-3b12-49e9-a900-3d069fa35ac4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:57:44 crc kubenswrapper[4802]: I1206 03:57:44.428767 4802 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/31be02e0-3b12-49e9-a900-3d069fa35ac4-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:44 crc kubenswrapper[4802]: I1206 03:57:44.428803 4802 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/31be02e0-3b12-49e9-a900-3d069fa35ac4-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:44 crc kubenswrapper[4802]: I1206 03:57:44.428815 4802 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/31be02e0-3b12-49e9-a900-3d069fa35ac4-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:44 crc kubenswrapper[4802]: I1206 03:57:44.428823 4802 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31be02e0-3b12-49e9-a900-3d069fa35ac4-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:44 crc kubenswrapper[4802]: I1206 03:57:44.428833 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kkcxt\" (UniqueName: \"kubernetes.io/projected/31be02e0-3b12-49e9-a900-3d069fa35ac4-kube-api-access-kkcxt\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:44 crc kubenswrapper[4802]: I1206 03:57:44.907148 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-4dwc7" event={"ID":"31be02e0-3b12-49e9-a900-3d069fa35ac4","Type":"ContainerDied","Data":"4c6a8faef9e7ceb14acd83bdd99e061a85f513446217176326f96a79e5e01f0d"} Dec 06 03:57:44 crc kubenswrapper[4802]: I1206 03:57:44.907198 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-4dwc7" Dec 06 03:57:44 crc kubenswrapper[4802]: I1206 03:57:44.909403 4802 scope.go:117] "RemoveContainer" containerID="53eb2ba35507f38304fc35d966b3530345562545e5fa5c41844dd46be2d6f21d" Dec 06 03:57:44 crc kubenswrapper[4802]: I1206 03:57:44.910707 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-flhmv" event={"ID":"de4cab50-e92b-4282-bd9d-1b9439373aad","Type":"ContainerStarted","Data":"cc3dcd6dd4b27d5dec3c150b5e3848e004011d94e8ad36cea99a4b8fa28d4c96"} Dec 06 03:57:44 crc kubenswrapper[4802]: I1206 03:57:44.937139 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-flhmv" podStartSLOduration=2.656742721 podStartE2EDuration="16.937120969s" podCreationTimestamp="2025-12-06 03:57:28 +0000 UTC" firstStartedPulling="2025-12-06 03:57:29.71546794 +0000 UTC m=+1042.587377132" lastFinishedPulling="2025-12-06 03:57:43.995846218 +0000 UTC m=+1056.867755380" observedRunningTime="2025-12-06 03:57:44.931341701 +0000 UTC m=+1057.803250853" watchObservedRunningTime="2025-12-06 03:57:44.937120969 +0000 UTC m=+1057.809030121" Dec 06 03:57:44 crc kubenswrapper[4802]: I1206 03:57:44.944660 4802 scope.go:117] "RemoveContainer" containerID="58e6ce4ec8424e87b10a3e1c0ccf3f5b0c841d67d4f4f675da676460244a9096" Dec 06 03:57:44 crc kubenswrapper[4802]: I1206 03:57:44.949936 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-4dwc7"] Dec 06 03:57:44 crc kubenswrapper[4802]: I1206 03:57:44.973401 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8554648995-4dwc7"] Dec 06 03:57:45 crc kubenswrapper[4802]: I1206 03:57:45.459382 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31be02e0-3b12-49e9-a900-3d069fa35ac4" path="/var/lib/kubelet/pods/31be02e0-3b12-49e9-a900-3d069fa35ac4/volumes" Dec 06 03:57:46 crc kubenswrapper[4802]: I1206 03:57:46.165463 4802 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-sbt8p" podUID="6e785e69-07cd-4c5c-8446-3d2a0d7e7b0b" containerName="ovn-controller" probeResult="failure" output=< Dec 06 03:57:46 crc kubenswrapper[4802]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 06 03:57:46 crc kubenswrapper[4802]: > Dec 06 03:57:47 crc kubenswrapper[4802]: I1206 03:57:47.481954 4802 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-8554648995-4dwc7" podUID="31be02e0-3b12-49e9-a900-3d069fa35ac4" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.111:5353: i/o timeout" Dec 06 03:57:49 crc kubenswrapper[4802]: I1206 03:57:49.951030 4802 generic.go:334] "Generic (PLEG): container finished" podID="8c8b6c49-6f71-403d-b639-4a17ae3bc802" containerID="f470e3dda7735fd4bf057dbf62cde306583097f22e2d9c550167fc04c5e92f72" exitCode=0 Dec 06 03:57:49 crc kubenswrapper[4802]: I1206 03:57:49.951115 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8c8b6c49-6f71-403d-b639-4a17ae3bc802","Type":"ContainerDied","Data":"f470e3dda7735fd4bf057dbf62cde306583097f22e2d9c550167fc04c5e92f72"} Dec 06 03:57:49 crc kubenswrapper[4802]: I1206 03:57:49.956286 4802 generic.go:334] "Generic (PLEG): container finished" podID="7b4f22a1-41dc-4a75-8999-53df29d2fb5d" containerID="fd1082f2aaf9365d3d9655c6924fa59bdb4c429364af5f69505c71489894ed4f" exitCode=0 Dec 06 03:57:49 crc kubenswrapper[4802]: I1206 03:57:49.956338 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7b4f22a1-41dc-4a75-8999-53df29d2fb5d","Type":"ContainerDied","Data":"fd1082f2aaf9365d3d9655c6924fa59bdb4c429364af5f69505c71489894ed4f"} Dec 06 03:57:51 crc kubenswrapper[4802]: I1206 03:57:51.160277 4802 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-sbt8p" podUID="6e785e69-07cd-4c5c-8446-3d2a0d7e7b0b" containerName="ovn-controller" probeResult="failure" output=< Dec 06 03:57:51 crc kubenswrapper[4802]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 06 03:57:51 crc kubenswrapper[4802]: > Dec 06 03:57:51 crc kubenswrapper[4802]: I1206 03:57:51.178276 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-49cxn" Dec 06 03:57:51 crc kubenswrapper[4802]: I1206 03:57:51.179038 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-49cxn" Dec 06 03:57:51 crc kubenswrapper[4802]: I1206 03:57:51.399271 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-sbt8p-config-vn5rk"] Dec 06 03:57:51 crc kubenswrapper[4802]: E1206 03:57:51.399711 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31be02e0-3b12-49e9-a900-3d069fa35ac4" containerName="init" Dec 06 03:57:51 crc kubenswrapper[4802]: I1206 03:57:51.399726 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="31be02e0-3b12-49e9-a900-3d069fa35ac4" containerName="init" Dec 06 03:57:51 crc kubenswrapper[4802]: E1206 03:57:51.399798 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31be02e0-3b12-49e9-a900-3d069fa35ac4" containerName="dnsmasq-dns" Dec 06 03:57:51 crc kubenswrapper[4802]: I1206 03:57:51.399808 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="31be02e0-3b12-49e9-a900-3d069fa35ac4" containerName="dnsmasq-dns" Dec 06 03:57:51 crc kubenswrapper[4802]: I1206 03:57:51.400000 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="31be02e0-3b12-49e9-a900-3d069fa35ac4" containerName="dnsmasq-dns" Dec 06 03:57:51 crc kubenswrapper[4802]: I1206 03:57:51.400727 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-sbt8p-config-vn5rk" Dec 06 03:57:51 crc kubenswrapper[4802]: I1206 03:57:51.402792 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 06 03:57:51 crc kubenswrapper[4802]: I1206 03:57:51.415307 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-sbt8p-config-vn5rk"] Dec 06 03:57:51 crc kubenswrapper[4802]: I1206 03:57:51.565920 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6e64e900-ec79-4264-a56f-200517be05e5-var-run-ovn\") pod \"ovn-controller-sbt8p-config-vn5rk\" (UID: \"6e64e900-ec79-4264-a56f-200517be05e5\") " pod="openstack/ovn-controller-sbt8p-config-vn5rk" Dec 06 03:57:51 crc kubenswrapper[4802]: I1206 03:57:51.565968 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6e64e900-ec79-4264-a56f-200517be05e5-var-log-ovn\") pod \"ovn-controller-sbt8p-config-vn5rk\" (UID: \"6e64e900-ec79-4264-a56f-200517be05e5\") " pod="openstack/ovn-controller-sbt8p-config-vn5rk" Dec 06 03:57:51 crc kubenswrapper[4802]: I1206 03:57:51.565996 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pknb\" (UniqueName: \"kubernetes.io/projected/6e64e900-ec79-4264-a56f-200517be05e5-kube-api-access-9pknb\") pod \"ovn-controller-sbt8p-config-vn5rk\" (UID: \"6e64e900-ec79-4264-a56f-200517be05e5\") " pod="openstack/ovn-controller-sbt8p-config-vn5rk" Dec 06 03:57:51 crc kubenswrapper[4802]: I1206 03:57:51.566026 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6e64e900-ec79-4264-a56f-200517be05e5-scripts\") pod \"ovn-controller-sbt8p-config-vn5rk\" (UID: \"6e64e900-ec79-4264-a56f-200517be05e5\") " pod="openstack/ovn-controller-sbt8p-config-vn5rk" Dec 06 03:57:51 crc kubenswrapper[4802]: I1206 03:57:51.566083 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6e64e900-ec79-4264-a56f-200517be05e5-additional-scripts\") pod \"ovn-controller-sbt8p-config-vn5rk\" (UID: \"6e64e900-ec79-4264-a56f-200517be05e5\") " pod="openstack/ovn-controller-sbt8p-config-vn5rk" Dec 06 03:57:51 crc kubenswrapper[4802]: I1206 03:57:51.566105 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6e64e900-ec79-4264-a56f-200517be05e5-var-run\") pod \"ovn-controller-sbt8p-config-vn5rk\" (UID: \"6e64e900-ec79-4264-a56f-200517be05e5\") " pod="openstack/ovn-controller-sbt8p-config-vn5rk" Dec 06 03:57:51 crc kubenswrapper[4802]: I1206 03:57:51.667430 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6e64e900-ec79-4264-a56f-200517be05e5-scripts\") pod \"ovn-controller-sbt8p-config-vn5rk\" (UID: \"6e64e900-ec79-4264-a56f-200517be05e5\") " pod="openstack/ovn-controller-sbt8p-config-vn5rk" Dec 06 03:57:51 crc kubenswrapper[4802]: I1206 03:57:51.667553 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6e64e900-ec79-4264-a56f-200517be05e5-additional-scripts\") pod \"ovn-controller-sbt8p-config-vn5rk\" (UID: \"6e64e900-ec79-4264-a56f-200517be05e5\") " pod="openstack/ovn-controller-sbt8p-config-vn5rk" Dec 06 03:57:51 crc kubenswrapper[4802]: I1206 03:57:51.667590 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6e64e900-ec79-4264-a56f-200517be05e5-var-run\") pod \"ovn-controller-sbt8p-config-vn5rk\" (UID: \"6e64e900-ec79-4264-a56f-200517be05e5\") " pod="openstack/ovn-controller-sbt8p-config-vn5rk" Dec 06 03:57:51 crc kubenswrapper[4802]: I1206 03:57:51.667693 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6e64e900-ec79-4264-a56f-200517be05e5-var-run-ovn\") pod \"ovn-controller-sbt8p-config-vn5rk\" (UID: \"6e64e900-ec79-4264-a56f-200517be05e5\") " pod="openstack/ovn-controller-sbt8p-config-vn5rk" Dec 06 03:57:51 crc kubenswrapper[4802]: I1206 03:57:51.667719 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6e64e900-ec79-4264-a56f-200517be05e5-var-log-ovn\") pod \"ovn-controller-sbt8p-config-vn5rk\" (UID: \"6e64e900-ec79-4264-a56f-200517be05e5\") " pod="openstack/ovn-controller-sbt8p-config-vn5rk" Dec 06 03:57:51 crc kubenswrapper[4802]: I1206 03:57:51.667742 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pknb\" (UniqueName: \"kubernetes.io/projected/6e64e900-ec79-4264-a56f-200517be05e5-kube-api-access-9pknb\") pod \"ovn-controller-sbt8p-config-vn5rk\" (UID: \"6e64e900-ec79-4264-a56f-200517be05e5\") " pod="openstack/ovn-controller-sbt8p-config-vn5rk" Dec 06 03:57:51 crc kubenswrapper[4802]: I1206 03:57:51.668583 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6e64e900-ec79-4264-a56f-200517be05e5-var-run-ovn\") pod \"ovn-controller-sbt8p-config-vn5rk\" (UID: \"6e64e900-ec79-4264-a56f-200517be05e5\") " pod="openstack/ovn-controller-sbt8p-config-vn5rk" Dec 06 03:57:51 crc kubenswrapper[4802]: I1206 03:57:51.668656 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6e64e900-ec79-4264-a56f-200517be05e5-var-run\") pod \"ovn-controller-sbt8p-config-vn5rk\" (UID: \"6e64e900-ec79-4264-a56f-200517be05e5\") " pod="openstack/ovn-controller-sbt8p-config-vn5rk" Dec 06 03:57:51 crc kubenswrapper[4802]: I1206 03:57:51.668691 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6e64e900-ec79-4264-a56f-200517be05e5-var-log-ovn\") pod \"ovn-controller-sbt8p-config-vn5rk\" (UID: \"6e64e900-ec79-4264-a56f-200517be05e5\") " pod="openstack/ovn-controller-sbt8p-config-vn5rk" Dec 06 03:57:51 crc kubenswrapper[4802]: I1206 03:57:51.668820 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6e64e900-ec79-4264-a56f-200517be05e5-additional-scripts\") pod \"ovn-controller-sbt8p-config-vn5rk\" (UID: \"6e64e900-ec79-4264-a56f-200517be05e5\") " pod="openstack/ovn-controller-sbt8p-config-vn5rk" Dec 06 03:57:51 crc kubenswrapper[4802]: I1206 03:57:51.669973 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6e64e900-ec79-4264-a56f-200517be05e5-scripts\") pod \"ovn-controller-sbt8p-config-vn5rk\" (UID: \"6e64e900-ec79-4264-a56f-200517be05e5\") " pod="openstack/ovn-controller-sbt8p-config-vn5rk" Dec 06 03:57:51 crc kubenswrapper[4802]: I1206 03:57:51.689283 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pknb\" (UniqueName: \"kubernetes.io/projected/6e64e900-ec79-4264-a56f-200517be05e5-kube-api-access-9pknb\") pod \"ovn-controller-sbt8p-config-vn5rk\" (UID: \"6e64e900-ec79-4264-a56f-200517be05e5\") " pod="openstack/ovn-controller-sbt8p-config-vn5rk" Dec 06 03:57:51 crc kubenswrapper[4802]: I1206 03:57:51.738431 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-sbt8p-config-vn5rk" Dec 06 03:57:55 crc kubenswrapper[4802]: I1206 03:57:55.010731 4802 generic.go:334] "Generic (PLEG): container finished" podID="de4cab50-e92b-4282-bd9d-1b9439373aad" containerID="cc3dcd6dd4b27d5dec3c150b5e3848e004011d94e8ad36cea99a4b8fa28d4c96" exitCode=0 Dec 06 03:57:55 crc kubenswrapper[4802]: I1206 03:57:55.011222 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-flhmv" event={"ID":"de4cab50-e92b-4282-bd9d-1b9439373aad","Type":"ContainerDied","Data":"cc3dcd6dd4b27d5dec3c150b5e3848e004011d94e8ad36cea99a4b8fa28d4c96"} Dec 06 03:57:56 crc kubenswrapper[4802]: I1206 03:57:56.165373 4802 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-sbt8p" podUID="6e785e69-07cd-4c5c-8446-3d2a0d7e7b0b" containerName="ovn-controller" probeResult="failure" output=< Dec 06 03:57:56 crc kubenswrapper[4802]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 06 03:57:56 crc kubenswrapper[4802]: > Dec 06 03:57:57 crc kubenswrapper[4802]: I1206 03:57:57.958122 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-flhmv" Dec 06 03:57:57 crc kubenswrapper[4802]: I1206 03:57:57.996472 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fp2sw\" (UniqueName: \"kubernetes.io/projected/de4cab50-e92b-4282-bd9d-1b9439373aad-kube-api-access-fp2sw\") pod \"de4cab50-e92b-4282-bd9d-1b9439373aad\" (UID: \"de4cab50-e92b-4282-bd9d-1b9439373aad\") " Dec 06 03:57:58 crc kubenswrapper[4802]: I1206 03:57:58.010352 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de4cab50-e92b-4282-bd9d-1b9439373aad-kube-api-access-fp2sw" (OuterVolumeSpecName: "kube-api-access-fp2sw") pod "de4cab50-e92b-4282-bd9d-1b9439373aad" (UID: "de4cab50-e92b-4282-bd9d-1b9439373aad"). InnerVolumeSpecName "kube-api-access-fp2sw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:57:58 crc kubenswrapper[4802]: I1206 03:57:58.051844 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-flhmv" event={"ID":"de4cab50-e92b-4282-bd9d-1b9439373aad","Type":"ContainerDied","Data":"fb63b9f045d40054ecdcfe61fb7fa7ab4384160fdc80f6eaf64870bffd41175e"} Dec 06 03:57:58 crc kubenswrapper[4802]: I1206 03:57:58.051904 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fb63b9f045d40054ecdcfe61fb7fa7ab4384160fdc80f6eaf64870bffd41175e" Dec 06 03:57:58 crc kubenswrapper[4802]: I1206 03:57:58.051961 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-flhmv" Dec 06 03:57:58 crc kubenswrapper[4802]: I1206 03:57:58.067170 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-sbt8p-config-vn5rk"] Dec 06 03:57:58 crc kubenswrapper[4802]: W1206 03:57:58.078054 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6e64e900_ec79_4264_a56f_200517be05e5.slice/crio-2aff2cd5190e2f4dcb498d70816e796430fdd5b3f160ee5cd440a1b65a0cd05b WatchSource:0}: Error finding container 2aff2cd5190e2f4dcb498d70816e796430fdd5b3f160ee5cd440a1b65a0cd05b: Status 404 returned error can't find the container with id 2aff2cd5190e2f4dcb498d70816e796430fdd5b3f160ee5cd440a1b65a0cd05b Dec 06 03:57:58 crc kubenswrapper[4802]: I1206 03:57:58.097595 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/de4cab50-e92b-4282-bd9d-1b9439373aad-dispersionconf\") pod \"de4cab50-e92b-4282-bd9d-1b9439373aad\" (UID: \"de4cab50-e92b-4282-bd9d-1b9439373aad\") " Dec 06 03:57:58 crc kubenswrapper[4802]: I1206 03:57:58.097680 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/de4cab50-e92b-4282-bd9d-1b9439373aad-scripts\") pod \"de4cab50-e92b-4282-bd9d-1b9439373aad\" (UID: \"de4cab50-e92b-4282-bd9d-1b9439373aad\") " Dec 06 03:57:58 crc kubenswrapper[4802]: I1206 03:57:58.097731 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de4cab50-e92b-4282-bd9d-1b9439373aad-combined-ca-bundle\") pod \"de4cab50-e92b-4282-bd9d-1b9439373aad\" (UID: \"de4cab50-e92b-4282-bd9d-1b9439373aad\") " Dec 06 03:57:58 crc kubenswrapper[4802]: I1206 03:57:58.097892 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/de4cab50-e92b-4282-bd9d-1b9439373aad-ring-data-devices\") pod \"de4cab50-e92b-4282-bd9d-1b9439373aad\" (UID: \"de4cab50-e92b-4282-bd9d-1b9439373aad\") " Dec 06 03:57:58 crc kubenswrapper[4802]: I1206 03:57:58.097918 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/de4cab50-e92b-4282-bd9d-1b9439373aad-swiftconf\") pod \"de4cab50-e92b-4282-bd9d-1b9439373aad\" (UID: \"de4cab50-e92b-4282-bd9d-1b9439373aad\") " Dec 06 03:57:58 crc kubenswrapper[4802]: I1206 03:57:58.097937 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/de4cab50-e92b-4282-bd9d-1b9439373aad-etc-swift\") pod \"de4cab50-e92b-4282-bd9d-1b9439373aad\" (UID: \"de4cab50-e92b-4282-bd9d-1b9439373aad\") " Dec 06 03:57:58 crc kubenswrapper[4802]: I1206 03:57:58.098366 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fp2sw\" (UniqueName: \"kubernetes.io/projected/de4cab50-e92b-4282-bd9d-1b9439373aad-kube-api-access-fp2sw\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:58 crc kubenswrapper[4802]: I1206 03:57:58.098642 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de4cab50-e92b-4282-bd9d-1b9439373aad-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "de4cab50-e92b-4282-bd9d-1b9439373aad" (UID: "de4cab50-e92b-4282-bd9d-1b9439373aad"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:57:58 crc kubenswrapper[4802]: I1206 03:57:58.099445 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de4cab50-e92b-4282-bd9d-1b9439373aad-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "de4cab50-e92b-4282-bd9d-1b9439373aad" (UID: "de4cab50-e92b-4282-bd9d-1b9439373aad"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:57:58 crc kubenswrapper[4802]: I1206 03:57:58.103723 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de4cab50-e92b-4282-bd9d-1b9439373aad-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "de4cab50-e92b-4282-bd9d-1b9439373aad" (UID: "de4cab50-e92b-4282-bd9d-1b9439373aad"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:57:58 crc kubenswrapper[4802]: I1206 03:57:58.115704 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de4cab50-e92b-4282-bd9d-1b9439373aad-scripts" (OuterVolumeSpecName: "scripts") pod "de4cab50-e92b-4282-bd9d-1b9439373aad" (UID: "de4cab50-e92b-4282-bd9d-1b9439373aad"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:57:58 crc kubenswrapper[4802]: I1206 03:57:58.117602 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de4cab50-e92b-4282-bd9d-1b9439373aad-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "de4cab50-e92b-4282-bd9d-1b9439373aad" (UID: "de4cab50-e92b-4282-bd9d-1b9439373aad"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:57:58 crc kubenswrapper[4802]: I1206 03:57:58.119393 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de4cab50-e92b-4282-bd9d-1b9439373aad-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "de4cab50-e92b-4282-bd9d-1b9439373aad" (UID: "de4cab50-e92b-4282-bd9d-1b9439373aad"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:57:58 crc kubenswrapper[4802]: I1206 03:57:58.199703 4802 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de4cab50-e92b-4282-bd9d-1b9439373aad-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:58 crc kubenswrapper[4802]: I1206 03:57:58.199740 4802 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/de4cab50-e92b-4282-bd9d-1b9439373aad-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:58 crc kubenswrapper[4802]: I1206 03:57:58.199761 4802 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/de4cab50-e92b-4282-bd9d-1b9439373aad-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:58 crc kubenswrapper[4802]: I1206 03:57:58.199771 4802 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/de4cab50-e92b-4282-bd9d-1b9439373aad-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:58 crc kubenswrapper[4802]: I1206 03:57:58.199780 4802 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/de4cab50-e92b-4282-bd9d-1b9439373aad-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:58 crc kubenswrapper[4802]: I1206 03:57:58.199788 4802 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/de4cab50-e92b-4282-bd9d-1b9439373aad-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:59 crc kubenswrapper[4802]: I1206 03:57:59.065550 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7b4f22a1-41dc-4a75-8999-53df29d2fb5d","Type":"ContainerStarted","Data":"ed17f5b81a54486ceb0680c299b5f84abc02bd1e687d265819e37dc8b4570d6b"} Dec 06 03:57:59 crc kubenswrapper[4802]: I1206 03:57:59.066433 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 06 03:57:59 crc kubenswrapper[4802]: I1206 03:57:59.068702 4802 generic.go:334] "Generic (PLEG): container finished" podID="6e64e900-ec79-4264-a56f-200517be05e5" containerID="bbc75983c8257f385dfab8690e21029a136f6c9ea7a2b913140132166d7871b0" exitCode=0 Dec 06 03:57:59 crc kubenswrapper[4802]: I1206 03:57:59.068796 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-sbt8p-config-vn5rk" event={"ID":"6e64e900-ec79-4264-a56f-200517be05e5","Type":"ContainerDied","Data":"bbc75983c8257f385dfab8690e21029a136f6c9ea7a2b913140132166d7871b0"} Dec 06 03:57:59 crc kubenswrapper[4802]: I1206 03:57:59.068817 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-sbt8p-config-vn5rk" event={"ID":"6e64e900-ec79-4264-a56f-200517be05e5","Type":"ContainerStarted","Data":"2aff2cd5190e2f4dcb498d70816e796430fdd5b3f160ee5cd440a1b65a0cd05b"} Dec 06 03:57:59 crc kubenswrapper[4802]: I1206 03:57:59.077359 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8c8b6c49-6f71-403d-b639-4a17ae3bc802","Type":"ContainerStarted","Data":"098aed31247101a27a8ff5c8dd2bfd3731dd532b7504c6cf5431f78994869852"} Dec 06 03:57:59 crc kubenswrapper[4802]: I1206 03:57:59.078042 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 06 03:57:59 crc kubenswrapper[4802]: I1206 03:57:59.094968 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=46.263392916 podStartE2EDuration="1m29.094947667s" podCreationTimestamp="2025-12-06 03:56:30 +0000 UTC" firstStartedPulling="2025-12-06 03:56:33.023284802 +0000 UTC m=+985.895193954" lastFinishedPulling="2025-12-06 03:57:15.854839543 +0000 UTC m=+1028.726748705" observedRunningTime="2025-12-06 03:57:59.091001319 +0000 UTC m=+1071.962910471" watchObservedRunningTime="2025-12-06 03:57:59.094947667 +0000 UTC m=+1071.966856819" Dec 06 03:57:59 crc kubenswrapper[4802]: I1206 03:57:59.121982 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=45.540329354 podStartE2EDuration="1m28.121961355s" podCreationTimestamp="2025-12-06 03:56:31 +0000 UTC" firstStartedPulling="2025-12-06 03:56:33.2669304 +0000 UTC m=+986.138839552" lastFinishedPulling="2025-12-06 03:57:15.848562401 +0000 UTC m=+1028.720471553" observedRunningTime="2025-12-06 03:57:59.117943516 +0000 UTC m=+1071.989852658" watchObservedRunningTime="2025-12-06 03:57:59.121961355 +0000 UTC m=+1071.993870507" Dec 06 03:58:00 crc kubenswrapper[4802]: I1206 03:58:00.087478 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-mlhfp" event={"ID":"5d947661-37cd-49c2-bd89-e927e4034ba4","Type":"ContainerStarted","Data":"46e6ea4ac88dfb898ea9b663b1a6d76d786318b1afea471fa0d10200b8a30b87"} Dec 06 03:58:00 crc kubenswrapper[4802]: I1206 03:58:00.109297 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-mlhfp" podStartSLOduration=2.203889683 podStartE2EDuration="29.109279544s" podCreationTimestamp="2025-12-06 03:57:31 +0000 UTC" firstStartedPulling="2025-12-06 03:57:32.073518805 +0000 UTC m=+1044.945427957" lastFinishedPulling="2025-12-06 03:57:58.978908656 +0000 UTC m=+1071.850817818" observedRunningTime="2025-12-06 03:58:00.108186694 +0000 UTC m=+1072.980095846" watchObservedRunningTime="2025-12-06 03:58:00.109279544 +0000 UTC m=+1072.981188696" Dec 06 03:58:00 crc kubenswrapper[4802]: I1206 03:58:00.233396 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/508d9d64-dd16-4d21-b492-052d123f1a6a-etc-swift\") pod \"swift-storage-0\" (UID: \"508d9d64-dd16-4d21-b492-052d123f1a6a\") " pod="openstack/swift-storage-0" Dec 06 03:58:00 crc kubenswrapper[4802]: I1206 03:58:00.239005 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/508d9d64-dd16-4d21-b492-052d123f1a6a-etc-swift\") pod \"swift-storage-0\" (UID: \"508d9d64-dd16-4d21-b492-052d123f1a6a\") " pod="openstack/swift-storage-0" Dec 06 03:58:00 crc kubenswrapper[4802]: I1206 03:58:00.425201 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-sbt8p-config-vn5rk" Dec 06 03:58:00 crc kubenswrapper[4802]: I1206 03:58:00.501741 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 06 03:58:00 crc kubenswrapper[4802]: I1206 03:58:00.539501 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6e64e900-ec79-4264-a56f-200517be05e5-var-run\") pod \"6e64e900-ec79-4264-a56f-200517be05e5\" (UID: \"6e64e900-ec79-4264-a56f-200517be05e5\") " Dec 06 03:58:00 crc kubenswrapper[4802]: I1206 03:58:00.539602 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6e64e900-ec79-4264-a56f-200517be05e5-var-run" (OuterVolumeSpecName: "var-run") pod "6e64e900-ec79-4264-a56f-200517be05e5" (UID: "6e64e900-ec79-4264-a56f-200517be05e5"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:58:00 crc kubenswrapper[4802]: I1206 03:58:00.539616 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6e64e900-ec79-4264-a56f-200517be05e5-scripts\") pod \"6e64e900-ec79-4264-a56f-200517be05e5\" (UID: \"6e64e900-ec79-4264-a56f-200517be05e5\") " Dec 06 03:58:00 crc kubenswrapper[4802]: I1206 03:58:00.539762 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6e64e900-ec79-4264-a56f-200517be05e5-var-log-ovn\") pod \"6e64e900-ec79-4264-a56f-200517be05e5\" (UID: \"6e64e900-ec79-4264-a56f-200517be05e5\") " Dec 06 03:58:00 crc kubenswrapper[4802]: I1206 03:58:00.539810 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6e64e900-ec79-4264-a56f-200517be05e5-additional-scripts\") pod \"6e64e900-ec79-4264-a56f-200517be05e5\" (UID: \"6e64e900-ec79-4264-a56f-200517be05e5\") " Dec 06 03:58:00 crc kubenswrapper[4802]: I1206 03:58:00.539848 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6e64e900-ec79-4264-a56f-200517be05e5-var-run-ovn\") pod \"6e64e900-ec79-4264-a56f-200517be05e5\" (UID: \"6e64e900-ec79-4264-a56f-200517be05e5\") " Dec 06 03:58:00 crc kubenswrapper[4802]: I1206 03:58:00.539869 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9pknb\" (UniqueName: \"kubernetes.io/projected/6e64e900-ec79-4264-a56f-200517be05e5-kube-api-access-9pknb\") pod \"6e64e900-ec79-4264-a56f-200517be05e5\" (UID: \"6e64e900-ec79-4264-a56f-200517be05e5\") " Dec 06 03:58:00 crc kubenswrapper[4802]: I1206 03:58:00.539901 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6e64e900-ec79-4264-a56f-200517be05e5-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "6e64e900-ec79-4264-a56f-200517be05e5" (UID: "6e64e900-ec79-4264-a56f-200517be05e5"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:58:00 crc kubenswrapper[4802]: I1206 03:58:00.539959 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6e64e900-ec79-4264-a56f-200517be05e5-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "6e64e900-ec79-4264-a56f-200517be05e5" (UID: "6e64e900-ec79-4264-a56f-200517be05e5"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:58:00 crc kubenswrapper[4802]: I1206 03:58:00.540370 4802 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6e64e900-ec79-4264-a56f-200517be05e5-var-run\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:00 crc kubenswrapper[4802]: I1206 03:58:00.540390 4802 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6e64e900-ec79-4264-a56f-200517be05e5-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:00 crc kubenswrapper[4802]: I1206 03:58:00.540399 4802 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6e64e900-ec79-4264-a56f-200517be05e5-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:00 crc kubenswrapper[4802]: I1206 03:58:00.540594 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e64e900-ec79-4264-a56f-200517be05e5-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "6e64e900-ec79-4264-a56f-200517be05e5" (UID: "6e64e900-ec79-4264-a56f-200517be05e5"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:58:00 crc kubenswrapper[4802]: I1206 03:58:00.541176 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e64e900-ec79-4264-a56f-200517be05e5-scripts" (OuterVolumeSpecName: "scripts") pod "6e64e900-ec79-4264-a56f-200517be05e5" (UID: "6e64e900-ec79-4264-a56f-200517be05e5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:58:00 crc kubenswrapper[4802]: I1206 03:58:00.549172 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e64e900-ec79-4264-a56f-200517be05e5-kube-api-access-9pknb" (OuterVolumeSpecName: "kube-api-access-9pknb") pod "6e64e900-ec79-4264-a56f-200517be05e5" (UID: "6e64e900-ec79-4264-a56f-200517be05e5"). InnerVolumeSpecName "kube-api-access-9pknb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:58:00 crc kubenswrapper[4802]: I1206 03:58:00.642691 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9pknb\" (UniqueName: \"kubernetes.io/projected/6e64e900-ec79-4264-a56f-200517be05e5-kube-api-access-9pknb\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:00 crc kubenswrapper[4802]: I1206 03:58:00.643043 4802 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6e64e900-ec79-4264-a56f-200517be05e5-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:00 crc kubenswrapper[4802]: I1206 03:58:00.643053 4802 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6e64e900-ec79-4264-a56f-200517be05e5-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:01 crc kubenswrapper[4802]: I1206 03:58:01.058534 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 06 03:58:01 crc kubenswrapper[4802]: I1206 03:58:01.095508 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-sbt8p-config-vn5rk" event={"ID":"6e64e900-ec79-4264-a56f-200517be05e5","Type":"ContainerDied","Data":"2aff2cd5190e2f4dcb498d70816e796430fdd5b3f160ee5cd440a1b65a0cd05b"} Dec 06 03:58:01 crc kubenswrapper[4802]: I1206 03:58:01.095557 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2aff2cd5190e2f4dcb498d70816e796430fdd5b3f160ee5cd440a1b65a0cd05b" Dec 06 03:58:01 crc kubenswrapper[4802]: I1206 03:58:01.095611 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-sbt8p-config-vn5rk" Dec 06 03:58:01 crc kubenswrapper[4802]: I1206 03:58:01.102414 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"508d9d64-dd16-4d21-b492-052d123f1a6a","Type":"ContainerStarted","Data":"4a3246fb238395d08cb1914ae051dea79e9b99fa7f6279dc9e199c0e817c4ce5"} Dec 06 03:58:01 crc kubenswrapper[4802]: I1206 03:58:01.163909 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-sbt8p" Dec 06 03:58:01 crc kubenswrapper[4802]: I1206 03:58:01.527534 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-sbt8p-config-vn5rk"] Dec 06 03:58:01 crc kubenswrapper[4802]: I1206 03:58:01.534657 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-sbt8p-config-vn5rk"] Dec 06 03:58:03 crc kubenswrapper[4802]: I1206 03:58:03.463822 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e64e900-ec79-4264-a56f-200517be05e5" path="/var/lib/kubelet/pods/6e64e900-ec79-4264-a56f-200517be05e5/volumes" Dec 06 03:58:04 crc kubenswrapper[4802]: I1206 03:58:04.132177 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"508d9d64-dd16-4d21-b492-052d123f1a6a","Type":"ContainerStarted","Data":"5b6df9558f3f9eefd2bf3f41c7244072f75e84955567a8a443abf8ab87639d3f"} Dec 06 03:58:05 crc kubenswrapper[4802]: I1206 03:58:05.139548 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"508d9d64-dd16-4d21-b492-052d123f1a6a","Type":"ContainerStarted","Data":"395567e5eb48b1b486a7a80f30a39793cec863c82667f420634ca90f5fa211e3"} Dec 06 03:58:06 crc kubenswrapper[4802]: I1206 03:58:06.152384 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"508d9d64-dd16-4d21-b492-052d123f1a6a","Type":"ContainerStarted","Data":"32465b95f62896df6dec12678b007c544d793e9e2d94dcfce85544dbc27d1f9f"} Dec 06 03:58:06 crc kubenswrapper[4802]: I1206 03:58:06.152436 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"508d9d64-dd16-4d21-b492-052d123f1a6a","Type":"ContainerStarted","Data":"525150eed9b649c1af8d01d1b3d9f97acf249ea85067c395f7602135b3a214a4"} Dec 06 03:58:09 crc kubenswrapper[4802]: I1206 03:58:09.176592 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"508d9d64-dd16-4d21-b492-052d123f1a6a","Type":"ContainerStarted","Data":"ff9440eef3fdfacda816e67b78255184192a7b53385974cfb923481dee672753"} Dec 06 03:58:10 crc kubenswrapper[4802]: I1206 03:58:10.188077 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"508d9d64-dd16-4d21-b492-052d123f1a6a","Type":"ContainerStarted","Data":"61e4235bedcd0154ae7e65826a0085e1fbfd1ed224a2938c7bedaee5f0c91677"} Dec 06 03:58:10 crc kubenswrapper[4802]: I1206 03:58:10.189007 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"508d9d64-dd16-4d21-b492-052d123f1a6a","Type":"ContainerStarted","Data":"ac518867049369a1c96b8c342110e1e0bb878cb70714603d8ebeca388ad82412"} Dec 06 03:58:10 crc kubenswrapper[4802]: I1206 03:58:10.189129 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"508d9d64-dd16-4d21-b492-052d123f1a6a","Type":"ContainerStarted","Data":"8028d6711925d625b8c5a2f9d697b7f446f811267314836bf1249c08c849e834"} Dec 06 03:58:12 crc kubenswrapper[4802]: I1206 03:58:12.214777 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"508d9d64-dd16-4d21-b492-052d123f1a6a","Type":"ContainerStarted","Data":"80a30636c3f057da4f2dc776c3e3b6b6df804e1b5c28747a3aa743f3c2d0641c"} Dec 06 03:58:12 crc kubenswrapper[4802]: I1206 03:58:12.391922 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 06 03:58:12 crc kubenswrapper[4802]: I1206 03:58:12.452609 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 06 03:58:12 crc kubenswrapper[4802]: I1206 03:58:12.723024 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-v5t59"] Dec 06 03:58:12 crc kubenswrapper[4802]: E1206 03:58:12.723299 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e64e900-ec79-4264-a56f-200517be05e5" containerName="ovn-config" Dec 06 03:58:12 crc kubenswrapper[4802]: I1206 03:58:12.723314 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e64e900-ec79-4264-a56f-200517be05e5" containerName="ovn-config" Dec 06 03:58:12 crc kubenswrapper[4802]: E1206 03:58:12.723330 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de4cab50-e92b-4282-bd9d-1b9439373aad" containerName="swift-ring-rebalance" Dec 06 03:58:12 crc kubenswrapper[4802]: I1206 03:58:12.723336 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="de4cab50-e92b-4282-bd9d-1b9439373aad" containerName="swift-ring-rebalance" Dec 06 03:58:12 crc kubenswrapper[4802]: I1206 03:58:12.723514 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="de4cab50-e92b-4282-bd9d-1b9439373aad" containerName="swift-ring-rebalance" Dec 06 03:58:12 crc kubenswrapper[4802]: I1206 03:58:12.723551 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e64e900-ec79-4264-a56f-200517be05e5" containerName="ovn-config" Dec 06 03:58:12 crc kubenswrapper[4802]: I1206 03:58:12.724019 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-v5t59" Dec 06 03:58:12 crc kubenswrapper[4802]: I1206 03:58:12.739293 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-v5t59"] Dec 06 03:58:12 crc kubenswrapper[4802]: I1206 03:58:12.813814 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-nv6kh"] Dec 06 03:58:12 crc kubenswrapper[4802]: I1206 03:58:12.814836 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-nv6kh" Dec 06 03:58:12 crc kubenswrapper[4802]: I1206 03:58:12.828146 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-ec26-account-create-update-fxz49"] Dec 06 03:58:12 crc kubenswrapper[4802]: I1206 03:58:12.830082 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-ec26-account-create-update-fxz49" Dec 06 03:58:12 crc kubenswrapper[4802]: I1206 03:58:12.834249 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-db-secret" Dec 06 03:58:12 crc kubenswrapper[4802]: I1206 03:58:12.842923 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-nv6kh"] Dec 06 03:58:12 crc kubenswrapper[4802]: I1206 03:58:12.865230 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ebeea130-3942-475e-ba17-51d624cde585-operator-scripts\") pod \"barbican-db-create-nv6kh\" (UID: \"ebeea130-3942-475e-ba17-51d624cde585\") " pod="openstack/barbican-db-create-nv6kh" Dec 06 03:58:12 crc kubenswrapper[4802]: I1206 03:58:12.865378 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c7d36678-2b9e-4380-8828-a32e159ce604-operator-scripts\") pod \"cinder-db-create-v5t59\" (UID: \"c7d36678-2b9e-4380-8828-a32e159ce604\") " pod="openstack/cinder-db-create-v5t59" Dec 06 03:58:12 crc kubenswrapper[4802]: I1206 03:58:12.865424 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jchl\" (UniqueName: \"kubernetes.io/projected/c7d36678-2b9e-4380-8828-a32e159ce604-kube-api-access-5jchl\") pod \"cinder-db-create-v5t59\" (UID: \"c7d36678-2b9e-4380-8828-a32e159ce604\") " pod="openstack/cinder-db-create-v5t59" Dec 06 03:58:12 crc kubenswrapper[4802]: I1206 03:58:12.865606 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fssrm\" (UniqueName: \"kubernetes.io/projected/ebeea130-3942-475e-ba17-51d624cde585-kube-api-access-fssrm\") pod \"barbican-db-create-nv6kh\" (UID: \"ebeea130-3942-475e-ba17-51d624cde585\") " pod="openstack/barbican-db-create-nv6kh" Dec 06 03:58:12 crc kubenswrapper[4802]: I1206 03:58:12.871460 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-ec26-account-create-update-fxz49"] Dec 06 03:58:12 crc kubenswrapper[4802]: I1206 03:58:12.942653 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-1352-account-create-update-jhtbv"] Dec 06 03:58:12 crc kubenswrapper[4802]: I1206 03:58:12.943701 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-1352-account-create-update-jhtbv" Dec 06 03:58:12 crc kubenswrapper[4802]: I1206 03:58:12.950137 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 06 03:58:12 crc kubenswrapper[4802]: I1206 03:58:12.967204 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c4bfa7b-7d20-4c15-b4a7-181a05dcaab4-operator-scripts\") pod \"heat-ec26-account-create-update-fxz49\" (UID: \"0c4bfa7b-7d20-4c15-b4a7-181a05dcaab4\") " pod="openstack/heat-ec26-account-create-update-fxz49" Dec 06 03:58:12 crc kubenswrapper[4802]: I1206 03:58:12.967280 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ebeea130-3942-475e-ba17-51d624cde585-operator-scripts\") pod \"barbican-db-create-nv6kh\" (UID: \"ebeea130-3942-475e-ba17-51d624cde585\") " pod="openstack/barbican-db-create-nv6kh" Dec 06 03:58:12 crc kubenswrapper[4802]: I1206 03:58:12.967315 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c7d36678-2b9e-4380-8828-a32e159ce604-operator-scripts\") pod \"cinder-db-create-v5t59\" (UID: \"c7d36678-2b9e-4380-8828-a32e159ce604\") " pod="openstack/cinder-db-create-v5t59" Dec 06 03:58:12 crc kubenswrapper[4802]: I1206 03:58:12.967333 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jchl\" (UniqueName: \"kubernetes.io/projected/c7d36678-2b9e-4380-8828-a32e159ce604-kube-api-access-5jchl\") pod \"cinder-db-create-v5t59\" (UID: \"c7d36678-2b9e-4380-8828-a32e159ce604\") " pod="openstack/cinder-db-create-v5t59" Dec 06 03:58:12 crc kubenswrapper[4802]: I1206 03:58:12.967387 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fssrm\" (UniqueName: \"kubernetes.io/projected/ebeea130-3942-475e-ba17-51d624cde585-kube-api-access-fssrm\") pod \"barbican-db-create-nv6kh\" (UID: \"ebeea130-3942-475e-ba17-51d624cde585\") " pod="openstack/barbican-db-create-nv6kh" Dec 06 03:58:12 crc kubenswrapper[4802]: I1206 03:58:12.967406 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrqc9\" (UniqueName: \"kubernetes.io/projected/0c4bfa7b-7d20-4c15-b4a7-181a05dcaab4-kube-api-access-qrqc9\") pod \"heat-ec26-account-create-update-fxz49\" (UID: \"0c4bfa7b-7d20-4c15-b4a7-181a05dcaab4\") " pod="openstack/heat-ec26-account-create-update-fxz49" Dec 06 03:58:12 crc kubenswrapper[4802]: I1206 03:58:12.968098 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ebeea130-3942-475e-ba17-51d624cde585-operator-scripts\") pod \"barbican-db-create-nv6kh\" (UID: \"ebeea130-3942-475e-ba17-51d624cde585\") " pod="openstack/barbican-db-create-nv6kh" Dec 06 03:58:12 crc kubenswrapper[4802]: I1206 03:58:12.968630 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c7d36678-2b9e-4380-8828-a32e159ce604-operator-scripts\") pod \"cinder-db-create-v5t59\" (UID: \"c7d36678-2b9e-4380-8828-a32e159ce604\") " pod="openstack/cinder-db-create-v5t59" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.012908 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-1352-account-create-update-jhtbv"] Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.018460 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fssrm\" (UniqueName: \"kubernetes.io/projected/ebeea130-3942-475e-ba17-51d624cde585-kube-api-access-fssrm\") pod \"barbican-db-create-nv6kh\" (UID: \"ebeea130-3942-475e-ba17-51d624cde585\") " pod="openstack/barbican-db-create-nv6kh" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.020461 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jchl\" (UniqueName: \"kubernetes.io/projected/c7d36678-2b9e-4380-8828-a32e159ce604-kube-api-access-5jchl\") pod \"cinder-db-create-v5t59\" (UID: \"c7d36678-2b9e-4380-8828-a32e159ce604\") " pod="openstack/cinder-db-create-v5t59" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.029879 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-create-gdjlc"] Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.031274 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-gdjlc" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.034575 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-e48e-account-create-update-zw8gt"] Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.035546 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-e48e-account-create-update-zw8gt" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.036631 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.040036 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-v5t59" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.070100 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrqc9\" (UniqueName: \"kubernetes.io/projected/0c4bfa7b-7d20-4c15-b4a7-181a05dcaab4-kube-api-access-qrqc9\") pod \"heat-ec26-account-create-update-fxz49\" (UID: \"0c4bfa7b-7d20-4c15-b4a7-181a05dcaab4\") " pod="openstack/heat-ec26-account-create-update-fxz49" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.070145 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7653cfc2-6e34-40f2-843a-9a644165b0fb-operator-scripts\") pod \"cinder-1352-account-create-update-jhtbv\" (UID: \"7653cfc2-6e34-40f2-843a-9a644165b0fb\") " pod="openstack/cinder-1352-account-create-update-jhtbv" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.070188 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c4bfa7b-7d20-4c15-b4a7-181a05dcaab4-operator-scripts\") pod \"heat-ec26-account-create-update-fxz49\" (UID: \"0c4bfa7b-7d20-4c15-b4a7-181a05dcaab4\") " pod="openstack/heat-ec26-account-create-update-fxz49" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.070244 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tbnch\" (UniqueName: \"kubernetes.io/projected/7653cfc2-6e34-40f2-843a-9a644165b0fb-kube-api-access-tbnch\") pod \"cinder-1352-account-create-update-jhtbv\" (UID: \"7653cfc2-6e34-40f2-843a-9a644165b0fb\") " pod="openstack/cinder-1352-account-create-update-jhtbv" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.072158 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c4bfa7b-7d20-4c15-b4a7-181a05dcaab4-operator-scripts\") pod \"heat-ec26-account-create-update-fxz49\" (UID: \"0c4bfa7b-7d20-4c15-b4a7-181a05dcaab4\") " pod="openstack/heat-ec26-account-create-update-fxz49" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.088582 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-e48e-account-create-update-zw8gt"] Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.104279 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrqc9\" (UniqueName: \"kubernetes.io/projected/0c4bfa7b-7d20-4c15-b4a7-181a05dcaab4-kube-api-access-qrqc9\") pod \"heat-ec26-account-create-update-fxz49\" (UID: \"0c4bfa7b-7d20-4c15-b4a7-181a05dcaab4\") " pod="openstack/heat-ec26-account-create-update-fxz49" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.114879 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-gdjlc"] Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.136594 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-mrxg9"] Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.137696 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-mrxg9" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.140423 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.140639 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.140798 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.142243 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-zk4kp" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.143088 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-mrxg9"] Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.162299 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-nv6kh" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.171721 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9lsf\" (UniqueName: \"kubernetes.io/projected/a076eb84-b8f5-4f75-922b-e44682ad8e67-kube-api-access-n9lsf\") pod \"barbican-e48e-account-create-update-zw8gt\" (UID: \"a076eb84-b8f5-4f75-922b-e44682ad8e67\") " pod="openstack/barbican-e48e-account-create-update-zw8gt" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.171849 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7653cfc2-6e34-40f2-843a-9a644165b0fb-operator-scripts\") pod \"cinder-1352-account-create-update-jhtbv\" (UID: \"7653cfc2-6e34-40f2-843a-9a644165b0fb\") " pod="openstack/cinder-1352-account-create-update-jhtbv" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.171893 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a076eb84-b8f5-4f75-922b-e44682ad8e67-operator-scripts\") pod \"barbican-e48e-account-create-update-zw8gt\" (UID: \"a076eb84-b8f5-4f75-922b-e44682ad8e67\") " pod="openstack/barbican-e48e-account-create-update-zw8gt" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.171917 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmtlr\" (UniqueName: \"kubernetes.io/projected/e065079d-0d3d-40b5-8103-556cb4b7f338-kube-api-access-fmtlr\") pod \"heat-db-create-gdjlc\" (UID: \"e065079d-0d3d-40b5-8103-556cb4b7f338\") " pod="openstack/heat-db-create-gdjlc" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.171953 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e065079d-0d3d-40b5-8103-556cb4b7f338-operator-scripts\") pod \"heat-db-create-gdjlc\" (UID: \"e065079d-0d3d-40b5-8103-556cb4b7f338\") " pod="openstack/heat-db-create-gdjlc" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.171982 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tbnch\" (UniqueName: \"kubernetes.io/projected/7653cfc2-6e34-40f2-843a-9a644165b0fb-kube-api-access-tbnch\") pod \"cinder-1352-account-create-update-jhtbv\" (UID: \"7653cfc2-6e34-40f2-843a-9a644165b0fb\") " pod="openstack/cinder-1352-account-create-update-jhtbv" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.172959 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7653cfc2-6e34-40f2-843a-9a644165b0fb-operator-scripts\") pod \"cinder-1352-account-create-update-jhtbv\" (UID: \"7653cfc2-6e34-40f2-843a-9a644165b0fb\") " pod="openstack/cinder-1352-account-create-update-jhtbv" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.175923 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-ec26-account-create-update-fxz49" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.201186 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tbnch\" (UniqueName: \"kubernetes.io/projected/7653cfc2-6e34-40f2-843a-9a644165b0fb-kube-api-access-tbnch\") pod \"cinder-1352-account-create-update-jhtbv\" (UID: \"7653cfc2-6e34-40f2-843a-9a644165b0fb\") " pod="openstack/cinder-1352-account-create-update-jhtbv" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.255128 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"508d9d64-dd16-4d21-b492-052d123f1a6a","Type":"ContainerStarted","Data":"573a2dbfa7201fff4b3df2fe7bcfa6103883006bc5c566f26bb53772d00ca688"} Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.273487 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e065079d-0d3d-40b5-8103-556cb4b7f338-operator-scripts\") pod \"heat-db-create-gdjlc\" (UID: \"e065079d-0d3d-40b5-8103-556cb4b7f338\") " pod="openstack/heat-db-create-gdjlc" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.273538 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6f5b891-75b4-4b8c-95b1-7b92ee267793-combined-ca-bundle\") pod \"keystone-db-sync-mrxg9\" (UID: \"c6f5b891-75b4-4b8c-95b1-7b92ee267793\") " pod="openstack/keystone-db-sync-mrxg9" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.273598 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5m9g\" (UniqueName: \"kubernetes.io/projected/c6f5b891-75b4-4b8c-95b1-7b92ee267793-kube-api-access-x5m9g\") pod \"keystone-db-sync-mrxg9\" (UID: \"c6f5b891-75b4-4b8c-95b1-7b92ee267793\") " pod="openstack/keystone-db-sync-mrxg9" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.273623 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9lsf\" (UniqueName: \"kubernetes.io/projected/a076eb84-b8f5-4f75-922b-e44682ad8e67-kube-api-access-n9lsf\") pod \"barbican-e48e-account-create-update-zw8gt\" (UID: \"a076eb84-b8f5-4f75-922b-e44682ad8e67\") " pod="openstack/barbican-e48e-account-create-update-zw8gt" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.273674 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6f5b891-75b4-4b8c-95b1-7b92ee267793-config-data\") pod \"keystone-db-sync-mrxg9\" (UID: \"c6f5b891-75b4-4b8c-95b1-7b92ee267793\") " pod="openstack/keystone-db-sync-mrxg9" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.273707 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a076eb84-b8f5-4f75-922b-e44682ad8e67-operator-scripts\") pod \"barbican-e48e-account-create-update-zw8gt\" (UID: \"a076eb84-b8f5-4f75-922b-e44682ad8e67\") " pod="openstack/barbican-e48e-account-create-update-zw8gt" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.273729 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmtlr\" (UniqueName: \"kubernetes.io/projected/e065079d-0d3d-40b5-8103-556cb4b7f338-kube-api-access-fmtlr\") pod \"heat-db-create-gdjlc\" (UID: \"e065079d-0d3d-40b5-8103-556cb4b7f338\") " pod="openstack/heat-db-create-gdjlc" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.274594 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e065079d-0d3d-40b5-8103-556cb4b7f338-operator-scripts\") pod \"heat-db-create-gdjlc\" (UID: \"e065079d-0d3d-40b5-8103-556cb4b7f338\") " pod="openstack/heat-db-create-gdjlc" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.274973 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a076eb84-b8f5-4f75-922b-e44682ad8e67-operator-scripts\") pod \"barbican-e48e-account-create-update-zw8gt\" (UID: \"a076eb84-b8f5-4f75-922b-e44682ad8e67\") " pod="openstack/barbican-e48e-account-create-update-zw8gt" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.296441 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmtlr\" (UniqueName: \"kubernetes.io/projected/e065079d-0d3d-40b5-8103-556cb4b7f338-kube-api-access-fmtlr\") pod \"heat-db-create-gdjlc\" (UID: \"e065079d-0d3d-40b5-8103-556cb4b7f338\") " pod="openstack/heat-db-create-gdjlc" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.298911 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9lsf\" (UniqueName: \"kubernetes.io/projected/a076eb84-b8f5-4f75-922b-e44682ad8e67-kube-api-access-n9lsf\") pod \"barbican-e48e-account-create-update-zw8gt\" (UID: \"a076eb84-b8f5-4f75-922b-e44682ad8e67\") " pod="openstack/barbican-e48e-account-create-update-zw8gt" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.377590 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6f5b891-75b4-4b8c-95b1-7b92ee267793-combined-ca-bundle\") pod \"keystone-db-sync-mrxg9\" (UID: \"c6f5b891-75b4-4b8c-95b1-7b92ee267793\") " pod="openstack/keystone-db-sync-mrxg9" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.377893 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5m9g\" (UniqueName: \"kubernetes.io/projected/c6f5b891-75b4-4b8c-95b1-7b92ee267793-kube-api-access-x5m9g\") pod \"keystone-db-sync-mrxg9\" (UID: \"c6f5b891-75b4-4b8c-95b1-7b92ee267793\") " pod="openstack/keystone-db-sync-mrxg9" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.377944 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6f5b891-75b4-4b8c-95b1-7b92ee267793-config-data\") pod \"keystone-db-sync-mrxg9\" (UID: \"c6f5b891-75b4-4b8c-95b1-7b92ee267793\") " pod="openstack/keystone-db-sync-mrxg9" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.380882 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6f5b891-75b4-4b8c-95b1-7b92ee267793-combined-ca-bundle\") pod \"keystone-db-sync-mrxg9\" (UID: \"c6f5b891-75b4-4b8c-95b1-7b92ee267793\") " pod="openstack/keystone-db-sync-mrxg9" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.387343 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6f5b891-75b4-4b8c-95b1-7b92ee267793-config-data\") pod \"keystone-db-sync-mrxg9\" (UID: \"c6f5b891-75b4-4b8c-95b1-7b92ee267793\") " pod="openstack/keystone-db-sync-mrxg9" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.402258 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5m9g\" (UniqueName: \"kubernetes.io/projected/c6f5b891-75b4-4b8c-95b1-7b92ee267793-kube-api-access-x5m9g\") pod \"keystone-db-sync-mrxg9\" (UID: \"c6f5b891-75b4-4b8c-95b1-7b92ee267793\") " pod="openstack/keystone-db-sync-mrxg9" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.430268 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-sxrmq"] Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.431317 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-sxrmq" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.443601 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-da9d-account-create-update-6t8rq"] Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.446845 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-da9d-account-create-update-6t8rq" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.452333 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.466470 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-sxrmq"] Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.466988 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-1352-account-create-update-jhtbv" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.476123 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-gdjlc" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.476486 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-da9d-account-create-update-6t8rq"] Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.483970 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-e48e-account-create-update-zw8gt" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.505408 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-mrxg9" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.580429 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrbmx\" (UniqueName: \"kubernetes.io/projected/180473dd-14d1-409d-8ca8-059c9883c762-kube-api-access-lrbmx\") pod \"neutron-da9d-account-create-update-6t8rq\" (UID: \"180473dd-14d1-409d-8ca8-059c9883c762\") " pod="openstack/neutron-da9d-account-create-update-6t8rq" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.580498 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e975aca2-19e6-4a92-a4b0-f0db131bdfec-operator-scripts\") pod \"neutron-db-create-sxrmq\" (UID: \"e975aca2-19e6-4a92-a4b0-f0db131bdfec\") " pod="openstack/neutron-db-create-sxrmq" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.580553 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtj2p\" (UniqueName: \"kubernetes.io/projected/e975aca2-19e6-4a92-a4b0-f0db131bdfec-kube-api-access-xtj2p\") pod \"neutron-db-create-sxrmq\" (UID: \"e975aca2-19e6-4a92-a4b0-f0db131bdfec\") " pod="openstack/neutron-db-create-sxrmq" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.580574 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/180473dd-14d1-409d-8ca8-059c9883c762-operator-scripts\") pod \"neutron-da9d-account-create-update-6t8rq\" (UID: \"180473dd-14d1-409d-8ca8-059c9883c762\") " pod="openstack/neutron-da9d-account-create-update-6t8rq" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.658686 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-v5t59"] Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.689717 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtj2p\" (UniqueName: \"kubernetes.io/projected/e975aca2-19e6-4a92-a4b0-f0db131bdfec-kube-api-access-xtj2p\") pod \"neutron-db-create-sxrmq\" (UID: \"e975aca2-19e6-4a92-a4b0-f0db131bdfec\") " pod="openstack/neutron-db-create-sxrmq" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.689919 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/180473dd-14d1-409d-8ca8-059c9883c762-operator-scripts\") pod \"neutron-da9d-account-create-update-6t8rq\" (UID: \"180473dd-14d1-409d-8ca8-059c9883c762\") " pod="openstack/neutron-da9d-account-create-update-6t8rq" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.690018 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrbmx\" (UniqueName: \"kubernetes.io/projected/180473dd-14d1-409d-8ca8-059c9883c762-kube-api-access-lrbmx\") pod \"neutron-da9d-account-create-update-6t8rq\" (UID: \"180473dd-14d1-409d-8ca8-059c9883c762\") " pod="openstack/neutron-da9d-account-create-update-6t8rq" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.690046 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e975aca2-19e6-4a92-a4b0-f0db131bdfec-operator-scripts\") pod \"neutron-db-create-sxrmq\" (UID: \"e975aca2-19e6-4a92-a4b0-f0db131bdfec\") " pod="openstack/neutron-db-create-sxrmq" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.690727 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/180473dd-14d1-409d-8ca8-059c9883c762-operator-scripts\") pod \"neutron-da9d-account-create-update-6t8rq\" (UID: \"180473dd-14d1-409d-8ca8-059c9883c762\") " pod="openstack/neutron-da9d-account-create-update-6t8rq" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.690830 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e975aca2-19e6-4a92-a4b0-f0db131bdfec-operator-scripts\") pod \"neutron-db-create-sxrmq\" (UID: \"e975aca2-19e6-4a92-a4b0-f0db131bdfec\") " pod="openstack/neutron-db-create-sxrmq" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.715183 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrbmx\" (UniqueName: \"kubernetes.io/projected/180473dd-14d1-409d-8ca8-059c9883c762-kube-api-access-lrbmx\") pod \"neutron-da9d-account-create-update-6t8rq\" (UID: \"180473dd-14d1-409d-8ca8-059c9883c762\") " pod="openstack/neutron-da9d-account-create-update-6t8rq" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.731559 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtj2p\" (UniqueName: \"kubernetes.io/projected/e975aca2-19e6-4a92-a4b0-f0db131bdfec-kube-api-access-xtj2p\") pod \"neutron-db-create-sxrmq\" (UID: \"e975aca2-19e6-4a92-a4b0-f0db131bdfec\") " pod="openstack/neutron-db-create-sxrmq" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.752279 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-ec26-account-create-update-fxz49"] Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.807160 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-sxrmq" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.831091 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-da9d-account-create-update-6t8rq" Dec 06 03:58:13 crc kubenswrapper[4802]: I1206 03:58:13.838777 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-nv6kh"] Dec 06 03:58:14 crc kubenswrapper[4802]: W1206 03:58:14.101407 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7653cfc2_6e34_40f2_843a_9a644165b0fb.slice/crio-11dbb64b94bb0fb158cd6d0630c07cbb51932162a2380b94b75c7de73609d7f0 WatchSource:0}: Error finding container 11dbb64b94bb0fb158cd6d0630c07cbb51932162a2380b94b75c7de73609d7f0: Status 404 returned error can't find the container with id 11dbb64b94bb0fb158cd6d0630c07cbb51932162a2380b94b75c7de73609d7f0 Dec 06 03:58:14 crc kubenswrapper[4802]: I1206 03:58:14.103757 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-1352-account-create-update-jhtbv"] Dec 06 03:58:14 crc kubenswrapper[4802]: I1206 03:58:14.235890 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-mrxg9"] Dec 06 03:58:14 crc kubenswrapper[4802]: W1206 03:58:14.238095 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc6f5b891_75b4_4b8c_95b1_7b92ee267793.slice/crio-be79cf42ee56ff8a94dc5994a49b498e922ade30f01ff048c8584d70d194354f WatchSource:0}: Error finding container be79cf42ee56ff8a94dc5994a49b498e922ade30f01ff048c8584d70d194354f: Status 404 returned error can't find the container with id be79cf42ee56ff8a94dc5994a49b498e922ade30f01ff048c8584d70d194354f Dec 06 03:58:14 crc kubenswrapper[4802]: I1206 03:58:14.307810 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-e48e-account-create-update-zw8gt"] Dec 06 03:58:14 crc kubenswrapper[4802]: I1206 03:58:14.351382 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-gdjlc"] Dec 06 03:58:14 crc kubenswrapper[4802]: I1206 03:58:14.424513 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-ec26-account-create-update-fxz49" event={"ID":"0c4bfa7b-7d20-4c15-b4a7-181a05dcaab4","Type":"ContainerStarted","Data":"31d1a7925e3900fe0bc88e5bf9be4fd7e3ae5930c127b7cdb99058f6b5dfdfa1"} Dec 06 03:58:14 crc kubenswrapper[4802]: I1206 03:58:14.464449 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-1352-account-create-update-jhtbv" event={"ID":"7653cfc2-6e34-40f2-843a-9a644165b0fb","Type":"ContainerStarted","Data":"11dbb64b94bb0fb158cd6d0630c07cbb51932162a2380b94b75c7de73609d7f0"} Dec 06 03:58:14 crc kubenswrapper[4802]: I1206 03:58:14.465498 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-v5t59" event={"ID":"c7d36678-2b9e-4380-8828-a32e159ce604","Type":"ContainerStarted","Data":"28f13551b8be593798d6d4e59dd25b369dcfd9450bf293e18c81c82f92ca1f76"} Dec 06 03:58:14 crc kubenswrapper[4802]: I1206 03:58:14.469204 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"508d9d64-dd16-4d21-b492-052d123f1a6a","Type":"ContainerStarted","Data":"ed671a17c4a78d04d4daecd4ccf4f63d5e256cd99fadcc3b614c0aa114444713"} Dec 06 03:58:14 crc kubenswrapper[4802]: I1206 03:58:14.470974 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-mrxg9" event={"ID":"c6f5b891-75b4-4b8c-95b1-7b92ee267793","Type":"ContainerStarted","Data":"be79cf42ee56ff8a94dc5994a49b498e922ade30f01ff048c8584d70d194354f"} Dec 06 03:58:14 crc kubenswrapper[4802]: I1206 03:58:14.471701 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-nv6kh" event={"ID":"ebeea130-3942-475e-ba17-51d624cde585","Type":"ContainerStarted","Data":"e09f65bd93757f8d7b981acb91feca1885c23fae5677b367bac63b404723fb58"} Dec 06 03:58:14 crc kubenswrapper[4802]: I1206 03:58:14.504571 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-da9d-account-create-update-6t8rq"] Dec 06 03:58:14 crc kubenswrapper[4802]: I1206 03:58:14.519635 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-sxrmq"] Dec 06 03:58:15 crc kubenswrapper[4802]: I1206 03:58:15.481463 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-gdjlc" event={"ID":"e065079d-0d3d-40b5-8103-556cb4b7f338","Type":"ContainerStarted","Data":"542b1354a0a67db73d00a6d2e248ce1ff1d1292e523b6542ee848019ce58023f"} Dec 06 03:58:15 crc kubenswrapper[4802]: I1206 03:58:15.482633 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-da9d-account-create-update-6t8rq" event={"ID":"180473dd-14d1-409d-8ca8-059c9883c762","Type":"ContainerStarted","Data":"f75862fae196449df53b292a3c814df1ab92c76586d3b028f034e08ebba86d6b"} Dec 06 03:58:15 crc kubenswrapper[4802]: I1206 03:58:15.488015 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"508d9d64-dd16-4d21-b492-052d123f1a6a","Type":"ContainerStarted","Data":"02798df10570803b3a302ffb506e54b435b83b23c96e61d7114d0e08f3614a21"} Dec 06 03:58:15 crc kubenswrapper[4802]: I1206 03:58:15.489277 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-sxrmq" event={"ID":"e975aca2-19e6-4a92-a4b0-f0db131bdfec","Type":"ContainerStarted","Data":"c925301f9fca45004ac1ed405fbf6cc576fd2e26e3827e6c097b61daebb23616"} Dec 06 03:58:15 crc kubenswrapper[4802]: I1206 03:58:15.490449 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-e48e-account-create-update-zw8gt" event={"ID":"a076eb84-b8f5-4f75-922b-e44682ad8e67","Type":"ContainerStarted","Data":"ec43af0cfb86d9e5a010f87e46140c544a979df97ab7cef43afe239b1b4f291a"} Dec 06 03:58:16 crc kubenswrapper[4802]: I1206 03:58:16.500074 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-ec26-account-create-update-fxz49" event={"ID":"0c4bfa7b-7d20-4c15-b4a7-181a05dcaab4","Type":"ContainerStarted","Data":"8af62147f7298df9918791bebb79842e0b68d2d15c42f552492667fa7bc01da7"} Dec 06 03:58:16 crc kubenswrapper[4802]: I1206 03:58:16.501834 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-1352-account-create-update-jhtbv" event={"ID":"7653cfc2-6e34-40f2-843a-9a644165b0fb","Type":"ContainerStarted","Data":"3f4b38426be7601d64bc873172769785c9046300daa2599fab6b14601ad2253e"} Dec 06 03:58:16 crc kubenswrapper[4802]: I1206 03:58:16.504202 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-v5t59" event={"ID":"c7d36678-2b9e-4380-8828-a32e159ce604","Type":"ContainerStarted","Data":"36903355ba2d3d638b25f476eaf7539b274899782c428bcb2387df084f4ddc5d"} Dec 06 03:58:16 crc kubenswrapper[4802]: I1206 03:58:16.506213 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-da9d-account-create-update-6t8rq" event={"ID":"180473dd-14d1-409d-8ca8-059c9883c762","Type":"ContainerStarted","Data":"f920d0003b0a0bffb8461417928e8983607d12b907c7583af66b6fadb491020a"} Dec 06 03:58:16 crc kubenswrapper[4802]: I1206 03:58:16.511790 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"508d9d64-dd16-4d21-b492-052d123f1a6a","Type":"ContainerStarted","Data":"84d6a8e0dbd2c00ece4b6dff35b97dc4f636ab48e111fc16911d95bd360957af"} Dec 06 03:58:16 crc kubenswrapper[4802]: I1206 03:58:16.513488 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-sxrmq" event={"ID":"e975aca2-19e6-4a92-a4b0-f0db131bdfec","Type":"ContainerStarted","Data":"8126e8f607f92361572ba1d15723069acd373ae541bfa0c6c20ffd7a082374ec"} Dec 06 03:58:16 crc kubenswrapper[4802]: I1206 03:58:16.514843 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-gdjlc" event={"ID":"e065079d-0d3d-40b5-8103-556cb4b7f338","Type":"ContainerStarted","Data":"b5003c23488a9b4291d460938302ad94691a2a22a3ad10889c45ae5722d34b63"} Dec 06 03:58:16 crc kubenswrapper[4802]: I1206 03:58:16.516570 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-nv6kh" event={"ID":"ebeea130-3942-475e-ba17-51d624cde585","Type":"ContainerStarted","Data":"1b96b595560864842e116cdbda2b107003083adfdc7dc7068ee10aebc4ffd2e6"} Dec 06 03:58:16 crc kubenswrapper[4802]: I1206 03:58:16.521605 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-e48e-account-create-update-zw8gt" event={"ID":"a076eb84-b8f5-4f75-922b-e44682ad8e67","Type":"ContainerStarted","Data":"1e3c95d017fa3e0e5f5a8706195ef03d408a5d227166a54b29857a70dde385c2"} Dec 06 03:58:16 crc kubenswrapper[4802]: I1206 03:58:16.522705 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-ec26-account-create-update-fxz49" podStartSLOduration=4.522692406 podStartE2EDuration="4.522692406s" podCreationTimestamp="2025-12-06 03:58:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:58:16.515645084 +0000 UTC m=+1089.387554236" watchObservedRunningTime="2025-12-06 03:58:16.522692406 +0000 UTC m=+1089.394601558" Dec 06 03:58:16 crc kubenswrapper[4802]: I1206 03:58:16.558454 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-1352-account-create-update-jhtbv" podStartSLOduration=4.558436643 podStartE2EDuration="4.558436643s" podCreationTimestamp="2025-12-06 03:58:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:58:16.549612902 +0000 UTC m=+1089.421522054" watchObservedRunningTime="2025-12-06 03:58:16.558436643 +0000 UTC m=+1089.430345805" Dec 06 03:58:16 crc kubenswrapper[4802]: I1206 03:58:16.567950 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-v5t59" podStartSLOduration=4.567929073 podStartE2EDuration="4.567929073s" podCreationTimestamp="2025-12-06 03:58:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:58:16.564066707 +0000 UTC m=+1089.435975859" watchObservedRunningTime="2025-12-06 03:58:16.567929073 +0000 UTC m=+1089.439838225" Dec 06 03:58:16 crc kubenswrapper[4802]: I1206 03:58:16.580597 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-e48e-account-create-update-zw8gt" podStartSLOduration=4.580578718 podStartE2EDuration="4.580578718s" podCreationTimestamp="2025-12-06 03:58:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:58:16.579244172 +0000 UTC m=+1089.451153324" watchObservedRunningTime="2025-12-06 03:58:16.580578718 +0000 UTC m=+1089.452487870" Dec 06 03:58:16 crc kubenswrapper[4802]: I1206 03:58:16.598739 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-nv6kh" podStartSLOduration=4.5987229549999995 podStartE2EDuration="4.598722955s" podCreationTimestamp="2025-12-06 03:58:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:58:16.591408704 +0000 UTC m=+1089.463317856" watchObservedRunningTime="2025-12-06 03:58:16.598722955 +0000 UTC m=+1089.470632107" Dec 06 03:58:17 crc kubenswrapper[4802]: I1206 03:58:17.566094 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-sxrmq" podStartSLOduration=4.566072167 podStartE2EDuration="4.566072167s" podCreationTimestamp="2025-12-06 03:58:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:58:17.548361404 +0000 UTC m=+1090.420270586" watchObservedRunningTime="2025-12-06 03:58:17.566072167 +0000 UTC m=+1090.437981319" Dec 06 03:58:17 crc kubenswrapper[4802]: I1206 03:58:17.570377 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-da9d-account-create-update-6t8rq" podStartSLOduration=4.570365734 podStartE2EDuration="4.570365734s" podCreationTimestamp="2025-12-06 03:58:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:58:17.565947554 +0000 UTC m=+1090.437856706" watchObservedRunningTime="2025-12-06 03:58:17.570365734 +0000 UTC m=+1090.442274886" Dec 06 03:58:17 crc kubenswrapper[4802]: I1206 03:58:17.585826 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-create-gdjlc" podStartSLOduration=5.585804456 podStartE2EDuration="5.585804456s" podCreationTimestamp="2025-12-06 03:58:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:58:17.577093809 +0000 UTC m=+1090.449002991" watchObservedRunningTime="2025-12-06 03:58:17.585804456 +0000 UTC m=+1090.457713608" Dec 06 03:58:25 crc kubenswrapper[4802]: I1206 03:58:25.634938 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"508d9d64-dd16-4d21-b492-052d123f1a6a","Type":"ContainerStarted","Data":"54d2602d9702976f34eaa445734cdf6765742ce7073a65133f18e08628f8f7c2"} Dec 06 03:58:26 crc kubenswrapper[4802]: I1206 03:58:26.644413 4802 generic.go:334] "Generic (PLEG): container finished" podID="0c4bfa7b-7d20-4c15-b4a7-181a05dcaab4" containerID="8af62147f7298df9918791bebb79842e0b68d2d15c42f552492667fa7bc01da7" exitCode=0 Dec 06 03:58:26 crc kubenswrapper[4802]: I1206 03:58:26.644508 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-ec26-account-create-update-fxz49" event={"ID":"0c4bfa7b-7d20-4c15-b4a7-181a05dcaab4","Type":"ContainerDied","Data":"8af62147f7298df9918791bebb79842e0b68d2d15c42f552492667fa7bc01da7"} Dec 06 03:58:26 crc kubenswrapper[4802]: I1206 03:58:26.646737 4802 generic.go:334] "Generic (PLEG): container finished" podID="7653cfc2-6e34-40f2-843a-9a644165b0fb" containerID="3f4b38426be7601d64bc873172769785c9046300daa2599fab6b14601ad2253e" exitCode=0 Dec 06 03:58:26 crc kubenswrapper[4802]: I1206 03:58:26.646809 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-1352-account-create-update-jhtbv" event={"ID":"7653cfc2-6e34-40f2-843a-9a644165b0fb","Type":"ContainerDied","Data":"3f4b38426be7601d64bc873172769785c9046300daa2599fab6b14601ad2253e"} Dec 06 03:58:26 crc kubenswrapper[4802]: I1206 03:58:26.648547 4802 generic.go:334] "Generic (PLEG): container finished" podID="180473dd-14d1-409d-8ca8-059c9883c762" containerID="f920d0003b0a0bffb8461417928e8983607d12b907c7583af66b6fadb491020a" exitCode=0 Dec 06 03:58:26 crc kubenswrapper[4802]: I1206 03:58:26.648614 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-da9d-account-create-update-6t8rq" event={"ID":"180473dd-14d1-409d-8ca8-059c9883c762","Type":"ContainerDied","Data":"f920d0003b0a0bffb8461417928e8983607d12b907c7583af66b6fadb491020a"} Dec 06 03:58:26 crc kubenswrapper[4802]: I1206 03:58:26.650050 4802 generic.go:334] "Generic (PLEG): container finished" podID="c7d36678-2b9e-4380-8828-a32e159ce604" containerID="36903355ba2d3d638b25f476eaf7539b274899782c428bcb2387df084f4ddc5d" exitCode=0 Dec 06 03:58:26 crc kubenswrapper[4802]: I1206 03:58:26.650137 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-v5t59" event={"ID":"c7d36678-2b9e-4380-8828-a32e159ce604","Type":"ContainerDied","Data":"36903355ba2d3d638b25f476eaf7539b274899782c428bcb2387df084f4ddc5d"} Dec 06 03:58:26 crc kubenswrapper[4802]: I1206 03:58:26.651520 4802 generic.go:334] "Generic (PLEG): container finished" podID="e975aca2-19e6-4a92-a4b0-f0db131bdfec" containerID="8126e8f607f92361572ba1d15723069acd373ae541bfa0c6c20ffd7a082374ec" exitCode=0 Dec 06 03:58:26 crc kubenswrapper[4802]: I1206 03:58:26.651576 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-sxrmq" event={"ID":"e975aca2-19e6-4a92-a4b0-f0db131bdfec","Type":"ContainerDied","Data":"8126e8f607f92361572ba1d15723069acd373ae541bfa0c6c20ffd7a082374ec"} Dec 06 03:58:26 crc kubenswrapper[4802]: I1206 03:58:26.653061 4802 generic.go:334] "Generic (PLEG): container finished" podID="ebeea130-3942-475e-ba17-51d624cde585" containerID="1b96b595560864842e116cdbda2b107003083adfdc7dc7068ee10aebc4ffd2e6" exitCode=0 Dec 06 03:58:26 crc kubenswrapper[4802]: I1206 03:58:26.653137 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-nv6kh" event={"ID":"ebeea130-3942-475e-ba17-51d624cde585","Type":"ContainerDied","Data":"1b96b595560864842e116cdbda2b107003083adfdc7dc7068ee10aebc4ffd2e6"} Dec 06 03:58:26 crc kubenswrapper[4802]: I1206 03:58:26.656040 4802 generic.go:334] "Generic (PLEG): container finished" podID="a076eb84-b8f5-4f75-922b-e44682ad8e67" containerID="1e3c95d017fa3e0e5f5a8706195ef03d408a5d227166a54b29857a70dde385c2" exitCode=0 Dec 06 03:58:26 crc kubenswrapper[4802]: I1206 03:58:26.656103 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-e48e-account-create-update-zw8gt" event={"ID":"a076eb84-b8f5-4f75-922b-e44682ad8e67","Type":"ContainerDied","Data":"1e3c95d017fa3e0e5f5a8706195ef03d408a5d227166a54b29857a70dde385c2"} Dec 06 03:58:26 crc kubenswrapper[4802]: I1206 03:58:26.657815 4802 generic.go:334] "Generic (PLEG): container finished" podID="e065079d-0d3d-40b5-8103-556cb4b7f338" containerID="b5003c23488a9b4291d460938302ad94691a2a22a3ad10889c45ae5722d34b63" exitCode=0 Dec 06 03:58:26 crc kubenswrapper[4802]: I1206 03:58:26.657893 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-gdjlc" event={"ID":"e065079d-0d3d-40b5-8103-556cb4b7f338","Type":"ContainerDied","Data":"b5003c23488a9b4291d460938302ad94691a2a22a3ad10889c45ae5722d34b63"} Dec 06 03:58:27 crc kubenswrapper[4802]: I1206 03:58:27.673440 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"508d9d64-dd16-4d21-b492-052d123f1a6a","Type":"ContainerStarted","Data":"3e2c4017bf03cf13918bc4da72d32c3f7f0ca4d2981e727333db08f6418e9096"} Dec 06 03:58:27 crc kubenswrapper[4802]: I1206 03:58:27.678872 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-mrxg9" event={"ID":"c6f5b891-75b4-4b8c-95b1-7b92ee267793","Type":"ContainerStarted","Data":"2e82f2fc1c3bfcea72061447fb05fb9dd187735d00c8db3b4ed7beb9e7607c72"} Dec 06 03:58:27 crc kubenswrapper[4802]: I1206 03:58:27.727297 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=49.953056197 podStartE2EDuration="1m0.727268746s" podCreationTimestamp="2025-12-06 03:57:27 +0000 UTC" firstStartedPulling="2025-12-06 03:58:01.073060099 +0000 UTC m=+1073.944969251" lastFinishedPulling="2025-12-06 03:58:11.847272648 +0000 UTC m=+1084.719181800" observedRunningTime="2025-12-06 03:58:27.708309588 +0000 UTC m=+1100.580218740" watchObservedRunningTime="2025-12-06 03:58:27.727268746 +0000 UTC m=+1100.599177908" Dec 06 03:58:27 crc kubenswrapper[4802]: I1206 03:58:27.745857 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-mrxg9" podStartSLOduration=1.60223666 podStartE2EDuration="14.745838394s" podCreationTimestamp="2025-12-06 03:58:13 +0000 UTC" firstStartedPulling="2025-12-06 03:58:14.241082711 +0000 UTC m=+1087.112991863" lastFinishedPulling="2025-12-06 03:58:27.384684445 +0000 UTC m=+1100.256593597" observedRunningTime="2025-12-06 03:58:27.737774803 +0000 UTC m=+1100.609683955" watchObservedRunningTime="2025-12-06 03:58:27.745838394 +0000 UTC m=+1100.617747546" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.073190 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-w49ls"] Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.077324 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-w49ls" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.090294 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.098955 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-w49ls"] Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.150090 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-1352-account-create-update-jhtbv" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.208718 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/37b0a45e-9768-4b03-bf26-d35d3c98f58a-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-w49ls\" (UID: \"37b0a45e-9768-4b03-bf26-d35d3c98f58a\") " pod="openstack/dnsmasq-dns-5c79d794d7-w49ls" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.208798 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhwmh\" (UniqueName: \"kubernetes.io/projected/37b0a45e-9768-4b03-bf26-d35d3c98f58a-kube-api-access-mhwmh\") pod \"dnsmasq-dns-5c79d794d7-w49ls\" (UID: \"37b0a45e-9768-4b03-bf26-d35d3c98f58a\") " pod="openstack/dnsmasq-dns-5c79d794d7-w49ls" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.208822 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/37b0a45e-9768-4b03-bf26-d35d3c98f58a-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-w49ls\" (UID: \"37b0a45e-9768-4b03-bf26-d35d3c98f58a\") " pod="openstack/dnsmasq-dns-5c79d794d7-w49ls" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.208850 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37b0a45e-9768-4b03-bf26-d35d3c98f58a-config\") pod \"dnsmasq-dns-5c79d794d7-w49ls\" (UID: \"37b0a45e-9768-4b03-bf26-d35d3c98f58a\") " pod="openstack/dnsmasq-dns-5c79d794d7-w49ls" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.208893 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/37b0a45e-9768-4b03-bf26-d35d3c98f58a-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-w49ls\" (UID: \"37b0a45e-9768-4b03-bf26-d35d3c98f58a\") " pod="openstack/dnsmasq-dns-5c79d794d7-w49ls" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.208939 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/37b0a45e-9768-4b03-bf26-d35d3c98f58a-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-w49ls\" (UID: \"37b0a45e-9768-4b03-bf26-d35d3c98f58a\") " pod="openstack/dnsmasq-dns-5c79d794d7-w49ls" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.314030 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tbnch\" (UniqueName: \"kubernetes.io/projected/7653cfc2-6e34-40f2-843a-9a644165b0fb-kube-api-access-tbnch\") pod \"7653cfc2-6e34-40f2-843a-9a644165b0fb\" (UID: \"7653cfc2-6e34-40f2-843a-9a644165b0fb\") " Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.314190 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7653cfc2-6e34-40f2-843a-9a644165b0fb-operator-scripts\") pod \"7653cfc2-6e34-40f2-843a-9a644165b0fb\" (UID: \"7653cfc2-6e34-40f2-843a-9a644165b0fb\") " Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.314525 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/37b0a45e-9768-4b03-bf26-d35d3c98f58a-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-w49ls\" (UID: \"37b0a45e-9768-4b03-bf26-d35d3c98f58a\") " pod="openstack/dnsmasq-dns-5c79d794d7-w49ls" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.314636 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/37b0a45e-9768-4b03-bf26-d35d3c98f58a-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-w49ls\" (UID: \"37b0a45e-9768-4b03-bf26-d35d3c98f58a\") " pod="openstack/dnsmasq-dns-5c79d794d7-w49ls" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.314664 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhwmh\" (UniqueName: \"kubernetes.io/projected/37b0a45e-9768-4b03-bf26-d35d3c98f58a-kube-api-access-mhwmh\") pod \"dnsmasq-dns-5c79d794d7-w49ls\" (UID: \"37b0a45e-9768-4b03-bf26-d35d3c98f58a\") " pod="openstack/dnsmasq-dns-5c79d794d7-w49ls" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.314685 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/37b0a45e-9768-4b03-bf26-d35d3c98f58a-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-w49ls\" (UID: \"37b0a45e-9768-4b03-bf26-d35d3c98f58a\") " pod="openstack/dnsmasq-dns-5c79d794d7-w49ls" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.314708 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37b0a45e-9768-4b03-bf26-d35d3c98f58a-config\") pod \"dnsmasq-dns-5c79d794d7-w49ls\" (UID: \"37b0a45e-9768-4b03-bf26-d35d3c98f58a\") " pod="openstack/dnsmasq-dns-5c79d794d7-w49ls" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.314756 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/37b0a45e-9768-4b03-bf26-d35d3c98f58a-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-w49ls\" (UID: \"37b0a45e-9768-4b03-bf26-d35d3c98f58a\") " pod="openstack/dnsmasq-dns-5c79d794d7-w49ls" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.315543 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/37b0a45e-9768-4b03-bf26-d35d3c98f58a-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-w49ls\" (UID: \"37b0a45e-9768-4b03-bf26-d35d3c98f58a\") " pod="openstack/dnsmasq-dns-5c79d794d7-w49ls" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.317762 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/37b0a45e-9768-4b03-bf26-d35d3c98f58a-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-w49ls\" (UID: \"37b0a45e-9768-4b03-bf26-d35d3c98f58a\") " pod="openstack/dnsmasq-dns-5c79d794d7-w49ls" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.322242 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7653cfc2-6e34-40f2-843a-9a644165b0fb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7653cfc2-6e34-40f2-843a-9a644165b0fb" (UID: "7653cfc2-6e34-40f2-843a-9a644165b0fb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.326361 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37b0a45e-9768-4b03-bf26-d35d3c98f58a-config\") pod \"dnsmasq-dns-5c79d794d7-w49ls\" (UID: \"37b0a45e-9768-4b03-bf26-d35d3c98f58a\") " pod="openstack/dnsmasq-dns-5c79d794d7-w49ls" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.328399 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/37b0a45e-9768-4b03-bf26-d35d3c98f58a-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-w49ls\" (UID: \"37b0a45e-9768-4b03-bf26-d35d3c98f58a\") " pod="openstack/dnsmasq-dns-5c79d794d7-w49ls" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.331900 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/37b0a45e-9768-4b03-bf26-d35d3c98f58a-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-w49ls\" (UID: \"37b0a45e-9768-4b03-bf26-d35d3c98f58a\") " pod="openstack/dnsmasq-dns-5c79d794d7-w49ls" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.334500 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7653cfc2-6e34-40f2-843a-9a644165b0fb-kube-api-access-tbnch" (OuterVolumeSpecName: "kube-api-access-tbnch") pod "7653cfc2-6e34-40f2-843a-9a644165b0fb" (UID: "7653cfc2-6e34-40f2-843a-9a644165b0fb"). InnerVolumeSpecName "kube-api-access-tbnch". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.351472 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhwmh\" (UniqueName: \"kubernetes.io/projected/37b0a45e-9768-4b03-bf26-d35d3c98f58a-kube-api-access-mhwmh\") pod \"dnsmasq-dns-5c79d794d7-w49ls\" (UID: \"37b0a45e-9768-4b03-bf26-d35d3c98f58a\") " pod="openstack/dnsmasq-dns-5c79d794d7-w49ls" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.416207 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tbnch\" (UniqueName: \"kubernetes.io/projected/7653cfc2-6e34-40f2-843a-9a644165b0fb-kube-api-access-tbnch\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.416252 4802 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7653cfc2-6e34-40f2-843a-9a644165b0fb-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.465584 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-w49ls" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.686160 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-gdjlc" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.701039 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-1352-account-create-update-jhtbv" event={"ID":"7653cfc2-6e34-40f2-843a-9a644165b0fb","Type":"ContainerDied","Data":"11dbb64b94bb0fb158cd6d0630c07cbb51932162a2380b94b75c7de73609d7f0"} Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.701061 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-1352-account-create-update-jhtbv" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.701087 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="11dbb64b94bb0fb158cd6d0630c07cbb51932162a2380b94b75c7de73609d7f0" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.704996 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-v5t59" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.710215 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-da9d-account-create-update-6t8rq" event={"ID":"180473dd-14d1-409d-8ca8-059c9883c762","Type":"ContainerDied","Data":"f75862fae196449df53b292a3c814df1ab92c76586d3b028f034e08ebba86d6b"} Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.710249 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f75862fae196449df53b292a3c814df1ab92c76586d3b028f034e08ebba86d6b" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.710500 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-e48e-account-create-update-zw8gt" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.711387 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-v5t59" event={"ID":"c7d36678-2b9e-4380-8828-a32e159ce604","Type":"ContainerDied","Data":"28f13551b8be593798d6d4e59dd25b369dcfd9450bf293e18c81c82f92ca1f76"} Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.711408 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="28f13551b8be593798d6d4e59dd25b369dcfd9450bf293e18c81c82f92ca1f76" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.715845 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-v5t59" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.720827 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-sxrmq" event={"ID":"e975aca2-19e6-4a92-a4b0-f0db131bdfec","Type":"ContainerDied","Data":"c925301f9fca45004ac1ed405fbf6cc576fd2e26e3827e6c097b61daebb23616"} Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.720869 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c925301f9fca45004ac1ed405fbf6cc576fd2e26e3827e6c097b61daebb23616" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.722882 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-nv6kh" event={"ID":"ebeea130-3942-475e-ba17-51d624cde585","Type":"ContainerDied","Data":"e09f65bd93757f8d7b981acb91feca1885c23fae5677b367bac63b404723fb58"} Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.722904 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e09f65bd93757f8d7b981acb91feca1885c23fae5677b367bac63b404723fb58" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.724085 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-e48e-account-create-update-zw8gt" event={"ID":"a076eb84-b8f5-4f75-922b-e44682ad8e67","Type":"ContainerDied","Data":"ec43af0cfb86d9e5a010f87e46140c544a979df97ab7cef43afe239b1b4f291a"} Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.724109 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ec43af0cfb86d9e5a010f87e46140c544a979df97ab7cef43afe239b1b4f291a" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.724169 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-e48e-account-create-update-zw8gt" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.726028 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-gdjlc" event={"ID":"e065079d-0d3d-40b5-8103-556cb4b7f338","Type":"ContainerDied","Data":"542b1354a0a67db73d00a6d2e248ce1ff1d1292e523b6542ee848019ce58023f"} Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.726062 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="542b1354a0a67db73d00a6d2e248ce1ff1d1292e523b6542ee848019ce58023f" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.726212 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-gdjlc" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.733331 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-ec26-account-create-update-fxz49" event={"ID":"0c4bfa7b-7d20-4c15-b4a7-181a05dcaab4","Type":"ContainerDied","Data":"31d1a7925e3900fe0bc88e5bf9be4fd7e3ae5930c127b7cdb99058f6b5dfdfa1"} Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.733367 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="31d1a7925e3900fe0bc88e5bf9be4fd7e3ae5930c127b7cdb99058f6b5dfdfa1" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.733944 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-da9d-account-create-update-6t8rq" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.750966 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-sxrmq" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.761673 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-nv6kh" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.792186 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-ec26-account-create-update-fxz49" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.823205 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fmtlr\" (UniqueName: \"kubernetes.io/projected/e065079d-0d3d-40b5-8103-556cb4b7f338-kube-api-access-fmtlr\") pod \"e065079d-0d3d-40b5-8103-556cb4b7f338\" (UID: \"e065079d-0d3d-40b5-8103-556cb4b7f338\") " Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.823315 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e065079d-0d3d-40b5-8103-556cb4b7f338-operator-scripts\") pod \"e065079d-0d3d-40b5-8103-556cb4b7f338\" (UID: \"e065079d-0d3d-40b5-8103-556cb4b7f338\") " Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.823371 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c7d36678-2b9e-4380-8828-a32e159ce604-operator-scripts\") pod \"c7d36678-2b9e-4380-8828-a32e159ce604\" (UID: \"c7d36678-2b9e-4380-8828-a32e159ce604\") " Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.823429 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lrbmx\" (UniqueName: \"kubernetes.io/projected/180473dd-14d1-409d-8ca8-059c9883c762-kube-api-access-lrbmx\") pod \"180473dd-14d1-409d-8ca8-059c9883c762\" (UID: \"180473dd-14d1-409d-8ca8-059c9883c762\") " Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.823450 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/180473dd-14d1-409d-8ca8-059c9883c762-operator-scripts\") pod \"180473dd-14d1-409d-8ca8-059c9883c762\" (UID: \"180473dd-14d1-409d-8ca8-059c9883c762\") " Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.823474 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e975aca2-19e6-4a92-a4b0-f0db131bdfec-operator-scripts\") pod \"e975aca2-19e6-4a92-a4b0-f0db131bdfec\" (UID: \"e975aca2-19e6-4a92-a4b0-f0db131bdfec\") " Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.823547 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a076eb84-b8f5-4f75-922b-e44682ad8e67-operator-scripts\") pod \"a076eb84-b8f5-4f75-922b-e44682ad8e67\" (UID: \"a076eb84-b8f5-4f75-922b-e44682ad8e67\") " Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.823593 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5jchl\" (UniqueName: \"kubernetes.io/projected/c7d36678-2b9e-4380-8828-a32e159ce604-kube-api-access-5jchl\") pod \"c7d36678-2b9e-4380-8828-a32e159ce604\" (UID: \"c7d36678-2b9e-4380-8828-a32e159ce604\") " Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.823622 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fssrm\" (UniqueName: \"kubernetes.io/projected/ebeea130-3942-475e-ba17-51d624cde585-kube-api-access-fssrm\") pod \"ebeea130-3942-475e-ba17-51d624cde585\" (UID: \"ebeea130-3942-475e-ba17-51d624cde585\") " Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.823645 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ebeea130-3942-475e-ba17-51d624cde585-operator-scripts\") pod \"ebeea130-3942-475e-ba17-51d624cde585\" (UID: \"ebeea130-3942-475e-ba17-51d624cde585\") " Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.823677 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xtj2p\" (UniqueName: \"kubernetes.io/projected/e975aca2-19e6-4a92-a4b0-f0db131bdfec-kube-api-access-xtj2p\") pod \"e975aca2-19e6-4a92-a4b0-f0db131bdfec\" (UID: \"e975aca2-19e6-4a92-a4b0-f0db131bdfec\") " Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.823706 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n9lsf\" (UniqueName: \"kubernetes.io/projected/a076eb84-b8f5-4f75-922b-e44682ad8e67-kube-api-access-n9lsf\") pod \"a076eb84-b8f5-4f75-922b-e44682ad8e67\" (UID: \"a076eb84-b8f5-4f75-922b-e44682ad8e67\") " Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.827156 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/180473dd-14d1-409d-8ca8-059c9883c762-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "180473dd-14d1-409d-8ca8-059c9883c762" (UID: "180473dd-14d1-409d-8ca8-059c9883c762"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.827185 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a076eb84-b8f5-4f75-922b-e44682ad8e67-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a076eb84-b8f5-4f75-922b-e44682ad8e67" (UID: "a076eb84-b8f5-4f75-922b-e44682ad8e67"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.827378 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e975aca2-19e6-4a92-a4b0-f0db131bdfec-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e975aca2-19e6-4a92-a4b0-f0db131bdfec" (UID: "e975aca2-19e6-4a92-a4b0-f0db131bdfec"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.827557 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e065079d-0d3d-40b5-8103-556cb4b7f338-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e065079d-0d3d-40b5-8103-556cb4b7f338" (UID: "e065079d-0d3d-40b5-8103-556cb4b7f338"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.827943 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c7d36678-2b9e-4380-8828-a32e159ce604-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c7d36678-2b9e-4380-8828-a32e159ce604" (UID: "c7d36678-2b9e-4380-8828-a32e159ce604"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.827939 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ebeea130-3942-475e-ba17-51d624cde585-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ebeea130-3942-475e-ba17-51d624cde585" (UID: "ebeea130-3942-475e-ba17-51d624cde585"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.830075 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e065079d-0d3d-40b5-8103-556cb4b7f338-kube-api-access-fmtlr" (OuterVolumeSpecName: "kube-api-access-fmtlr") pod "e065079d-0d3d-40b5-8103-556cb4b7f338" (UID: "e065079d-0d3d-40b5-8103-556cb4b7f338"). InnerVolumeSpecName "kube-api-access-fmtlr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.833163 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7d36678-2b9e-4380-8828-a32e159ce604-kube-api-access-5jchl" (OuterVolumeSpecName: "kube-api-access-5jchl") pod "c7d36678-2b9e-4380-8828-a32e159ce604" (UID: "c7d36678-2b9e-4380-8828-a32e159ce604"). InnerVolumeSpecName "kube-api-access-5jchl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.835087 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/180473dd-14d1-409d-8ca8-059c9883c762-kube-api-access-lrbmx" (OuterVolumeSpecName: "kube-api-access-lrbmx") pod "180473dd-14d1-409d-8ca8-059c9883c762" (UID: "180473dd-14d1-409d-8ca8-059c9883c762"). InnerVolumeSpecName "kube-api-access-lrbmx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.835158 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ebeea130-3942-475e-ba17-51d624cde585-kube-api-access-fssrm" (OuterVolumeSpecName: "kube-api-access-fssrm") pod "ebeea130-3942-475e-ba17-51d624cde585" (UID: "ebeea130-3942-475e-ba17-51d624cde585"). InnerVolumeSpecName "kube-api-access-fssrm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.835158 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e975aca2-19e6-4a92-a4b0-f0db131bdfec-kube-api-access-xtj2p" (OuterVolumeSpecName: "kube-api-access-xtj2p") pod "e975aca2-19e6-4a92-a4b0-f0db131bdfec" (UID: "e975aca2-19e6-4a92-a4b0-f0db131bdfec"). InnerVolumeSpecName "kube-api-access-xtj2p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.835178 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a076eb84-b8f5-4f75-922b-e44682ad8e67-kube-api-access-n9lsf" (OuterVolumeSpecName: "kube-api-access-n9lsf") pod "a076eb84-b8f5-4f75-922b-e44682ad8e67" (UID: "a076eb84-b8f5-4f75-922b-e44682ad8e67"). InnerVolumeSpecName "kube-api-access-n9lsf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.927173 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qrqc9\" (UniqueName: \"kubernetes.io/projected/0c4bfa7b-7d20-4c15-b4a7-181a05dcaab4-kube-api-access-qrqc9\") pod \"0c4bfa7b-7d20-4c15-b4a7-181a05dcaab4\" (UID: \"0c4bfa7b-7d20-4c15-b4a7-181a05dcaab4\") " Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.927323 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c4bfa7b-7d20-4c15-b4a7-181a05dcaab4-operator-scripts\") pod \"0c4bfa7b-7d20-4c15-b4a7-181a05dcaab4\" (UID: \"0c4bfa7b-7d20-4c15-b4a7-181a05dcaab4\") " Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.927779 4802 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a076eb84-b8f5-4f75-922b-e44682ad8e67-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.927796 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5jchl\" (UniqueName: \"kubernetes.io/projected/c7d36678-2b9e-4380-8828-a32e159ce604-kube-api-access-5jchl\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.927808 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fssrm\" (UniqueName: \"kubernetes.io/projected/ebeea130-3942-475e-ba17-51d624cde585-kube-api-access-fssrm\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.927817 4802 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ebeea130-3942-475e-ba17-51d624cde585-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.927826 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xtj2p\" (UniqueName: \"kubernetes.io/projected/e975aca2-19e6-4a92-a4b0-f0db131bdfec-kube-api-access-xtj2p\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.927852 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n9lsf\" (UniqueName: \"kubernetes.io/projected/a076eb84-b8f5-4f75-922b-e44682ad8e67-kube-api-access-n9lsf\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.927862 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fmtlr\" (UniqueName: \"kubernetes.io/projected/e065079d-0d3d-40b5-8103-556cb4b7f338-kube-api-access-fmtlr\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.927871 4802 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e065079d-0d3d-40b5-8103-556cb4b7f338-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.927879 4802 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c7d36678-2b9e-4380-8828-a32e159ce604-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.927888 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lrbmx\" (UniqueName: \"kubernetes.io/projected/180473dd-14d1-409d-8ca8-059c9883c762-kube-api-access-lrbmx\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.927896 4802 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/180473dd-14d1-409d-8ca8-059c9883c762-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.927904 4802 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e975aca2-19e6-4a92-a4b0-f0db131bdfec-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.927966 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c4bfa7b-7d20-4c15-b4a7-181a05dcaab4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0c4bfa7b-7d20-4c15-b4a7-181a05dcaab4" (UID: "0c4bfa7b-7d20-4c15-b4a7-181a05dcaab4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:58:28 crc kubenswrapper[4802]: I1206 03:58:28.930366 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c4bfa7b-7d20-4c15-b4a7-181a05dcaab4-kube-api-access-qrqc9" (OuterVolumeSpecName: "kube-api-access-qrqc9") pod "0c4bfa7b-7d20-4c15-b4a7-181a05dcaab4" (UID: "0c4bfa7b-7d20-4c15-b4a7-181a05dcaab4"). InnerVolumeSpecName "kube-api-access-qrqc9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:58:29 crc kubenswrapper[4802]: I1206 03:58:29.028808 4802 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c4bfa7b-7d20-4c15-b4a7-181a05dcaab4-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:29 crc kubenswrapper[4802]: I1206 03:58:29.028830 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qrqc9\" (UniqueName: \"kubernetes.io/projected/0c4bfa7b-7d20-4c15-b4a7-181a05dcaab4-kube-api-access-qrqc9\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:29 crc kubenswrapper[4802]: I1206 03:58:29.075128 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-w49ls"] Dec 06 03:58:29 crc kubenswrapper[4802]: W1206 03:58:29.093561 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod37b0a45e_9768_4b03_bf26_d35d3c98f58a.slice/crio-64f380ed649f9607bbd62872585717105bf77daeca9a78e75fcd52c063e8050e WatchSource:0}: Error finding container 64f380ed649f9607bbd62872585717105bf77daeca9a78e75fcd52c063e8050e: Status 404 returned error can't find the container with id 64f380ed649f9607bbd62872585717105bf77daeca9a78e75fcd52c063e8050e Dec 06 03:58:29 crc kubenswrapper[4802]: I1206 03:58:29.743082 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-w49ls" event={"ID":"37b0a45e-9768-4b03-bf26-d35d3c98f58a","Type":"ContainerStarted","Data":"64f380ed649f9607bbd62872585717105bf77daeca9a78e75fcd52c063e8050e"} Dec 06 03:58:29 crc kubenswrapper[4802]: I1206 03:58:29.743115 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-da9d-account-create-update-6t8rq" Dec 06 03:58:29 crc kubenswrapper[4802]: I1206 03:58:29.743709 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-sxrmq" Dec 06 03:58:29 crc kubenswrapper[4802]: I1206 03:58:29.743115 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-ec26-account-create-update-fxz49" Dec 06 03:58:29 crc kubenswrapper[4802]: I1206 03:58:29.743785 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-nv6kh" Dec 06 03:58:30 crc kubenswrapper[4802]: I1206 03:58:30.753420 4802 generic.go:334] "Generic (PLEG): container finished" podID="37b0a45e-9768-4b03-bf26-d35d3c98f58a" containerID="c8ea67068f3ec73d843aaf2410f2c467b06159bad710370a55376c11e644a279" exitCode=0 Dec 06 03:58:30 crc kubenswrapper[4802]: I1206 03:58:30.753458 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-w49ls" event={"ID":"37b0a45e-9768-4b03-bf26-d35d3c98f58a","Type":"ContainerDied","Data":"c8ea67068f3ec73d843aaf2410f2c467b06159bad710370a55376c11e644a279"} Dec 06 03:58:31 crc kubenswrapper[4802]: I1206 03:58:31.768014 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-w49ls" event={"ID":"37b0a45e-9768-4b03-bf26-d35d3c98f58a","Type":"ContainerStarted","Data":"99b3ae2fcbdcf8dd15fdf93bc19b0329b9e53253fa4267acb98239a199207902"} Dec 06 03:58:31 crc kubenswrapper[4802]: I1206 03:58:31.768449 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c79d794d7-w49ls" Dec 06 03:58:31 crc kubenswrapper[4802]: I1206 03:58:31.799991 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c79d794d7-w49ls" podStartSLOduration=3.799966063 podStartE2EDuration="3.799966063s" podCreationTimestamp="2025-12-06 03:58:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:58:31.79288918 +0000 UTC m=+1104.664798332" watchObservedRunningTime="2025-12-06 03:58:31.799966063 +0000 UTC m=+1104.671875235" Dec 06 03:58:35 crc kubenswrapper[4802]: I1206 03:58:35.800797 4802 generic.go:334] "Generic (PLEG): container finished" podID="c6f5b891-75b4-4b8c-95b1-7b92ee267793" containerID="2e82f2fc1c3bfcea72061447fb05fb9dd187735d00c8db3b4ed7beb9e7607c72" exitCode=0 Dec 06 03:58:35 crc kubenswrapper[4802]: I1206 03:58:35.800898 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-mrxg9" event={"ID":"c6f5b891-75b4-4b8c-95b1-7b92ee267793","Type":"ContainerDied","Data":"2e82f2fc1c3bfcea72061447fb05fb9dd187735d00c8db3b4ed7beb9e7607c72"} Dec 06 03:58:36 crc kubenswrapper[4802]: I1206 03:58:36.829269 4802 generic.go:334] "Generic (PLEG): container finished" podID="5d947661-37cd-49c2-bd89-e927e4034ba4" containerID="46e6ea4ac88dfb898ea9b663b1a6d76d786318b1afea471fa0d10200b8a30b87" exitCode=0 Dec 06 03:58:36 crc kubenswrapper[4802]: I1206 03:58:36.829461 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-mlhfp" event={"ID":"5d947661-37cd-49c2-bd89-e927e4034ba4","Type":"ContainerDied","Data":"46e6ea4ac88dfb898ea9b663b1a6d76d786318b1afea471fa0d10200b8a30b87"} Dec 06 03:58:37 crc kubenswrapper[4802]: I1206 03:58:37.164670 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-mrxg9" Dec 06 03:58:37 crc kubenswrapper[4802]: I1206 03:58:37.262342 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5m9g\" (UniqueName: \"kubernetes.io/projected/c6f5b891-75b4-4b8c-95b1-7b92ee267793-kube-api-access-x5m9g\") pod \"c6f5b891-75b4-4b8c-95b1-7b92ee267793\" (UID: \"c6f5b891-75b4-4b8c-95b1-7b92ee267793\") " Dec 06 03:58:37 crc kubenswrapper[4802]: I1206 03:58:37.262457 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6f5b891-75b4-4b8c-95b1-7b92ee267793-combined-ca-bundle\") pod \"c6f5b891-75b4-4b8c-95b1-7b92ee267793\" (UID: \"c6f5b891-75b4-4b8c-95b1-7b92ee267793\") " Dec 06 03:58:37 crc kubenswrapper[4802]: I1206 03:58:37.262699 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6f5b891-75b4-4b8c-95b1-7b92ee267793-config-data\") pod \"c6f5b891-75b4-4b8c-95b1-7b92ee267793\" (UID: \"c6f5b891-75b4-4b8c-95b1-7b92ee267793\") " Dec 06 03:58:37 crc kubenswrapper[4802]: I1206 03:58:37.270685 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6f5b891-75b4-4b8c-95b1-7b92ee267793-kube-api-access-x5m9g" (OuterVolumeSpecName: "kube-api-access-x5m9g") pod "c6f5b891-75b4-4b8c-95b1-7b92ee267793" (UID: "c6f5b891-75b4-4b8c-95b1-7b92ee267793"). InnerVolumeSpecName "kube-api-access-x5m9g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:58:37 crc kubenswrapper[4802]: I1206 03:58:37.293410 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6f5b891-75b4-4b8c-95b1-7b92ee267793-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c6f5b891-75b4-4b8c-95b1-7b92ee267793" (UID: "c6f5b891-75b4-4b8c-95b1-7b92ee267793"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:58:37 crc kubenswrapper[4802]: I1206 03:58:37.332924 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6f5b891-75b4-4b8c-95b1-7b92ee267793-config-data" (OuterVolumeSpecName: "config-data") pod "c6f5b891-75b4-4b8c-95b1-7b92ee267793" (UID: "c6f5b891-75b4-4b8c-95b1-7b92ee267793"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:58:37 crc kubenswrapper[4802]: I1206 03:58:37.366530 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6f5b891-75b4-4b8c-95b1-7b92ee267793-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:37 crc kubenswrapper[4802]: I1206 03:58:37.366598 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5m9g\" (UniqueName: \"kubernetes.io/projected/c6f5b891-75b4-4b8c-95b1-7b92ee267793-kube-api-access-x5m9g\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:37 crc kubenswrapper[4802]: I1206 03:58:37.366615 4802 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6f5b891-75b4-4b8c-95b1-7b92ee267793-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:37 crc kubenswrapper[4802]: I1206 03:58:37.843980 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-mrxg9" event={"ID":"c6f5b891-75b4-4b8c-95b1-7b92ee267793","Type":"ContainerDied","Data":"be79cf42ee56ff8a94dc5994a49b498e922ade30f01ff048c8584d70d194354f"} Dec 06 03:58:37 crc kubenswrapper[4802]: I1206 03:58:37.844032 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="be79cf42ee56ff8a94dc5994a49b498e922ade30f01ff048c8584d70d194354f" Dec 06 03:58:37 crc kubenswrapper[4802]: I1206 03:58:37.844028 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-mrxg9" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.100035 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-w49ls"] Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.100536 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c79d794d7-w49ls" podUID="37b0a45e-9768-4b03-bf26-d35d3c98f58a" containerName="dnsmasq-dns" containerID="cri-o://99b3ae2fcbdcf8dd15fdf93bc19b0329b9e53253fa4267acb98239a199207902" gracePeriod=10 Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.104266 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c79d794d7-w49ls" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.127678 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-ddx8q"] Dec 06 03:58:38 crc kubenswrapper[4802]: E1206 03:58:38.130891 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6f5b891-75b4-4b8c-95b1-7b92ee267793" containerName="keystone-db-sync" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.130909 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6f5b891-75b4-4b8c-95b1-7b92ee267793" containerName="keystone-db-sync" Dec 06 03:58:38 crc kubenswrapper[4802]: E1206 03:58:38.130919 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7653cfc2-6e34-40f2-843a-9a644165b0fb" containerName="mariadb-account-create-update" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.130926 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="7653cfc2-6e34-40f2-843a-9a644165b0fb" containerName="mariadb-account-create-update" Dec 06 03:58:38 crc kubenswrapper[4802]: E1206 03:58:38.130943 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7d36678-2b9e-4380-8828-a32e159ce604" containerName="mariadb-database-create" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.130950 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7d36678-2b9e-4380-8828-a32e159ce604" containerName="mariadb-database-create" Dec 06 03:58:38 crc kubenswrapper[4802]: E1206 03:58:38.130969 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e975aca2-19e6-4a92-a4b0-f0db131bdfec" containerName="mariadb-database-create" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.130977 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="e975aca2-19e6-4a92-a4b0-f0db131bdfec" containerName="mariadb-database-create" Dec 06 03:58:38 crc kubenswrapper[4802]: E1206 03:58:38.131004 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebeea130-3942-475e-ba17-51d624cde585" containerName="mariadb-database-create" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.131013 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebeea130-3942-475e-ba17-51d624cde585" containerName="mariadb-database-create" Dec 06 03:58:38 crc kubenswrapper[4802]: E1206 03:58:38.131023 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e065079d-0d3d-40b5-8103-556cb4b7f338" containerName="mariadb-database-create" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.131032 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="e065079d-0d3d-40b5-8103-556cb4b7f338" containerName="mariadb-database-create" Dec 06 03:58:38 crc kubenswrapper[4802]: E1206 03:58:38.131183 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c4bfa7b-7d20-4c15-b4a7-181a05dcaab4" containerName="mariadb-account-create-update" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.131195 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c4bfa7b-7d20-4c15-b4a7-181a05dcaab4" containerName="mariadb-account-create-update" Dec 06 03:58:38 crc kubenswrapper[4802]: E1206 03:58:38.131208 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="180473dd-14d1-409d-8ca8-059c9883c762" containerName="mariadb-account-create-update" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.131216 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="180473dd-14d1-409d-8ca8-059c9883c762" containerName="mariadb-account-create-update" Dec 06 03:58:38 crc kubenswrapper[4802]: E1206 03:58:38.131230 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a076eb84-b8f5-4f75-922b-e44682ad8e67" containerName="mariadb-account-create-update" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.131238 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="a076eb84-b8f5-4f75-922b-e44682ad8e67" containerName="mariadb-account-create-update" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.131437 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="180473dd-14d1-409d-8ca8-059c9883c762" containerName="mariadb-account-create-update" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.131457 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="e975aca2-19e6-4a92-a4b0-f0db131bdfec" containerName="mariadb-database-create" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.131471 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="e065079d-0d3d-40b5-8103-556cb4b7f338" containerName="mariadb-database-create" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.131483 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6f5b891-75b4-4b8c-95b1-7b92ee267793" containerName="keystone-db-sync" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.131498 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="a076eb84-b8f5-4f75-922b-e44682ad8e67" containerName="mariadb-account-create-update" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.131510 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="7653cfc2-6e34-40f2-843a-9a644165b0fb" containerName="mariadb-account-create-update" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.131524 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c4bfa7b-7d20-4c15-b4a7-181a05dcaab4" containerName="mariadb-account-create-update" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.131533 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7d36678-2b9e-4380-8828-a32e159ce604" containerName="mariadb-database-create" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.131544 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebeea130-3942-475e-ba17-51d624cde585" containerName="mariadb-database-create" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.132239 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-ddx8q" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.143277 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.143375 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.143277 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.143582 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.143713 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-zk4kp" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.152036 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b868669f-shq86"] Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.154886 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b868669f-shq86" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.169162 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-ddx8q"] Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.181457 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b24d756b-e69d-42c8-8f6c-0ea2a3d57d00-scripts\") pod \"keystone-bootstrap-ddx8q\" (UID: \"b24d756b-e69d-42c8-8f6c-0ea2a3d57d00\") " pod="openstack/keystone-bootstrap-ddx8q" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.181509 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cr2zr\" (UniqueName: \"kubernetes.io/projected/b24d756b-e69d-42c8-8f6c-0ea2a3d57d00-kube-api-access-cr2zr\") pod \"keystone-bootstrap-ddx8q\" (UID: \"b24d756b-e69d-42c8-8f6c-0ea2a3d57d00\") " pod="openstack/keystone-bootstrap-ddx8q" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.181557 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b24d756b-e69d-42c8-8f6c-0ea2a3d57d00-credential-keys\") pod \"keystone-bootstrap-ddx8q\" (UID: \"b24d756b-e69d-42c8-8f6c-0ea2a3d57d00\") " pod="openstack/keystone-bootstrap-ddx8q" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.181595 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b24d756b-e69d-42c8-8f6c-0ea2a3d57d00-config-data\") pod \"keystone-bootstrap-ddx8q\" (UID: \"b24d756b-e69d-42c8-8f6c-0ea2a3d57d00\") " pod="openstack/keystone-bootstrap-ddx8q" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.181630 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b24d756b-e69d-42c8-8f6c-0ea2a3d57d00-fernet-keys\") pod \"keystone-bootstrap-ddx8q\" (UID: \"b24d756b-e69d-42c8-8f6c-0ea2a3d57d00\") " pod="openstack/keystone-bootstrap-ddx8q" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.181661 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b24d756b-e69d-42c8-8f6c-0ea2a3d57d00-combined-ca-bundle\") pod \"keystone-bootstrap-ddx8q\" (UID: \"b24d756b-e69d-42c8-8f6c-0ea2a3d57d00\") " pod="openstack/keystone-bootstrap-ddx8q" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.200117 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b868669f-shq86"] Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.250043 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-2hzlt"] Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.251371 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-2hzlt" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.260744 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-kxppf" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.261115 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.282744 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czslj\" (UniqueName: \"kubernetes.io/projected/32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4-kube-api-access-czslj\") pod \"dnsmasq-dns-5b868669f-shq86\" (UID: \"32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4\") " pod="openstack/dnsmasq-dns-5b868669f-shq86" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.282807 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b24d756b-e69d-42c8-8f6c-0ea2a3d57d00-scripts\") pod \"keystone-bootstrap-ddx8q\" (UID: \"b24d756b-e69d-42c8-8f6c-0ea2a3d57d00\") " pod="openstack/keystone-bootstrap-ddx8q" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.282857 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4-dns-svc\") pod \"dnsmasq-dns-5b868669f-shq86\" (UID: \"32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4\") " pod="openstack/dnsmasq-dns-5b868669f-shq86" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.282874 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4-ovsdbserver-sb\") pod \"dnsmasq-dns-5b868669f-shq86\" (UID: \"32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4\") " pod="openstack/dnsmasq-dns-5b868669f-shq86" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.282903 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cr2zr\" (UniqueName: \"kubernetes.io/projected/b24d756b-e69d-42c8-8f6c-0ea2a3d57d00-kube-api-access-cr2zr\") pod \"keystone-bootstrap-ddx8q\" (UID: \"b24d756b-e69d-42c8-8f6c-0ea2a3d57d00\") " pod="openstack/keystone-bootstrap-ddx8q" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.282922 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4-dns-swift-storage-0\") pod \"dnsmasq-dns-5b868669f-shq86\" (UID: \"32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4\") " pod="openstack/dnsmasq-dns-5b868669f-shq86" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.282938 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4-config\") pod \"dnsmasq-dns-5b868669f-shq86\" (UID: \"32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4\") " pod="openstack/dnsmasq-dns-5b868669f-shq86" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.282975 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b24d756b-e69d-42c8-8f6c-0ea2a3d57d00-credential-keys\") pod \"keystone-bootstrap-ddx8q\" (UID: \"b24d756b-e69d-42c8-8f6c-0ea2a3d57d00\") " pod="openstack/keystone-bootstrap-ddx8q" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.282998 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b24d756b-e69d-42c8-8f6c-0ea2a3d57d00-config-data\") pod \"keystone-bootstrap-ddx8q\" (UID: \"b24d756b-e69d-42c8-8f6c-0ea2a3d57d00\") " pod="openstack/keystone-bootstrap-ddx8q" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.283028 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b24d756b-e69d-42c8-8f6c-0ea2a3d57d00-fernet-keys\") pod \"keystone-bootstrap-ddx8q\" (UID: \"b24d756b-e69d-42c8-8f6c-0ea2a3d57d00\") " pod="openstack/keystone-bootstrap-ddx8q" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.283056 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b24d756b-e69d-42c8-8f6c-0ea2a3d57d00-combined-ca-bundle\") pod \"keystone-bootstrap-ddx8q\" (UID: \"b24d756b-e69d-42c8-8f6c-0ea2a3d57d00\") " pod="openstack/keystone-bootstrap-ddx8q" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.283085 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4-ovsdbserver-nb\") pod \"dnsmasq-dns-5b868669f-shq86\" (UID: \"32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4\") " pod="openstack/dnsmasq-dns-5b868669f-shq86" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.305803 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b24d756b-e69d-42c8-8f6c-0ea2a3d57d00-scripts\") pod \"keystone-bootstrap-ddx8q\" (UID: \"b24d756b-e69d-42c8-8f6c-0ea2a3d57d00\") " pod="openstack/keystone-bootstrap-ddx8q" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.305891 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b24d756b-e69d-42c8-8f6c-0ea2a3d57d00-combined-ca-bundle\") pod \"keystone-bootstrap-ddx8q\" (UID: \"b24d756b-e69d-42c8-8f6c-0ea2a3d57d00\") " pod="openstack/keystone-bootstrap-ddx8q" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.306316 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b24d756b-e69d-42c8-8f6c-0ea2a3d57d00-config-data\") pod \"keystone-bootstrap-ddx8q\" (UID: \"b24d756b-e69d-42c8-8f6c-0ea2a3d57d00\") " pod="openstack/keystone-bootstrap-ddx8q" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.310470 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b24d756b-e69d-42c8-8f6c-0ea2a3d57d00-credential-keys\") pod \"keystone-bootstrap-ddx8q\" (UID: \"b24d756b-e69d-42c8-8f6c-0ea2a3d57d00\") " pod="openstack/keystone-bootstrap-ddx8q" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.325209 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b24d756b-e69d-42c8-8f6c-0ea2a3d57d00-fernet-keys\") pod \"keystone-bootstrap-ddx8q\" (UID: \"b24d756b-e69d-42c8-8f6c-0ea2a3d57d00\") " pod="openstack/keystone-bootstrap-ddx8q" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.330112 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cr2zr\" (UniqueName: \"kubernetes.io/projected/b24d756b-e69d-42c8-8f6c-0ea2a3d57d00-kube-api-access-cr2zr\") pod \"keystone-bootstrap-ddx8q\" (UID: \"b24d756b-e69d-42c8-8f6c-0ea2a3d57d00\") " pod="openstack/keystone-bootstrap-ddx8q" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.392726 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czslj\" (UniqueName: \"kubernetes.io/projected/32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4-kube-api-access-czslj\") pod \"dnsmasq-dns-5b868669f-shq86\" (UID: \"32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4\") " pod="openstack/dnsmasq-dns-5b868669f-shq86" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.393018 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4-ovsdbserver-sb\") pod \"dnsmasq-dns-5b868669f-shq86\" (UID: \"32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4\") " pod="openstack/dnsmasq-dns-5b868669f-shq86" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.393039 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4-dns-svc\") pod \"dnsmasq-dns-5b868669f-shq86\" (UID: \"32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4\") " pod="openstack/dnsmasq-dns-5b868669f-shq86" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.393063 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4-dns-swift-storage-0\") pod \"dnsmasq-dns-5b868669f-shq86\" (UID: \"32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4\") " pod="openstack/dnsmasq-dns-5b868669f-shq86" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.393080 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4-config\") pod \"dnsmasq-dns-5b868669f-shq86\" (UID: \"32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4\") " pod="openstack/dnsmasq-dns-5b868669f-shq86" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.393107 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a639667-be04-429f-a2ce-b20a011425f8-combined-ca-bundle\") pod \"heat-db-sync-2hzlt\" (UID: \"2a639667-be04-429f-a2ce-b20a011425f8\") " pod="openstack/heat-db-sync-2hzlt" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.393159 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a639667-be04-429f-a2ce-b20a011425f8-config-data\") pod \"heat-db-sync-2hzlt\" (UID: \"2a639667-be04-429f-a2ce-b20a011425f8\") " pod="openstack/heat-db-sync-2hzlt" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.393205 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ssjb\" (UniqueName: \"kubernetes.io/projected/2a639667-be04-429f-a2ce-b20a011425f8-kube-api-access-6ssjb\") pod \"heat-db-sync-2hzlt\" (UID: \"2a639667-be04-429f-a2ce-b20a011425f8\") " pod="openstack/heat-db-sync-2hzlt" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.393225 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4-ovsdbserver-nb\") pod \"dnsmasq-dns-5b868669f-shq86\" (UID: \"32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4\") " pod="openstack/dnsmasq-dns-5b868669f-shq86" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.394043 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4-ovsdbserver-nb\") pod \"dnsmasq-dns-5b868669f-shq86\" (UID: \"32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4\") " pod="openstack/dnsmasq-dns-5b868669f-shq86" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.394078 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4-config\") pod \"dnsmasq-dns-5b868669f-shq86\" (UID: \"32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4\") " pod="openstack/dnsmasq-dns-5b868669f-shq86" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.394274 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-2hzlt"] Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.394556 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4-dns-svc\") pod \"dnsmasq-dns-5b868669f-shq86\" (UID: \"32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4\") " pod="openstack/dnsmasq-dns-5b868669f-shq86" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.394608 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4-dns-swift-storage-0\") pod \"dnsmasq-dns-5b868669f-shq86\" (UID: \"32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4\") " pod="openstack/dnsmasq-dns-5b868669f-shq86" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.394830 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4-ovsdbserver-sb\") pod \"dnsmasq-dns-5b868669f-shq86\" (UID: \"32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4\") " pod="openstack/dnsmasq-dns-5b868669f-shq86" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.436621 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czslj\" (UniqueName: \"kubernetes.io/projected/32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4-kube-api-access-czslj\") pod \"dnsmasq-dns-5b868669f-shq86\" (UID: \"32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4\") " pod="openstack/dnsmasq-dns-5b868669f-shq86" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.469642 4802 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5c79d794d7-w49ls" podUID="37b0a45e-9768-4b03-bf26-d35d3c98f58a" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.132:5353: connect: connection refused" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.483855 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-ddx8q" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.487041 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b868669f-shq86"] Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.487798 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b868669f-shq86" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.495664 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a639667-be04-429f-a2ce-b20a011425f8-combined-ca-bundle\") pod \"heat-db-sync-2hzlt\" (UID: \"2a639667-be04-429f-a2ce-b20a011425f8\") " pod="openstack/heat-db-sync-2hzlt" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.495768 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a639667-be04-429f-a2ce-b20a011425f8-config-data\") pod \"heat-db-sync-2hzlt\" (UID: \"2a639667-be04-429f-a2ce-b20a011425f8\") " pod="openstack/heat-db-sync-2hzlt" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.495854 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ssjb\" (UniqueName: \"kubernetes.io/projected/2a639667-be04-429f-a2ce-b20a011425f8-kube-api-access-6ssjb\") pod \"heat-db-sync-2hzlt\" (UID: \"2a639667-be04-429f-a2ce-b20a011425f8\") " pod="openstack/heat-db-sync-2hzlt" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.505393 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a639667-be04-429f-a2ce-b20a011425f8-combined-ca-bundle\") pod \"heat-db-sync-2hzlt\" (UID: \"2a639667-be04-429f-a2ce-b20a011425f8\") " pod="openstack/heat-db-sync-2hzlt" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.506914 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a639667-be04-429f-a2ce-b20a011425f8-config-data\") pod \"heat-db-sync-2hzlt\" (UID: \"2a639667-be04-429f-a2ce-b20a011425f8\") " pod="openstack/heat-db-sync-2hzlt" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.537256 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-mlhfp" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.573945 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-qs8hv"] Dec 06 03:58:38 crc kubenswrapper[4802]: E1206 03:58:38.574313 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d947661-37cd-49c2-bd89-e927e4034ba4" containerName="glance-db-sync" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.574325 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d947661-37cd-49c2-bd89-e927e4034ba4" containerName="glance-db-sync" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.574477 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d947661-37cd-49c2-bd89-e927e4034ba4" containerName="glance-db-sync" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.575122 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-qs8hv" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.591252 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ssjb\" (UniqueName: \"kubernetes.io/projected/2a639667-be04-429f-a2ce-b20a011425f8-kube-api-access-6ssjb\") pod \"heat-db-sync-2hzlt\" (UID: \"2a639667-be04-429f-a2ce-b20a011425f8\") " pod="openstack/heat-db-sync-2hzlt" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.596617 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xt576\" (UniqueName: \"kubernetes.io/projected/5d947661-37cd-49c2-bd89-e927e4034ba4-kube-api-access-xt576\") pod \"5d947661-37cd-49c2-bd89-e927e4034ba4\" (UID: \"5d947661-37cd-49c2-bd89-e927e4034ba4\") " Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.596672 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5d947661-37cd-49c2-bd89-e927e4034ba4-db-sync-config-data\") pod \"5d947661-37cd-49c2-bd89-e927e4034ba4\" (UID: \"5d947661-37cd-49c2-bd89-e927e4034ba4\") " Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.597078 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d947661-37cd-49c2-bd89-e927e4034ba4-combined-ca-bundle\") pod \"5d947661-37cd-49c2-bd89-e927e4034ba4\" (UID: \"5d947661-37cd-49c2-bd89-e927e4034ba4\") " Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.597114 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d947661-37cd-49c2-bd89-e927e4034ba4-config-data\") pod \"5d947661-37cd-49c2-bd89-e927e4034ba4\" (UID: \"5d947661-37cd-49c2-bd89-e927e4034ba4\") " Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.600007 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.600717 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.600948 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-nr9zb" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.621953 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d947661-37cd-49c2-bd89-e927e4034ba4-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "5d947661-37cd-49c2-bd89-e927e4034ba4" (UID: "5d947661-37cd-49c2-bd89-e927e4034ba4"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.628869 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-qs8hv"] Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.632022 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d947661-37cd-49c2-bd89-e927e4034ba4-kube-api-access-xt576" (OuterVolumeSpecName: "kube-api-access-xt576") pod "5d947661-37cd-49c2-bd89-e927e4034ba4" (UID: "5d947661-37cd-49c2-bd89-e927e4034ba4"). InnerVolumeSpecName "kube-api-access-xt576". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.697899 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-5hj5p"] Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.699050 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-5hj5p" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.699058 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d947661-37cd-49c2-bd89-e927e4034ba4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5d947661-37cd-49c2-bd89-e927e4034ba4" (UID: "5d947661-37cd-49c2-bd89-e927e4034ba4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.728998 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8c25c327-e3cd-4f28-855a-249e4bb0c5f6-config\") pod \"neutron-db-sync-qs8hv\" (UID: \"8c25c327-e3cd-4f28-855a-249e4bb0c5f6\") " pod="openstack/neutron-db-sync-qs8hv" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.729058 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c25c327-e3cd-4f28-855a-249e4bb0c5f6-combined-ca-bundle\") pod \"neutron-db-sync-qs8hv\" (UID: \"8c25c327-e3cd-4f28-855a-249e4bb0c5f6\") " pod="openstack/neutron-db-sync-qs8hv" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.729121 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lt8jq\" (UniqueName: \"kubernetes.io/projected/8c25c327-e3cd-4f28-855a-249e4bb0c5f6-kube-api-access-lt8jq\") pod \"neutron-db-sync-qs8hv\" (UID: \"8c25c327-e3cd-4f28-855a-249e4bb0c5f6\") " pod="openstack/neutron-db-sync-qs8hv" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.729577 4802 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d947661-37cd-49c2-bd89-e927e4034ba4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.729594 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xt576\" (UniqueName: \"kubernetes.io/projected/5d947661-37cd-49c2-bd89-e927e4034ba4-kube-api-access-xt576\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.729605 4802 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5d947661-37cd-49c2-bd89-e927e4034ba4-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.736443 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-5k5s6" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.744958 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-4gbch"] Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.750962 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.760199 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.763336 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf78879c9-4gbch" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.808235 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d947661-37cd-49c2-bd89-e927e4034ba4-config-data" (OuterVolumeSpecName: "config-data") pod "5d947661-37cd-49c2-bd89-e927e4034ba4" (UID: "5d947661-37cd-49c2-bd89-e927e4034ba4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.813611 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-2hzlt" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.823279 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-5hj5p"] Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.832464 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lt8jq\" (UniqueName: \"kubernetes.io/projected/8c25c327-e3cd-4f28-855a-249e4bb0c5f6-kube-api-access-lt8jq\") pod \"neutron-db-sync-qs8hv\" (UID: \"8c25c327-e3cd-4f28-855a-249e4bb0c5f6\") " pod="openstack/neutron-db-sync-qs8hv" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.832553 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/21cae3ac-9ae8-4c31-972d-7a3c25e8bd95-dns-svc\") pod \"dnsmasq-dns-cf78879c9-4gbch\" (UID: \"21cae3ac-9ae8-4c31-972d-7a3c25e8bd95\") " pod="openstack/dnsmasq-dns-cf78879c9-4gbch" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.832583 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/21cae3ac-9ae8-4c31-972d-7a3c25e8bd95-dns-swift-storage-0\") pod \"dnsmasq-dns-cf78879c9-4gbch\" (UID: \"21cae3ac-9ae8-4c31-972d-7a3c25e8bd95\") " pod="openstack/dnsmasq-dns-cf78879c9-4gbch" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.832723 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/21cae3ac-9ae8-4c31-972d-7a3c25e8bd95-ovsdbserver-sb\") pod \"dnsmasq-dns-cf78879c9-4gbch\" (UID: \"21cae3ac-9ae8-4c31-972d-7a3c25e8bd95\") " pod="openstack/dnsmasq-dns-cf78879c9-4gbch" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.832819 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f1644907-b4b1-4e23-a273-19d535bd7f88-etc-machine-id\") pod \"cinder-db-sync-5hj5p\" (UID: \"f1644907-b4b1-4e23-a273-19d535bd7f88\") " pod="openstack/cinder-db-sync-5hj5p" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.832932 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1644907-b4b1-4e23-a273-19d535bd7f88-combined-ca-bundle\") pod \"cinder-db-sync-5hj5p\" (UID: \"f1644907-b4b1-4e23-a273-19d535bd7f88\") " pod="openstack/cinder-db-sync-5hj5p" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.833069 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f1644907-b4b1-4e23-a273-19d535bd7f88-db-sync-config-data\") pod \"cinder-db-sync-5hj5p\" (UID: \"f1644907-b4b1-4e23-a273-19d535bd7f88\") " pod="openstack/cinder-db-sync-5hj5p" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.855731 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21cae3ac-9ae8-4c31-972d-7a3c25e8bd95-config\") pod \"dnsmasq-dns-cf78879c9-4gbch\" (UID: \"21cae3ac-9ae8-4c31-972d-7a3c25e8bd95\") " pod="openstack/dnsmasq-dns-cf78879c9-4gbch" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.855769 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4c4tx\" (UniqueName: \"kubernetes.io/projected/21cae3ac-9ae8-4c31-972d-7a3c25e8bd95-kube-api-access-4c4tx\") pod \"dnsmasq-dns-cf78879c9-4gbch\" (UID: \"21cae3ac-9ae8-4c31-972d-7a3c25e8bd95\") " pod="openstack/dnsmasq-dns-cf78879c9-4gbch" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.855787 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pln97\" (UniqueName: \"kubernetes.io/projected/f1644907-b4b1-4e23-a273-19d535bd7f88-kube-api-access-pln97\") pod \"cinder-db-sync-5hj5p\" (UID: \"f1644907-b4b1-4e23-a273-19d535bd7f88\") " pod="openstack/cinder-db-sync-5hj5p" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.855821 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/21cae3ac-9ae8-4c31-972d-7a3c25e8bd95-ovsdbserver-nb\") pod \"dnsmasq-dns-cf78879c9-4gbch\" (UID: \"21cae3ac-9ae8-4c31-972d-7a3c25e8bd95\") " pod="openstack/dnsmasq-dns-cf78879c9-4gbch" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.855870 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f1644907-b4b1-4e23-a273-19d535bd7f88-scripts\") pod \"cinder-db-sync-5hj5p\" (UID: \"f1644907-b4b1-4e23-a273-19d535bd7f88\") " pod="openstack/cinder-db-sync-5hj5p" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.855971 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8c25c327-e3cd-4f28-855a-249e4bb0c5f6-config\") pod \"neutron-db-sync-qs8hv\" (UID: \"8c25c327-e3cd-4f28-855a-249e4bb0c5f6\") " pod="openstack/neutron-db-sync-qs8hv" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.855992 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c25c327-e3cd-4f28-855a-249e4bb0c5f6-combined-ca-bundle\") pod \"neutron-db-sync-qs8hv\" (UID: \"8c25c327-e3cd-4f28-855a-249e4bb0c5f6\") " pod="openstack/neutron-db-sync-qs8hv" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.856024 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1644907-b4b1-4e23-a273-19d535bd7f88-config-data\") pod \"cinder-db-sync-5hj5p\" (UID: \"f1644907-b4b1-4e23-a273-19d535bd7f88\") " pod="openstack/cinder-db-sync-5hj5p" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.856169 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d947661-37cd-49c2-bd89-e927e4034ba4-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.858466 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-qtg4k"] Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.859766 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-qtg4k" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.862006 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c25c327-e3cd-4f28-855a-249e4bb0c5f6-combined-ca-bundle\") pod \"neutron-db-sync-qs8hv\" (UID: \"8c25c327-e3cd-4f28-855a-249e4bb0c5f6\") " pod="openstack/neutron-db-sync-qs8hv" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.864781 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-qfhm9" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.867981 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/8c25c327-e3cd-4f28-855a-249e4bb0c5f6-config\") pod \"neutron-db-sync-qs8hv\" (UID: \"8c25c327-e3cd-4f28-855a-249e4bb0c5f6\") " pod="openstack/neutron-db-sync-qs8hv" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.868431 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.917648 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-dxvz6"] Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.919029 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-dxvz6" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.924350 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lt8jq\" (UniqueName: \"kubernetes.io/projected/8c25c327-e3cd-4f28-855a-249e4bb0c5f6-kube-api-access-lt8jq\") pod \"neutron-db-sync-qs8hv\" (UID: \"8c25c327-e3cd-4f28-855a-249e4bb0c5f6\") " pod="openstack/neutron-db-sync-qs8hv" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.924675 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.924999 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-xvxk2" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.925163 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.935526 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-mlhfp" event={"ID":"5d947661-37cd-49c2-bd89-e927e4034ba4","Type":"ContainerDied","Data":"ef95b6034d02153fe5279d159ad7e5567b7b3122f29307525c7760694c91ff4f"} Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.935568 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ef95b6034d02153fe5279d159ad7e5567b7b3122f29307525c7760694c91ff4f" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.935645 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-mlhfp" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.945824 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-4gbch"] Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.952645 4802 generic.go:334] "Generic (PLEG): container finished" podID="37b0a45e-9768-4b03-bf26-d35d3c98f58a" containerID="99b3ae2fcbdcf8dd15fdf93bc19b0329b9e53253fa4267acb98239a199207902" exitCode=0 Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.952707 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-w49ls" event={"ID":"37b0a45e-9768-4b03-bf26-d35d3c98f58a","Type":"ContainerDied","Data":"99b3ae2fcbdcf8dd15fdf93bc19b0329b9e53253fa4267acb98239a199207902"} Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.957284 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-qs8hv" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.957512 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fzfx\" (UniqueName: \"kubernetes.io/projected/5c636931-d435-47ec-b770-80dc1ee60756-kube-api-access-4fzfx\") pod \"barbican-db-sync-qtg4k\" (UID: \"5c636931-d435-47ec-b770-80dc1ee60756\") " pod="openstack/barbican-db-sync-qtg4k" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.957550 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f1644907-b4b1-4e23-a273-19d535bd7f88-etc-machine-id\") pod \"cinder-db-sync-5hj5p\" (UID: \"f1644907-b4b1-4e23-a273-19d535bd7f88\") " pod="openstack/cinder-db-sync-5hj5p" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.957585 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1644907-b4b1-4e23-a273-19d535bd7f88-combined-ca-bundle\") pod \"cinder-db-sync-5hj5p\" (UID: \"f1644907-b4b1-4e23-a273-19d535bd7f88\") " pod="openstack/cinder-db-sync-5hj5p" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.957614 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f1644907-b4b1-4e23-a273-19d535bd7f88-db-sync-config-data\") pod \"cinder-db-sync-5hj5p\" (UID: \"f1644907-b4b1-4e23-a273-19d535bd7f88\") " pod="openstack/cinder-db-sync-5hj5p" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.957640 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21cae3ac-9ae8-4c31-972d-7a3c25e8bd95-config\") pod \"dnsmasq-dns-cf78879c9-4gbch\" (UID: \"21cae3ac-9ae8-4c31-972d-7a3c25e8bd95\") " pod="openstack/dnsmasq-dns-cf78879c9-4gbch" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.957656 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4c4tx\" (UniqueName: \"kubernetes.io/projected/21cae3ac-9ae8-4c31-972d-7a3c25e8bd95-kube-api-access-4c4tx\") pod \"dnsmasq-dns-cf78879c9-4gbch\" (UID: \"21cae3ac-9ae8-4c31-972d-7a3c25e8bd95\") " pod="openstack/dnsmasq-dns-cf78879c9-4gbch" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.957669 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pln97\" (UniqueName: \"kubernetes.io/projected/f1644907-b4b1-4e23-a273-19d535bd7f88-kube-api-access-pln97\") pod \"cinder-db-sync-5hj5p\" (UID: \"f1644907-b4b1-4e23-a273-19d535bd7f88\") " pod="openstack/cinder-db-sync-5hj5p" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.957688 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/21cae3ac-9ae8-4c31-972d-7a3c25e8bd95-ovsdbserver-nb\") pod \"dnsmasq-dns-cf78879c9-4gbch\" (UID: \"21cae3ac-9ae8-4c31-972d-7a3c25e8bd95\") " pod="openstack/dnsmasq-dns-cf78879c9-4gbch" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.957712 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f1644907-b4b1-4e23-a273-19d535bd7f88-scripts\") pod \"cinder-db-sync-5hj5p\" (UID: \"f1644907-b4b1-4e23-a273-19d535bd7f88\") " pod="openstack/cinder-db-sync-5hj5p" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.957730 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c636931-d435-47ec-b770-80dc1ee60756-combined-ca-bundle\") pod \"barbican-db-sync-qtg4k\" (UID: \"5c636931-d435-47ec-b770-80dc1ee60756\") " pod="openstack/barbican-db-sync-qtg4k" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.957793 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1644907-b4b1-4e23-a273-19d535bd7f88-config-data\") pod \"cinder-db-sync-5hj5p\" (UID: \"f1644907-b4b1-4e23-a273-19d535bd7f88\") " pod="openstack/cinder-db-sync-5hj5p" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.957820 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/21cae3ac-9ae8-4c31-972d-7a3c25e8bd95-dns-svc\") pod \"dnsmasq-dns-cf78879c9-4gbch\" (UID: \"21cae3ac-9ae8-4c31-972d-7a3c25e8bd95\") " pod="openstack/dnsmasq-dns-cf78879c9-4gbch" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.957835 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/21cae3ac-9ae8-4c31-972d-7a3c25e8bd95-dns-swift-storage-0\") pod \"dnsmasq-dns-cf78879c9-4gbch\" (UID: \"21cae3ac-9ae8-4c31-972d-7a3c25e8bd95\") " pod="openstack/dnsmasq-dns-cf78879c9-4gbch" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.957863 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/21cae3ac-9ae8-4c31-972d-7a3c25e8bd95-ovsdbserver-sb\") pod \"dnsmasq-dns-cf78879c9-4gbch\" (UID: \"21cae3ac-9ae8-4c31-972d-7a3c25e8bd95\") " pod="openstack/dnsmasq-dns-cf78879c9-4gbch" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.957900 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5c636931-d435-47ec-b770-80dc1ee60756-db-sync-config-data\") pod \"barbican-db-sync-qtg4k\" (UID: \"5c636931-d435-47ec-b770-80dc1ee60756\") " pod="openstack/barbican-db-sync-qtg4k" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.958102 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f1644907-b4b1-4e23-a273-19d535bd7f88-etc-machine-id\") pod \"cinder-db-sync-5hj5p\" (UID: \"f1644907-b4b1-4e23-a273-19d535bd7f88\") " pod="openstack/cinder-db-sync-5hj5p" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.958970 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/21cae3ac-9ae8-4c31-972d-7a3c25e8bd95-ovsdbserver-nb\") pod \"dnsmasq-dns-cf78879c9-4gbch\" (UID: \"21cae3ac-9ae8-4c31-972d-7a3c25e8bd95\") " pod="openstack/dnsmasq-dns-cf78879c9-4gbch" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.959843 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21cae3ac-9ae8-4c31-972d-7a3c25e8bd95-config\") pod \"dnsmasq-dns-cf78879c9-4gbch\" (UID: \"21cae3ac-9ae8-4c31-972d-7a3c25e8bd95\") " pod="openstack/dnsmasq-dns-cf78879c9-4gbch" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.961526 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/21cae3ac-9ae8-4c31-972d-7a3c25e8bd95-dns-svc\") pod \"dnsmasq-dns-cf78879c9-4gbch\" (UID: \"21cae3ac-9ae8-4c31-972d-7a3c25e8bd95\") " pod="openstack/dnsmasq-dns-cf78879c9-4gbch" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.961548 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/21cae3ac-9ae8-4c31-972d-7a3c25e8bd95-dns-swift-storage-0\") pod \"dnsmasq-dns-cf78879c9-4gbch\" (UID: \"21cae3ac-9ae8-4c31-972d-7a3c25e8bd95\") " pod="openstack/dnsmasq-dns-cf78879c9-4gbch" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.964742 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1644907-b4b1-4e23-a273-19d535bd7f88-config-data\") pod \"cinder-db-sync-5hj5p\" (UID: \"f1644907-b4b1-4e23-a273-19d535bd7f88\") " pod="openstack/cinder-db-sync-5hj5p" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.973000 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f1644907-b4b1-4e23-a273-19d535bd7f88-db-sync-config-data\") pod \"cinder-db-sync-5hj5p\" (UID: \"f1644907-b4b1-4e23-a273-19d535bd7f88\") " pod="openstack/cinder-db-sync-5hj5p" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.975760 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/21cae3ac-9ae8-4c31-972d-7a3c25e8bd95-ovsdbserver-sb\") pod \"dnsmasq-dns-cf78879c9-4gbch\" (UID: \"21cae3ac-9ae8-4c31-972d-7a3c25e8bd95\") " pod="openstack/dnsmasq-dns-cf78879c9-4gbch" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.976667 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f1644907-b4b1-4e23-a273-19d535bd7f88-scripts\") pod \"cinder-db-sync-5hj5p\" (UID: \"f1644907-b4b1-4e23-a273-19d535bd7f88\") " pod="openstack/cinder-db-sync-5hj5p" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.980641 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1644907-b4b1-4e23-a273-19d535bd7f88-combined-ca-bundle\") pod \"cinder-db-sync-5hj5p\" (UID: \"f1644907-b4b1-4e23-a273-19d535bd7f88\") " pod="openstack/cinder-db-sync-5hj5p" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.986793 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pln97\" (UniqueName: \"kubernetes.io/projected/f1644907-b4b1-4e23-a273-19d535bd7f88-kube-api-access-pln97\") pod \"cinder-db-sync-5hj5p\" (UID: \"f1644907-b4b1-4e23-a273-19d535bd7f88\") " pod="openstack/cinder-db-sync-5hj5p" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.988294 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-qtg4k"] Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.993049 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4c4tx\" (UniqueName: \"kubernetes.io/projected/21cae3ac-9ae8-4c31-972d-7a3c25e8bd95-kube-api-access-4c4tx\") pod \"dnsmasq-dns-cf78879c9-4gbch\" (UID: \"21cae3ac-9ae8-4c31-972d-7a3c25e8bd95\") " pod="openstack/dnsmasq-dns-cf78879c9-4gbch" Dec 06 03:58:38 crc kubenswrapper[4802]: I1206 03:58:38.996122 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-dxvz6"] Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.004297 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.007478 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.010133 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-w49ls" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.012620 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.012846 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.027943 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.061998 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4abd9ed3-07e6-46a6-8539-85fc5c624145-combined-ca-bundle\") pod \"placement-db-sync-dxvz6\" (UID: \"4abd9ed3-07e6-46a6-8539-85fc5c624145\") " pod="openstack/placement-db-sync-dxvz6" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.062086 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4abd9ed3-07e6-46a6-8539-85fc5c624145-config-data\") pod \"placement-db-sync-dxvz6\" (UID: \"4abd9ed3-07e6-46a6-8539-85fc5c624145\") " pod="openstack/placement-db-sync-dxvz6" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.062111 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c636931-d435-47ec-b770-80dc1ee60756-combined-ca-bundle\") pod \"barbican-db-sync-qtg4k\" (UID: \"5c636931-d435-47ec-b770-80dc1ee60756\") " pod="openstack/barbican-db-sync-qtg4k" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.062144 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h57kj\" (UniqueName: \"kubernetes.io/projected/4abd9ed3-07e6-46a6-8539-85fc5c624145-kube-api-access-h57kj\") pod \"placement-db-sync-dxvz6\" (UID: \"4abd9ed3-07e6-46a6-8539-85fc5c624145\") " pod="openstack/placement-db-sync-dxvz6" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.062173 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4abd9ed3-07e6-46a6-8539-85fc5c624145-logs\") pod \"placement-db-sync-dxvz6\" (UID: \"4abd9ed3-07e6-46a6-8539-85fc5c624145\") " pod="openstack/placement-db-sync-dxvz6" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.062231 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4abd9ed3-07e6-46a6-8539-85fc5c624145-scripts\") pod \"placement-db-sync-dxvz6\" (UID: \"4abd9ed3-07e6-46a6-8539-85fc5c624145\") " pod="openstack/placement-db-sync-dxvz6" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.062248 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5c636931-d435-47ec-b770-80dc1ee60756-db-sync-config-data\") pod \"barbican-db-sync-qtg4k\" (UID: \"5c636931-d435-47ec-b770-80dc1ee60756\") " pod="openstack/barbican-db-sync-qtg4k" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.062271 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fzfx\" (UniqueName: \"kubernetes.io/projected/5c636931-d435-47ec-b770-80dc1ee60756-kube-api-access-4fzfx\") pod \"barbican-db-sync-qtg4k\" (UID: \"5c636931-d435-47ec-b770-80dc1ee60756\") " pod="openstack/barbican-db-sync-qtg4k" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.073788 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c636931-d435-47ec-b770-80dc1ee60756-combined-ca-bundle\") pod \"barbican-db-sync-qtg4k\" (UID: \"5c636931-d435-47ec-b770-80dc1ee60756\") " pod="openstack/barbican-db-sync-qtg4k" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.074429 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5c636931-d435-47ec-b770-80dc1ee60756-db-sync-config-data\") pod \"barbican-db-sync-qtg4k\" (UID: \"5c636931-d435-47ec-b770-80dc1ee60756\") " pod="openstack/barbican-db-sync-qtg4k" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.086514 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-5hj5p" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.098486 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fzfx\" (UniqueName: \"kubernetes.io/projected/5c636931-d435-47ec-b770-80dc1ee60756-kube-api-access-4fzfx\") pod \"barbican-db-sync-qtg4k\" (UID: \"5c636931-d435-47ec-b770-80dc1ee60756\") " pod="openstack/barbican-db-sync-qtg4k" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.111358 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf78879c9-4gbch" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.164189 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/37b0a45e-9768-4b03-bf26-d35d3c98f58a-ovsdbserver-sb\") pod \"37b0a45e-9768-4b03-bf26-d35d3c98f58a\" (UID: \"37b0a45e-9768-4b03-bf26-d35d3c98f58a\") " Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.164250 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mhwmh\" (UniqueName: \"kubernetes.io/projected/37b0a45e-9768-4b03-bf26-d35d3c98f58a-kube-api-access-mhwmh\") pod \"37b0a45e-9768-4b03-bf26-d35d3c98f58a\" (UID: \"37b0a45e-9768-4b03-bf26-d35d3c98f58a\") " Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.164359 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/37b0a45e-9768-4b03-bf26-d35d3c98f58a-dns-swift-storage-0\") pod \"37b0a45e-9768-4b03-bf26-d35d3c98f58a\" (UID: \"37b0a45e-9768-4b03-bf26-d35d3c98f58a\") " Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.164468 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/37b0a45e-9768-4b03-bf26-d35d3c98f58a-ovsdbserver-nb\") pod \"37b0a45e-9768-4b03-bf26-d35d3c98f58a\" (UID: \"37b0a45e-9768-4b03-bf26-d35d3c98f58a\") " Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.164499 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/37b0a45e-9768-4b03-bf26-d35d3c98f58a-dns-svc\") pod \"37b0a45e-9768-4b03-bf26-d35d3c98f58a\" (UID: \"37b0a45e-9768-4b03-bf26-d35d3c98f58a\") " Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.164523 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37b0a45e-9768-4b03-bf26-d35d3c98f58a-config\") pod \"37b0a45e-9768-4b03-bf26-d35d3c98f58a\" (UID: \"37b0a45e-9768-4b03-bf26-d35d3c98f58a\") " Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.164743 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4d4f4fff-a6ec-4fef-82b4-6ba392fd297c-run-httpd\") pod \"ceilometer-0\" (UID: \"4d4f4fff-a6ec-4fef-82b4-6ba392fd297c\") " pod="openstack/ceilometer-0" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.164777 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d4f4fff-a6ec-4fef-82b4-6ba392fd297c-config-data\") pod \"ceilometer-0\" (UID: \"4d4f4fff-a6ec-4fef-82b4-6ba392fd297c\") " pod="openstack/ceilometer-0" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.164820 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4abd9ed3-07e6-46a6-8539-85fc5c624145-config-data\") pod \"placement-db-sync-dxvz6\" (UID: \"4abd9ed3-07e6-46a6-8539-85fc5c624145\") " pod="openstack/placement-db-sync-dxvz6" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.164836 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4d4f4fff-a6ec-4fef-82b4-6ba392fd297c-log-httpd\") pod \"ceilometer-0\" (UID: \"4d4f4fff-a6ec-4fef-82b4-6ba392fd297c\") " pod="openstack/ceilometer-0" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.164874 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h57kj\" (UniqueName: \"kubernetes.io/projected/4abd9ed3-07e6-46a6-8539-85fc5c624145-kube-api-access-h57kj\") pod \"placement-db-sync-dxvz6\" (UID: \"4abd9ed3-07e6-46a6-8539-85fc5c624145\") " pod="openstack/placement-db-sync-dxvz6" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.164903 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4abd9ed3-07e6-46a6-8539-85fc5c624145-logs\") pod \"placement-db-sync-dxvz6\" (UID: \"4abd9ed3-07e6-46a6-8539-85fc5c624145\") " pod="openstack/placement-db-sync-dxvz6" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.164928 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4d4f4fff-a6ec-4fef-82b4-6ba392fd297c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4d4f4fff-a6ec-4fef-82b4-6ba392fd297c\") " pod="openstack/ceilometer-0" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.164965 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4abd9ed3-07e6-46a6-8539-85fc5c624145-scripts\") pod \"placement-db-sync-dxvz6\" (UID: \"4abd9ed3-07e6-46a6-8539-85fc5c624145\") " pod="openstack/placement-db-sync-dxvz6" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.164981 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjnp6\" (UniqueName: \"kubernetes.io/projected/4d4f4fff-a6ec-4fef-82b4-6ba392fd297c-kube-api-access-fjnp6\") pod \"ceilometer-0\" (UID: \"4d4f4fff-a6ec-4fef-82b4-6ba392fd297c\") " pod="openstack/ceilometer-0" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.165003 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d4f4fff-a6ec-4fef-82b4-6ba392fd297c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4d4f4fff-a6ec-4fef-82b4-6ba392fd297c\") " pod="openstack/ceilometer-0" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.165029 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4abd9ed3-07e6-46a6-8539-85fc5c624145-combined-ca-bundle\") pod \"placement-db-sync-dxvz6\" (UID: \"4abd9ed3-07e6-46a6-8539-85fc5c624145\") " pod="openstack/placement-db-sync-dxvz6" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.165050 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d4f4fff-a6ec-4fef-82b4-6ba392fd297c-scripts\") pod \"ceilometer-0\" (UID: \"4d4f4fff-a6ec-4fef-82b4-6ba392fd297c\") " pod="openstack/ceilometer-0" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.170084 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4abd9ed3-07e6-46a6-8539-85fc5c624145-logs\") pod \"placement-db-sync-dxvz6\" (UID: \"4abd9ed3-07e6-46a6-8539-85fc5c624145\") " pod="openstack/placement-db-sync-dxvz6" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.181158 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4abd9ed3-07e6-46a6-8539-85fc5c624145-scripts\") pod \"placement-db-sync-dxvz6\" (UID: \"4abd9ed3-07e6-46a6-8539-85fc5c624145\") " pod="openstack/placement-db-sync-dxvz6" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.188451 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4abd9ed3-07e6-46a6-8539-85fc5c624145-combined-ca-bundle\") pod \"placement-db-sync-dxvz6\" (UID: \"4abd9ed3-07e6-46a6-8539-85fc5c624145\") " pod="openstack/placement-db-sync-dxvz6" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.200116 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4abd9ed3-07e6-46a6-8539-85fc5c624145-config-data\") pod \"placement-db-sync-dxvz6\" (UID: \"4abd9ed3-07e6-46a6-8539-85fc5c624145\") " pod="openstack/placement-db-sync-dxvz6" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.214288 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h57kj\" (UniqueName: \"kubernetes.io/projected/4abd9ed3-07e6-46a6-8539-85fc5c624145-kube-api-access-h57kj\") pod \"placement-db-sync-dxvz6\" (UID: \"4abd9ed3-07e6-46a6-8539-85fc5c624145\") " pod="openstack/placement-db-sync-dxvz6" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.215058 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37b0a45e-9768-4b03-bf26-d35d3c98f58a-kube-api-access-mhwmh" (OuterVolumeSpecName: "kube-api-access-mhwmh") pod "37b0a45e-9768-4b03-bf26-d35d3c98f58a" (UID: "37b0a45e-9768-4b03-bf26-d35d3c98f58a"). InnerVolumeSpecName "kube-api-access-mhwmh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.243227 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-qtg4k" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.253996 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-4gbch"] Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.267215 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d4f4fff-a6ec-4fef-82b4-6ba392fd297c-scripts\") pod \"ceilometer-0\" (UID: \"4d4f4fff-a6ec-4fef-82b4-6ba392fd297c\") " pod="openstack/ceilometer-0" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.267271 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4d4f4fff-a6ec-4fef-82b4-6ba392fd297c-run-httpd\") pod \"ceilometer-0\" (UID: \"4d4f4fff-a6ec-4fef-82b4-6ba392fd297c\") " pod="openstack/ceilometer-0" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.270578 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4d4f4fff-a6ec-4fef-82b4-6ba392fd297c-run-httpd\") pod \"ceilometer-0\" (UID: \"4d4f4fff-a6ec-4fef-82b4-6ba392fd297c\") " pod="openstack/ceilometer-0" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.270785 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d4f4fff-a6ec-4fef-82b4-6ba392fd297c-config-data\") pod \"ceilometer-0\" (UID: \"4d4f4fff-a6ec-4fef-82b4-6ba392fd297c\") " pod="openstack/ceilometer-0" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.270892 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4d4f4fff-a6ec-4fef-82b4-6ba392fd297c-log-httpd\") pod \"ceilometer-0\" (UID: \"4d4f4fff-a6ec-4fef-82b4-6ba392fd297c\") " pod="openstack/ceilometer-0" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.271063 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4d4f4fff-a6ec-4fef-82b4-6ba392fd297c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4d4f4fff-a6ec-4fef-82b4-6ba392fd297c\") " pod="openstack/ceilometer-0" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.271159 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjnp6\" (UniqueName: \"kubernetes.io/projected/4d4f4fff-a6ec-4fef-82b4-6ba392fd297c-kube-api-access-fjnp6\") pod \"ceilometer-0\" (UID: \"4d4f4fff-a6ec-4fef-82b4-6ba392fd297c\") " pod="openstack/ceilometer-0" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.271205 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d4f4fff-a6ec-4fef-82b4-6ba392fd297c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4d4f4fff-a6ec-4fef-82b4-6ba392fd297c\") " pod="openstack/ceilometer-0" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.271296 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mhwmh\" (UniqueName: \"kubernetes.io/projected/37b0a45e-9768-4b03-bf26-d35d3c98f58a-kube-api-access-mhwmh\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.271834 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4d4f4fff-a6ec-4fef-82b4-6ba392fd297c-log-httpd\") pod \"ceilometer-0\" (UID: \"4d4f4fff-a6ec-4fef-82b4-6ba392fd297c\") " pod="openstack/ceilometer-0" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.292538 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4d4f4fff-a6ec-4fef-82b4-6ba392fd297c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4d4f4fff-a6ec-4fef-82b4-6ba392fd297c\") " pod="openstack/ceilometer-0" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.297130 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-z4x9c"] Dec 06 03:58:39 crc kubenswrapper[4802]: E1206 03:58:39.297577 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37b0a45e-9768-4b03-bf26-d35d3c98f58a" containerName="init" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.297596 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="37b0a45e-9768-4b03-bf26-d35d3c98f58a" containerName="init" Dec 06 03:58:39 crc kubenswrapper[4802]: E1206 03:58:39.297625 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37b0a45e-9768-4b03-bf26-d35d3c98f58a" containerName="dnsmasq-dns" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.297631 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="37b0a45e-9768-4b03-bf26-d35d3c98f58a" containerName="dnsmasq-dns" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.304020 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="37b0a45e-9768-4b03-bf26-d35d3c98f58a" containerName="dnsmasq-dns" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.305029 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-z4x9c" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.306326 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d4f4fff-a6ec-4fef-82b4-6ba392fd297c-scripts\") pod \"ceilometer-0\" (UID: \"4d4f4fff-a6ec-4fef-82b4-6ba392fd297c\") " pod="openstack/ceilometer-0" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.306448 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d4f4fff-a6ec-4fef-82b4-6ba392fd297c-config-data\") pod \"ceilometer-0\" (UID: \"4d4f4fff-a6ec-4fef-82b4-6ba392fd297c\") " pod="openstack/ceilometer-0" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.307379 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-dxvz6" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.317424 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d4f4fff-a6ec-4fef-82b4-6ba392fd297c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4d4f4fff-a6ec-4fef-82b4-6ba392fd297c\") " pod="openstack/ceilometer-0" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.331310 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-z4x9c"] Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.384605 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjnp6\" (UniqueName: \"kubernetes.io/projected/4d4f4fff-a6ec-4fef-82b4-6ba392fd297c-kube-api-access-fjnp6\") pod \"ceilometer-0\" (UID: \"4d4f4fff-a6ec-4fef-82b4-6ba392fd297c\") " pod="openstack/ceilometer-0" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.431737 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37b0a45e-9768-4b03-bf26-d35d3c98f58a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "37b0a45e-9768-4b03-bf26-d35d3c98f58a" (UID: "37b0a45e-9768-4b03-bf26-d35d3c98f58a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.467195 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37b0a45e-9768-4b03-bf26-d35d3c98f58a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "37b0a45e-9768-4b03-bf26-d35d3c98f58a" (UID: "37b0a45e-9768-4b03-bf26-d35d3c98f58a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.470195 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37b0a45e-9768-4b03-bf26-d35d3c98f58a-config" (OuterVolumeSpecName: "config") pod "37b0a45e-9768-4b03-bf26-d35d3c98f58a" (UID: "37b0a45e-9768-4b03-bf26-d35d3c98f58a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.471217 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37b0a45e-9768-4b03-bf26-d35d3c98f58a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "37b0a45e-9768-4b03-bf26-d35d3c98f58a" (UID: "37b0a45e-9768-4b03-bf26-d35d3c98f58a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.475329 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f91ac3c0-8611-458e-bb4a-cfacb60d8b5b-config\") pod \"dnsmasq-dns-56df8fb6b7-z4x9c\" (UID: \"f91ac3c0-8611-458e-bb4a-cfacb60d8b5b\") " pod="openstack/dnsmasq-dns-56df8fb6b7-z4x9c" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.475390 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f91ac3c0-8611-458e-bb4a-cfacb60d8b5b-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-z4x9c\" (UID: \"f91ac3c0-8611-458e-bb4a-cfacb60d8b5b\") " pod="openstack/dnsmasq-dns-56df8fb6b7-z4x9c" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.475411 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f91ac3c0-8611-458e-bb4a-cfacb60d8b5b-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-z4x9c\" (UID: \"f91ac3c0-8611-458e-bb4a-cfacb60d8b5b\") " pod="openstack/dnsmasq-dns-56df8fb6b7-z4x9c" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.475432 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f91ac3c0-8611-458e-bb4a-cfacb60d8b5b-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-z4x9c\" (UID: \"f91ac3c0-8611-458e-bb4a-cfacb60d8b5b\") " pod="openstack/dnsmasq-dns-56df8fb6b7-z4x9c" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.475451 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f91ac3c0-8611-458e-bb4a-cfacb60d8b5b-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-z4x9c\" (UID: \"f91ac3c0-8611-458e-bb4a-cfacb60d8b5b\") " pod="openstack/dnsmasq-dns-56df8fb6b7-z4x9c" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.475530 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hsfz5\" (UniqueName: \"kubernetes.io/projected/f91ac3c0-8611-458e-bb4a-cfacb60d8b5b-kube-api-access-hsfz5\") pod \"dnsmasq-dns-56df8fb6b7-z4x9c\" (UID: \"f91ac3c0-8611-458e-bb4a-cfacb60d8b5b\") " pod="openstack/dnsmasq-dns-56df8fb6b7-z4x9c" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.475645 4802 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/37b0a45e-9768-4b03-bf26-d35d3c98f58a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.475658 4802 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37b0a45e-9768-4b03-bf26-d35d3c98f58a-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.475696 4802 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/37b0a45e-9768-4b03-bf26-d35d3c98f58a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.475707 4802 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/37b0a45e-9768-4b03-bf26-d35d3c98f58a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.489725 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37b0a45e-9768-4b03-bf26-d35d3c98f58a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "37b0a45e-9768-4b03-bf26-d35d3c98f58a" (UID: "37b0a45e-9768-4b03-bf26-d35d3c98f58a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:58:39 crc kubenswrapper[4802]: W1206 03:58:39.516920 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb24d756b_e69d_42c8_8f6c_0ea2a3d57d00.slice/crio-bb2592161c9b1b7b1be4815a96ef98534ba2a1228433ba4eeaeb74cab39d3fba WatchSource:0}: Error finding container bb2592161c9b1b7b1be4815a96ef98534ba2a1228433ba4eeaeb74cab39d3fba: Status 404 returned error can't find the container with id bb2592161c9b1b7b1be4815a96ef98534ba2a1228433ba4eeaeb74cab39d3fba Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.582586 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hsfz5\" (UniqueName: \"kubernetes.io/projected/f91ac3c0-8611-458e-bb4a-cfacb60d8b5b-kube-api-access-hsfz5\") pod \"dnsmasq-dns-56df8fb6b7-z4x9c\" (UID: \"f91ac3c0-8611-458e-bb4a-cfacb60d8b5b\") " pod="openstack/dnsmasq-dns-56df8fb6b7-z4x9c" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.582775 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f91ac3c0-8611-458e-bb4a-cfacb60d8b5b-config\") pod \"dnsmasq-dns-56df8fb6b7-z4x9c\" (UID: \"f91ac3c0-8611-458e-bb4a-cfacb60d8b5b\") " pod="openstack/dnsmasq-dns-56df8fb6b7-z4x9c" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.582818 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f91ac3c0-8611-458e-bb4a-cfacb60d8b5b-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-z4x9c\" (UID: \"f91ac3c0-8611-458e-bb4a-cfacb60d8b5b\") " pod="openstack/dnsmasq-dns-56df8fb6b7-z4x9c" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.582839 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f91ac3c0-8611-458e-bb4a-cfacb60d8b5b-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-z4x9c\" (UID: \"f91ac3c0-8611-458e-bb4a-cfacb60d8b5b\") " pod="openstack/dnsmasq-dns-56df8fb6b7-z4x9c" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.582869 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f91ac3c0-8611-458e-bb4a-cfacb60d8b5b-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-z4x9c\" (UID: \"f91ac3c0-8611-458e-bb4a-cfacb60d8b5b\") " pod="openstack/dnsmasq-dns-56df8fb6b7-z4x9c" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.582887 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f91ac3c0-8611-458e-bb4a-cfacb60d8b5b-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-z4x9c\" (UID: \"f91ac3c0-8611-458e-bb4a-cfacb60d8b5b\") " pod="openstack/dnsmasq-dns-56df8fb6b7-z4x9c" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.582935 4802 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/37b0a45e-9768-4b03-bf26-d35d3c98f58a-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.583780 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f91ac3c0-8611-458e-bb4a-cfacb60d8b5b-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-z4x9c\" (UID: \"f91ac3c0-8611-458e-bb4a-cfacb60d8b5b\") " pod="openstack/dnsmasq-dns-56df8fb6b7-z4x9c" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.583974 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f91ac3c0-8611-458e-bb4a-cfacb60d8b5b-config\") pod \"dnsmasq-dns-56df8fb6b7-z4x9c\" (UID: \"f91ac3c0-8611-458e-bb4a-cfacb60d8b5b\") " pod="openstack/dnsmasq-dns-56df8fb6b7-z4x9c" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.584715 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f91ac3c0-8611-458e-bb4a-cfacb60d8b5b-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-z4x9c\" (UID: \"f91ac3c0-8611-458e-bb4a-cfacb60d8b5b\") " pod="openstack/dnsmasq-dns-56df8fb6b7-z4x9c" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.585554 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f91ac3c0-8611-458e-bb4a-cfacb60d8b5b-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-z4x9c\" (UID: \"f91ac3c0-8611-458e-bb4a-cfacb60d8b5b\") " pod="openstack/dnsmasq-dns-56df8fb6b7-z4x9c" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.586546 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f91ac3c0-8611-458e-bb4a-cfacb60d8b5b-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-z4x9c\" (UID: \"f91ac3c0-8611-458e-bb4a-cfacb60d8b5b\") " pod="openstack/dnsmasq-dns-56df8fb6b7-z4x9c" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.591333 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-ddx8q"] Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.600793 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hsfz5\" (UniqueName: \"kubernetes.io/projected/f91ac3c0-8611-458e-bb4a-cfacb60d8b5b-kube-api-access-hsfz5\") pod \"dnsmasq-dns-56df8fb6b7-z4x9c\" (UID: \"f91ac3c0-8611-458e-bb4a-cfacb60d8b5b\") " pod="openstack/dnsmasq-dns-56df8fb6b7-z4x9c" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.606709 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-z4x9c" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.648948 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.714861 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-2hzlt"] Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.763144 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b868669f-shq86"] Dec 06 03:58:39 crc kubenswrapper[4802]: W1206 03:58:39.769571 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod32ee40e7_4fe2_42c3_b4c9_0dc894eb17e4.slice/crio-1e29f83ab6924ad19beac3e29ec455149ed2b785c2300d934b4405453660d316 WatchSource:0}: Error finding container 1e29f83ab6924ad19beac3e29ec455149ed2b785c2300d934b4405453660d316: Status 404 returned error can't find the container with id 1e29f83ab6924ad19beac3e29ec455149ed2b785c2300d934b4405453660d316 Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.961205 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-ddx8q" event={"ID":"b24d756b-e69d-42c8-8f6c-0ea2a3d57d00","Type":"ContainerStarted","Data":"bb2592161c9b1b7b1be4815a96ef98534ba2a1228433ba4eeaeb74cab39d3fba"} Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.962208 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b868669f-shq86" event={"ID":"32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4","Type":"ContainerStarted","Data":"1e29f83ab6924ad19beac3e29ec455149ed2b785c2300d934b4405453660d316"} Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.964061 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-2hzlt" event={"ID":"2a639667-be04-429f-a2ce-b20a011425f8","Type":"ContainerStarted","Data":"3173bdb2f47b8aaeecd93e04af3df770499eef5d3268913395d23e8481c67cc6"} Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.965738 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-w49ls" event={"ID":"37b0a45e-9768-4b03-bf26-d35d3c98f58a","Type":"ContainerDied","Data":"64f380ed649f9607bbd62872585717105bf77daeca9a78e75fcd52c063e8050e"} Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.965823 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-w49ls" Dec 06 03:58:39 crc kubenswrapper[4802]: I1206 03:58:39.965892 4802 scope.go:117] "RemoveContainer" containerID="99b3ae2fcbdcf8dd15fdf93bc19b0329b9e53253fa4267acb98239a199207902" Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.052559 4802 scope.go:117] "RemoveContainer" containerID="c8ea67068f3ec73d843aaf2410f2c467b06159bad710370a55376c11e644a279" Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.055340 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-w49ls"] Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.094928 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-w49ls"] Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.116140 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-4gbch"] Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.125495 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-qs8hv"] Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.158222 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.161013 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.165363 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-n458c" Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.165638 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.166043 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.193784 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.213024 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-5hj5p"] Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.225181 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-qtg4k"] Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.236982 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-dxvz6"] Dec 06 03:58:40 crc kubenswrapper[4802]: W1206 03:58:40.263867 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4abd9ed3_07e6_46a6_8539_85fc5c624145.slice/crio-91add2d6376f69256916a934641c5d1403c2313455ee03e98ad61e269d1bff21 WatchSource:0}: Error finding container 91add2d6376f69256916a934641c5d1403c2313455ee03e98ad61e269d1bff21: Status 404 returned error can't find the container with id 91add2d6376f69256916a934641c5d1403c2313455ee03e98ad61e269d1bff21 Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.327739 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cb49a4b-5eee-498e-a5ba-584e6ee6d79a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3cb49a4b-5eee-498e-a5ba-584e6ee6d79a\") " pod="openstack/glance-default-external-api-0" Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.327807 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4snvb\" (UniqueName: \"kubernetes.io/projected/3cb49a4b-5eee-498e-a5ba-584e6ee6d79a-kube-api-access-4snvb\") pod \"glance-default-external-api-0\" (UID: \"3cb49a4b-5eee-498e-a5ba-584e6ee6d79a\") " pod="openstack/glance-default-external-api-0" Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.327842 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3cb49a4b-5eee-498e-a5ba-584e6ee6d79a-scripts\") pod \"glance-default-external-api-0\" (UID: \"3cb49a4b-5eee-498e-a5ba-584e6ee6d79a\") " pod="openstack/glance-default-external-api-0" Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.327876 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3cb49a4b-5eee-498e-a5ba-584e6ee6d79a-logs\") pod \"glance-default-external-api-0\" (UID: \"3cb49a4b-5eee-498e-a5ba-584e6ee6d79a\") " pod="openstack/glance-default-external-api-0" Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.327917 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3cb49a4b-5eee-498e-a5ba-584e6ee6d79a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3cb49a4b-5eee-498e-a5ba-584e6ee6d79a\") " pod="openstack/glance-default-external-api-0" Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.327935 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"3cb49a4b-5eee-498e-a5ba-584e6ee6d79a\") " pod="openstack/glance-default-external-api-0" Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.327993 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cb49a4b-5eee-498e-a5ba-584e6ee6d79a-config-data\") pod \"glance-default-external-api-0\" (UID: \"3cb49a4b-5eee-498e-a5ba-584e6ee6d79a\") " pod="openstack/glance-default-external-api-0" Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.348739 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-z4x9c"] Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.393940 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 03:58:40 crc kubenswrapper[4802]: E1206 03:58:40.400219 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle config-data glance httpd-run kube-api-access-4snvb logs scripts], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/glance-default-external-api-0" podUID="3cb49a4b-5eee-498e-a5ba-584e6ee6d79a" Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.429651 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cb49a4b-5eee-498e-a5ba-584e6ee6d79a-config-data\") pod \"glance-default-external-api-0\" (UID: \"3cb49a4b-5eee-498e-a5ba-584e6ee6d79a\") " pod="openstack/glance-default-external-api-0" Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.429695 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cb49a4b-5eee-498e-a5ba-584e6ee6d79a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3cb49a4b-5eee-498e-a5ba-584e6ee6d79a\") " pod="openstack/glance-default-external-api-0" Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.429731 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4snvb\" (UniqueName: \"kubernetes.io/projected/3cb49a4b-5eee-498e-a5ba-584e6ee6d79a-kube-api-access-4snvb\") pod \"glance-default-external-api-0\" (UID: \"3cb49a4b-5eee-498e-a5ba-584e6ee6d79a\") " pod="openstack/glance-default-external-api-0" Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.429775 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3cb49a4b-5eee-498e-a5ba-584e6ee6d79a-scripts\") pod \"glance-default-external-api-0\" (UID: \"3cb49a4b-5eee-498e-a5ba-584e6ee6d79a\") " pod="openstack/glance-default-external-api-0" Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.429807 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3cb49a4b-5eee-498e-a5ba-584e6ee6d79a-logs\") pod \"glance-default-external-api-0\" (UID: \"3cb49a4b-5eee-498e-a5ba-584e6ee6d79a\") " pod="openstack/glance-default-external-api-0" Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.429847 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3cb49a4b-5eee-498e-a5ba-584e6ee6d79a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3cb49a4b-5eee-498e-a5ba-584e6ee6d79a\") " pod="openstack/glance-default-external-api-0" Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.429864 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"3cb49a4b-5eee-498e-a5ba-584e6ee6d79a\") " pod="openstack/glance-default-external-api-0" Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.430189 4802 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"3cb49a4b-5eee-498e-a5ba-584e6ee6d79a\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-external-api-0" Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.430538 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.432248 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.457163 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.457740 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3cb49a4b-5eee-498e-a5ba-584e6ee6d79a-logs\") pod \"glance-default-external-api-0\" (UID: \"3cb49a4b-5eee-498e-a5ba-584e6ee6d79a\") " pod="openstack/glance-default-external-api-0" Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.458266 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3cb49a4b-5eee-498e-a5ba-584e6ee6d79a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3cb49a4b-5eee-498e-a5ba-584e6ee6d79a\") " pod="openstack/glance-default-external-api-0" Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.463064 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cb49a4b-5eee-498e-a5ba-584e6ee6d79a-config-data\") pod \"glance-default-external-api-0\" (UID: \"3cb49a4b-5eee-498e-a5ba-584e6ee6d79a\") " pod="openstack/glance-default-external-api-0" Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.471047 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.495190 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3cb49a4b-5eee-498e-a5ba-584e6ee6d79a-scripts\") pod \"glance-default-external-api-0\" (UID: \"3cb49a4b-5eee-498e-a5ba-584e6ee6d79a\") " pod="openstack/glance-default-external-api-0" Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.502058 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cb49a4b-5eee-498e-a5ba-584e6ee6d79a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3cb49a4b-5eee-498e-a5ba-584e6ee6d79a\") " pod="openstack/glance-default-external-api-0" Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.517943 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4snvb\" (UniqueName: \"kubernetes.io/projected/3cb49a4b-5eee-498e-a5ba-584e6ee6d79a-kube-api-access-4snvb\") pod \"glance-default-external-api-0\" (UID: \"3cb49a4b-5eee-498e-a5ba-584e6ee6d79a\") " pod="openstack/glance-default-external-api-0" Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.524495 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.532709 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ec8cedbe-d81f-470c-a964-83e5a47a9db2-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ec8cedbe-d81f-470c-a964-83e5a47a9db2\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.532787 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"ec8cedbe-d81f-470c-a964-83e5a47a9db2\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.532806 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec8cedbe-d81f-470c-a964-83e5a47a9db2-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ec8cedbe-d81f-470c-a964-83e5a47a9db2\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.532862 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec8cedbe-d81f-470c-a964-83e5a47a9db2-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ec8cedbe-d81f-470c-a964-83e5a47a9db2\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.532924 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec8cedbe-d81f-470c-a964-83e5a47a9db2-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ec8cedbe-d81f-470c-a964-83e5a47a9db2\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.532943 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8b4t\" (UniqueName: \"kubernetes.io/projected/ec8cedbe-d81f-470c-a964-83e5a47a9db2-kube-api-access-t8b4t\") pod \"glance-default-internal-api-0\" (UID: \"ec8cedbe-d81f-470c-a964-83e5a47a9db2\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.532976 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec8cedbe-d81f-470c-a964-83e5a47a9db2-logs\") pod \"glance-default-internal-api-0\" (UID: \"ec8cedbe-d81f-470c-a964-83e5a47a9db2\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.562567 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"3cb49a4b-5eee-498e-a5ba-584e6ee6d79a\") " pod="openstack/glance-default-external-api-0" Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.601774 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.636057 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.637503 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"ec8cedbe-d81f-470c-a964-83e5a47a9db2\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.637666 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec8cedbe-d81f-470c-a964-83e5a47a9db2-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ec8cedbe-d81f-470c-a964-83e5a47a9db2\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.637888 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec8cedbe-d81f-470c-a964-83e5a47a9db2-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ec8cedbe-d81f-470c-a964-83e5a47a9db2\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.638072 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec8cedbe-d81f-470c-a964-83e5a47a9db2-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ec8cedbe-d81f-470c-a964-83e5a47a9db2\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.638212 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8b4t\" (UniqueName: \"kubernetes.io/projected/ec8cedbe-d81f-470c-a964-83e5a47a9db2-kube-api-access-t8b4t\") pod \"glance-default-internal-api-0\" (UID: \"ec8cedbe-d81f-470c-a964-83e5a47a9db2\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.638383 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec8cedbe-d81f-470c-a964-83e5a47a9db2-logs\") pod \"glance-default-internal-api-0\" (UID: \"ec8cedbe-d81f-470c-a964-83e5a47a9db2\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.638511 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ec8cedbe-d81f-470c-a964-83e5a47a9db2-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ec8cedbe-d81f-470c-a964-83e5a47a9db2\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.639175 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ec8cedbe-d81f-470c-a964-83e5a47a9db2-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ec8cedbe-d81f-470c-a964-83e5a47a9db2\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.642342 4802 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"ec8cedbe-d81f-470c-a964-83e5a47a9db2\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-internal-api-0" Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.660577 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec8cedbe-d81f-470c-a964-83e5a47a9db2-logs\") pod \"glance-default-internal-api-0\" (UID: \"ec8cedbe-d81f-470c-a964-83e5a47a9db2\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.698663 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8b4t\" (UniqueName: \"kubernetes.io/projected/ec8cedbe-d81f-470c-a964-83e5a47a9db2-kube-api-access-t8b4t\") pod \"glance-default-internal-api-0\" (UID: \"ec8cedbe-d81f-470c-a964-83e5a47a9db2\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.710246 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec8cedbe-d81f-470c-a964-83e5a47a9db2-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ec8cedbe-d81f-470c-a964-83e5a47a9db2\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.715821 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec8cedbe-d81f-470c-a964-83e5a47a9db2-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ec8cedbe-d81f-470c-a964-83e5a47a9db2\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.716219 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec8cedbe-d81f-470c-a964-83e5a47a9db2-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ec8cedbe-d81f-470c-a964-83e5a47a9db2\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.717631 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"ec8cedbe-d81f-470c-a964-83e5a47a9db2\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.767119 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.984143 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-qs8hv" event={"ID":"8c25c327-e3cd-4f28-855a-249e4bb0c5f6","Type":"ContainerStarted","Data":"22256b01fbd251cd8845a65d030b3e708b70e6bb74755d572798823340fae615"} Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.984402 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-qs8hv" event={"ID":"8c25c327-e3cd-4f28-855a-249e4bb0c5f6","Type":"ContainerStarted","Data":"acbf648bc77efc79c07ddf28a963e7543f0acb21850d33865a63152a8d883989"} Dec 06 03:58:40 crc kubenswrapper[4802]: I1206 03:58:40.990520 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-qtg4k" event={"ID":"5c636931-d435-47ec-b770-80dc1ee60756","Type":"ContainerStarted","Data":"8c51e9bbf9a382e680ba73c38c4086dfd7247c8036b228086da905f8a953b122"} Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.000940 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-dxvz6" event={"ID":"4abd9ed3-07e6-46a6-8539-85fc5c624145","Type":"ContainerStarted","Data":"91add2d6376f69256916a934641c5d1403c2313455ee03e98ad61e269d1bff21"} Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.010198 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-qs8hv" podStartSLOduration=3.010181762 podStartE2EDuration="3.010181762s" podCreationTimestamp="2025-12-06 03:58:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:58:41.009173254 +0000 UTC m=+1113.881082406" watchObservedRunningTime="2025-12-06 03:58:41.010181762 +0000 UTC m=+1113.882090914" Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.011196 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-ddx8q" event={"ID":"b24d756b-e69d-42c8-8f6c-0ea2a3d57d00","Type":"ContainerStarted","Data":"0f19de9d3ffb77620b7ecce801747092e84300ff9f114eff03946ce3f3ce3405"} Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.015225 4802 generic.go:334] "Generic (PLEG): container finished" podID="32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4" containerID="c2ff792b218bcc8de5c4b7ab67c39dbb1764fe613342bd19c0cb2dace45e3ae2" exitCode=0 Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.015286 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b868669f-shq86" event={"ID":"32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4","Type":"ContainerDied","Data":"c2ff792b218bcc8de5c4b7ab67c39dbb1764fe613342bd19c0cb2dace45e3ae2"} Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.023448 4802 generic.go:334] "Generic (PLEG): container finished" podID="f91ac3c0-8611-458e-bb4a-cfacb60d8b5b" containerID="0167fa890261a5700e8a3554578562ee7c6e7304e40de2db2330b755086ae8e9" exitCode=0 Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.023505 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-z4x9c" event={"ID":"f91ac3c0-8611-458e-bb4a-cfacb60d8b5b","Type":"ContainerDied","Data":"0167fa890261a5700e8a3554578562ee7c6e7304e40de2db2330b755086ae8e9"} Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.023526 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-z4x9c" event={"ID":"f91ac3c0-8611-458e-bb4a-cfacb60d8b5b","Type":"ContainerStarted","Data":"1ff3e5fe6bf0f9d9b42183731e7c24bb6aaa5a6c9bec17525aef35cd8c01b62d"} Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.024808 4802 generic.go:334] "Generic (PLEG): container finished" podID="21cae3ac-9ae8-4c31-972d-7a3c25e8bd95" containerID="689c5edb7366110bb9cfec8e85ab178dae1afcc1a4e826332201b599523b2f0e" exitCode=0 Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.024842 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf78879c9-4gbch" event={"ID":"21cae3ac-9ae8-4c31-972d-7a3c25e8bd95","Type":"ContainerDied","Data":"689c5edb7366110bb9cfec8e85ab178dae1afcc1a4e826332201b599523b2f0e"} Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.024857 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf78879c9-4gbch" event={"ID":"21cae3ac-9ae8-4c31-972d-7a3c25e8bd95","Type":"ContainerStarted","Data":"ea989ab32e47bbd10957b293ee5ced5df3f979526e98923a1e7bfd65a13f36f6"} Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.048901 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-ddx8q" podStartSLOduration=3.048884625 podStartE2EDuration="3.048884625s" podCreationTimestamp="2025-12-06 03:58:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:58:41.043257942 +0000 UTC m=+1113.915167094" watchObservedRunningTime="2025-12-06 03:58:41.048884625 +0000 UTC m=+1113.920793777" Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.050922 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-5hj5p" event={"ID":"f1644907-b4b1-4e23-a273-19d535bd7f88","Type":"ContainerStarted","Data":"e0760f7b4403432817d1e174766269537a2b9a6108c03ff28e47ddde093058e0"} Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.059987 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.059986 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4d4f4fff-a6ec-4fef-82b4-6ba392fd297c","Type":"ContainerStarted","Data":"5ebc226f052371f321b24045dae747ef258f0a92b1f6b491170158e35a15d1ca"} Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.109224 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.153136 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4snvb\" (UniqueName: \"kubernetes.io/projected/3cb49a4b-5eee-498e-a5ba-584e6ee6d79a-kube-api-access-4snvb\") pod \"3cb49a4b-5eee-498e-a5ba-584e6ee6d79a\" (UID: \"3cb49a4b-5eee-498e-a5ba-584e6ee6d79a\") " Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.153186 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cb49a4b-5eee-498e-a5ba-584e6ee6d79a-config-data\") pod \"3cb49a4b-5eee-498e-a5ba-584e6ee6d79a\" (UID: \"3cb49a4b-5eee-498e-a5ba-584e6ee6d79a\") " Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.153234 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3cb49a4b-5eee-498e-a5ba-584e6ee6d79a-logs\") pod \"3cb49a4b-5eee-498e-a5ba-584e6ee6d79a\" (UID: \"3cb49a4b-5eee-498e-a5ba-584e6ee6d79a\") " Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.153264 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3cb49a4b-5eee-498e-a5ba-584e6ee6d79a-httpd-run\") pod \"3cb49a4b-5eee-498e-a5ba-584e6ee6d79a\" (UID: \"3cb49a4b-5eee-498e-a5ba-584e6ee6d79a\") " Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.153339 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cb49a4b-5eee-498e-a5ba-584e6ee6d79a-combined-ca-bundle\") pod \"3cb49a4b-5eee-498e-a5ba-584e6ee6d79a\" (UID: \"3cb49a4b-5eee-498e-a5ba-584e6ee6d79a\") " Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.153376 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3cb49a4b-5eee-498e-a5ba-584e6ee6d79a-scripts\") pod \"3cb49a4b-5eee-498e-a5ba-584e6ee6d79a\" (UID: \"3cb49a4b-5eee-498e-a5ba-584e6ee6d79a\") " Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.153409 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"3cb49a4b-5eee-498e-a5ba-584e6ee6d79a\" (UID: \"3cb49a4b-5eee-498e-a5ba-584e6ee6d79a\") " Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.154786 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3cb49a4b-5eee-498e-a5ba-584e6ee6d79a-logs" (OuterVolumeSpecName: "logs") pod "3cb49a4b-5eee-498e-a5ba-584e6ee6d79a" (UID: "3cb49a4b-5eee-498e-a5ba-584e6ee6d79a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.158168 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3cb49a4b-5eee-498e-a5ba-584e6ee6d79a-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "3cb49a4b-5eee-498e-a5ba-584e6ee6d79a" (UID: "3cb49a4b-5eee-498e-a5ba-584e6ee6d79a"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.160429 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "3cb49a4b-5eee-498e-a5ba-584e6ee6d79a" (UID: "3cb49a4b-5eee-498e-a5ba-584e6ee6d79a"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.161159 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb49a4b-5eee-498e-a5ba-584e6ee6d79a-kube-api-access-4snvb" (OuterVolumeSpecName: "kube-api-access-4snvb") pod "3cb49a4b-5eee-498e-a5ba-584e6ee6d79a" (UID: "3cb49a4b-5eee-498e-a5ba-584e6ee6d79a"). InnerVolumeSpecName "kube-api-access-4snvb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.163012 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cb49a4b-5eee-498e-a5ba-584e6ee6d79a-config-data" (OuterVolumeSpecName: "config-data") pod "3cb49a4b-5eee-498e-a5ba-584e6ee6d79a" (UID: "3cb49a4b-5eee-498e-a5ba-584e6ee6d79a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.166589 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cb49a4b-5eee-498e-a5ba-584e6ee6d79a-scripts" (OuterVolumeSpecName: "scripts") pod "3cb49a4b-5eee-498e-a5ba-584e6ee6d79a" (UID: "3cb49a4b-5eee-498e-a5ba-584e6ee6d79a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.168028 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cb49a4b-5eee-498e-a5ba-584e6ee6d79a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3cb49a4b-5eee-498e-a5ba-584e6ee6d79a" (UID: "3cb49a4b-5eee-498e-a5ba-584e6ee6d79a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.255942 4802 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cb49a4b-5eee-498e-a5ba-584e6ee6d79a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.255980 4802 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3cb49a4b-5eee-498e-a5ba-584e6ee6d79a-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.256033 4802 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.256048 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4snvb\" (UniqueName: \"kubernetes.io/projected/3cb49a4b-5eee-498e-a5ba-584e6ee6d79a-kube-api-access-4snvb\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.256061 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cb49a4b-5eee-498e-a5ba-584e6ee6d79a-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.256069 4802 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3cb49a4b-5eee-498e-a5ba-584e6ee6d79a-logs\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.256103 4802 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3cb49a4b-5eee-498e-a5ba-584e6ee6d79a-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.282731 4802 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.312244 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.358398 4802 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.486925 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37b0a45e-9768-4b03-bf26-d35d3c98f58a" path="/var/lib/kubelet/pods/37b0a45e-9768-4b03-bf26-d35d3c98f58a/volumes" Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.537195 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b868669f-shq86" Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.562452 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4-ovsdbserver-nb\") pod \"32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4\" (UID: \"32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4\") " Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.562525 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4-dns-swift-storage-0\") pod \"32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4\" (UID: \"32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4\") " Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.562564 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-czslj\" (UniqueName: \"kubernetes.io/projected/32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4-kube-api-access-czslj\") pod \"32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4\" (UID: \"32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4\") " Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.562607 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4-ovsdbserver-sb\") pod \"32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4\" (UID: \"32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4\") " Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.562715 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4-dns-svc\") pod \"32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4\" (UID: \"32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4\") " Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.562856 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4-config\") pod \"32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4\" (UID: \"32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4\") " Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.576033 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4-kube-api-access-czslj" (OuterVolumeSpecName: "kube-api-access-czslj") pod "32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4" (UID: "32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4"). InnerVolumeSpecName "kube-api-access-czslj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.604649 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4-config" (OuterVolumeSpecName: "config") pod "32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4" (UID: "32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.663126 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4" (UID: "32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.663629 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4" (UID: "32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.665731 4802 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.665913 4802 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.666065 4802 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.666204 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-czslj\" (UniqueName: \"kubernetes.io/projected/32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4-kube-api-access-czslj\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.671118 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4" (UID: "32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.672648 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4" (UID: "32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.677807 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf78879c9-4gbch" Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.767141 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/21cae3ac-9ae8-4c31-972d-7a3c25e8bd95-ovsdbserver-nb\") pod \"21cae3ac-9ae8-4c31-972d-7a3c25e8bd95\" (UID: \"21cae3ac-9ae8-4c31-972d-7a3c25e8bd95\") " Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.767204 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/21cae3ac-9ae8-4c31-972d-7a3c25e8bd95-dns-swift-storage-0\") pod \"21cae3ac-9ae8-4c31-972d-7a3c25e8bd95\" (UID: \"21cae3ac-9ae8-4c31-972d-7a3c25e8bd95\") " Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.767225 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4c4tx\" (UniqueName: \"kubernetes.io/projected/21cae3ac-9ae8-4c31-972d-7a3c25e8bd95-kube-api-access-4c4tx\") pod \"21cae3ac-9ae8-4c31-972d-7a3c25e8bd95\" (UID: \"21cae3ac-9ae8-4c31-972d-7a3c25e8bd95\") " Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.767251 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21cae3ac-9ae8-4c31-972d-7a3c25e8bd95-config\") pod \"21cae3ac-9ae8-4c31-972d-7a3c25e8bd95\" (UID: \"21cae3ac-9ae8-4c31-972d-7a3c25e8bd95\") " Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.767308 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/21cae3ac-9ae8-4c31-972d-7a3c25e8bd95-dns-svc\") pod \"21cae3ac-9ae8-4c31-972d-7a3c25e8bd95\" (UID: \"21cae3ac-9ae8-4c31-972d-7a3c25e8bd95\") " Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.767329 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/21cae3ac-9ae8-4c31-972d-7a3c25e8bd95-ovsdbserver-sb\") pod \"21cae3ac-9ae8-4c31-972d-7a3c25e8bd95\" (UID: \"21cae3ac-9ae8-4c31-972d-7a3c25e8bd95\") " Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.768187 4802 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.768201 4802 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.778918 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21cae3ac-9ae8-4c31-972d-7a3c25e8bd95-kube-api-access-4c4tx" (OuterVolumeSpecName: "kube-api-access-4c4tx") pod "21cae3ac-9ae8-4c31-972d-7a3c25e8bd95" (UID: "21cae3ac-9ae8-4c31-972d-7a3c25e8bd95"). InnerVolumeSpecName "kube-api-access-4c4tx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.794333 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21cae3ac-9ae8-4c31-972d-7a3c25e8bd95-config" (OuterVolumeSpecName: "config") pod "21cae3ac-9ae8-4c31-972d-7a3c25e8bd95" (UID: "21cae3ac-9ae8-4c31-972d-7a3c25e8bd95"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.806218 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21cae3ac-9ae8-4c31-972d-7a3c25e8bd95-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "21cae3ac-9ae8-4c31-972d-7a3c25e8bd95" (UID: "21cae3ac-9ae8-4c31-972d-7a3c25e8bd95"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.810220 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21cae3ac-9ae8-4c31-972d-7a3c25e8bd95-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "21cae3ac-9ae8-4c31-972d-7a3c25e8bd95" (UID: "21cae3ac-9ae8-4c31-972d-7a3c25e8bd95"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.818010 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21cae3ac-9ae8-4c31-972d-7a3c25e8bd95-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "21cae3ac-9ae8-4c31-972d-7a3c25e8bd95" (UID: "21cae3ac-9ae8-4c31-972d-7a3c25e8bd95"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.820168 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21cae3ac-9ae8-4c31-972d-7a3c25e8bd95-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "21cae3ac-9ae8-4c31-972d-7a3c25e8bd95" (UID: "21cae3ac-9ae8-4c31-972d-7a3c25e8bd95"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.870084 4802 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/21cae3ac-9ae8-4c31-972d-7a3c25e8bd95-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.870115 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4c4tx\" (UniqueName: \"kubernetes.io/projected/21cae3ac-9ae8-4c31-972d-7a3c25e8bd95-kube-api-access-4c4tx\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.870126 4802 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21cae3ac-9ae8-4c31-972d-7a3c25e8bd95-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.870133 4802 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/21cae3ac-9ae8-4c31-972d-7a3c25e8bd95-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.870142 4802 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/21cae3ac-9ae8-4c31-972d-7a3c25e8bd95-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:41 crc kubenswrapper[4802]: I1206 03:58:41.870149 4802 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/21cae3ac-9ae8-4c31-972d-7a3c25e8bd95-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:42 crc kubenswrapper[4802]: I1206 03:58:42.087020 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b868669f-shq86" event={"ID":"32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4","Type":"ContainerDied","Data":"1e29f83ab6924ad19beac3e29ec455149ed2b785c2300d934b4405453660d316"} Dec 06 03:58:42 crc kubenswrapper[4802]: I1206 03:58:42.087275 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b868669f-shq86" Dec 06 03:58:42 crc kubenswrapper[4802]: I1206 03:58:42.087413 4802 scope.go:117] "RemoveContainer" containerID="c2ff792b218bcc8de5c4b7ab67c39dbb1764fe613342bd19c0cb2dace45e3ae2" Dec 06 03:58:42 crc kubenswrapper[4802]: I1206 03:58:42.095817 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-z4x9c" event={"ID":"f91ac3c0-8611-458e-bb4a-cfacb60d8b5b","Type":"ContainerStarted","Data":"dbc39ed94f5b7975108e7a698f80ef79b3a5290662fc25802033727d945074ef"} Dec 06 03:58:42 crc kubenswrapper[4802]: I1206 03:58:42.096979 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-56df8fb6b7-z4x9c" Dec 06 03:58:42 crc kubenswrapper[4802]: I1206 03:58:42.103722 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ec8cedbe-d81f-470c-a964-83e5a47a9db2","Type":"ContainerStarted","Data":"4f34db756bb1581336b9b9813aff7202e5d250ae710d7ad6961772015e92c7d2"} Dec 06 03:58:42 crc kubenswrapper[4802]: I1206 03:58:42.108839 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf78879c9-4gbch" event={"ID":"21cae3ac-9ae8-4c31-972d-7a3c25e8bd95","Type":"ContainerDied","Data":"ea989ab32e47bbd10957b293ee5ced5df3f979526e98923a1e7bfd65a13f36f6"} Dec 06 03:58:42 crc kubenswrapper[4802]: I1206 03:58:42.109253 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 03:58:42 crc kubenswrapper[4802]: I1206 03:58:42.111101 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf78879c9-4gbch" Dec 06 03:58:42 crc kubenswrapper[4802]: I1206 03:58:42.118465 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-56df8fb6b7-z4x9c" podStartSLOduration=3.11844918 podStartE2EDuration="3.11844918s" podCreationTimestamp="2025-12-06 03:58:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:58:42.116102446 +0000 UTC m=+1114.988011598" watchObservedRunningTime="2025-12-06 03:58:42.11844918 +0000 UTC m=+1114.990358332" Dec 06 03:58:42 crc kubenswrapper[4802]: I1206 03:58:42.191206 4802 scope.go:117] "RemoveContainer" containerID="689c5edb7366110bb9cfec8e85ab178dae1afcc1a4e826332201b599523b2f0e" Dec 06 03:58:42 crc kubenswrapper[4802]: I1206 03:58:42.193429 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 03:58:42 crc kubenswrapper[4802]: I1206 03:58:42.215889 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 03:58:42 crc kubenswrapper[4802]: I1206 03:58:42.243707 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 03:58:42 crc kubenswrapper[4802]: E1206 03:58:42.244156 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4" containerName="init" Dec 06 03:58:42 crc kubenswrapper[4802]: I1206 03:58:42.244171 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4" containerName="init" Dec 06 03:58:42 crc kubenswrapper[4802]: E1206 03:58:42.244188 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21cae3ac-9ae8-4c31-972d-7a3c25e8bd95" containerName="init" Dec 06 03:58:42 crc kubenswrapper[4802]: I1206 03:58:42.244195 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="21cae3ac-9ae8-4c31-972d-7a3c25e8bd95" containerName="init" Dec 06 03:58:42 crc kubenswrapper[4802]: I1206 03:58:42.244786 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="21cae3ac-9ae8-4c31-972d-7a3c25e8bd95" containerName="init" Dec 06 03:58:42 crc kubenswrapper[4802]: I1206 03:58:42.244804 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4" containerName="init" Dec 06 03:58:42 crc kubenswrapper[4802]: I1206 03:58:42.246608 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 03:58:42 crc kubenswrapper[4802]: I1206 03:58:42.262838 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 06 03:58:42 crc kubenswrapper[4802]: I1206 03:58:42.264310 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b868669f-shq86"] Dec 06 03:58:42 crc kubenswrapper[4802]: I1206 03:58:42.278970 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b868669f-shq86"] Dec 06 03:58:42 crc kubenswrapper[4802]: I1206 03:58:42.291816 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 03:58:42 crc kubenswrapper[4802]: I1206 03:58:42.309215 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-4gbch"] Dec 06 03:58:42 crc kubenswrapper[4802]: I1206 03:58:42.319353 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-4gbch"] Dec 06 03:58:42 crc kubenswrapper[4802]: I1206 03:58:42.388211 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqt6l\" (UniqueName: \"kubernetes.io/projected/63be777b-838a-499d-a04c-c522ae8fcb2f-kube-api-access-lqt6l\") pod \"glance-default-external-api-0\" (UID: \"63be777b-838a-499d-a04c-c522ae8fcb2f\") " pod="openstack/glance-default-external-api-0" Dec 06 03:58:42 crc kubenswrapper[4802]: I1206 03:58:42.388475 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/63be777b-838a-499d-a04c-c522ae8fcb2f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"63be777b-838a-499d-a04c-c522ae8fcb2f\") " pod="openstack/glance-default-external-api-0" Dec 06 03:58:42 crc kubenswrapper[4802]: I1206 03:58:42.388852 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63be777b-838a-499d-a04c-c522ae8fcb2f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"63be777b-838a-499d-a04c-c522ae8fcb2f\") " pod="openstack/glance-default-external-api-0" Dec 06 03:58:42 crc kubenswrapper[4802]: I1206 03:58:42.389050 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/63be777b-838a-499d-a04c-c522ae8fcb2f-logs\") pod \"glance-default-external-api-0\" (UID: \"63be777b-838a-499d-a04c-c522ae8fcb2f\") " pod="openstack/glance-default-external-api-0" Dec 06 03:58:42 crc kubenswrapper[4802]: I1206 03:58:42.389978 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"63be777b-838a-499d-a04c-c522ae8fcb2f\") " pod="openstack/glance-default-external-api-0" Dec 06 03:58:42 crc kubenswrapper[4802]: I1206 03:58:42.390129 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63be777b-838a-499d-a04c-c522ae8fcb2f-config-data\") pod \"glance-default-external-api-0\" (UID: \"63be777b-838a-499d-a04c-c522ae8fcb2f\") " pod="openstack/glance-default-external-api-0" Dec 06 03:58:42 crc kubenswrapper[4802]: I1206 03:58:42.390318 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63be777b-838a-499d-a04c-c522ae8fcb2f-scripts\") pod \"glance-default-external-api-0\" (UID: \"63be777b-838a-499d-a04c-c522ae8fcb2f\") " pod="openstack/glance-default-external-api-0" Dec 06 03:58:42 crc kubenswrapper[4802]: I1206 03:58:42.506222 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63be777b-838a-499d-a04c-c522ae8fcb2f-config-data\") pod \"glance-default-external-api-0\" (UID: \"63be777b-838a-499d-a04c-c522ae8fcb2f\") " pod="openstack/glance-default-external-api-0" Dec 06 03:58:42 crc kubenswrapper[4802]: I1206 03:58:42.506535 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63be777b-838a-499d-a04c-c522ae8fcb2f-scripts\") pod \"glance-default-external-api-0\" (UID: \"63be777b-838a-499d-a04c-c522ae8fcb2f\") " pod="openstack/glance-default-external-api-0" Dec 06 03:58:42 crc kubenswrapper[4802]: I1206 03:58:42.506568 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqt6l\" (UniqueName: \"kubernetes.io/projected/63be777b-838a-499d-a04c-c522ae8fcb2f-kube-api-access-lqt6l\") pod \"glance-default-external-api-0\" (UID: \"63be777b-838a-499d-a04c-c522ae8fcb2f\") " pod="openstack/glance-default-external-api-0" Dec 06 03:58:42 crc kubenswrapper[4802]: I1206 03:58:42.506598 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/63be777b-838a-499d-a04c-c522ae8fcb2f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"63be777b-838a-499d-a04c-c522ae8fcb2f\") " pod="openstack/glance-default-external-api-0" Dec 06 03:58:42 crc kubenswrapper[4802]: I1206 03:58:42.506645 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63be777b-838a-499d-a04c-c522ae8fcb2f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"63be777b-838a-499d-a04c-c522ae8fcb2f\") " pod="openstack/glance-default-external-api-0" Dec 06 03:58:42 crc kubenswrapper[4802]: I1206 03:58:42.506667 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/63be777b-838a-499d-a04c-c522ae8fcb2f-logs\") pod \"glance-default-external-api-0\" (UID: \"63be777b-838a-499d-a04c-c522ae8fcb2f\") " pod="openstack/glance-default-external-api-0" Dec 06 03:58:42 crc kubenswrapper[4802]: I1206 03:58:42.506729 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"63be777b-838a-499d-a04c-c522ae8fcb2f\") " pod="openstack/glance-default-external-api-0" Dec 06 03:58:42 crc kubenswrapper[4802]: I1206 03:58:42.506979 4802 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"63be777b-838a-499d-a04c-c522ae8fcb2f\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-external-api-0" Dec 06 03:58:42 crc kubenswrapper[4802]: I1206 03:58:42.508913 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/63be777b-838a-499d-a04c-c522ae8fcb2f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"63be777b-838a-499d-a04c-c522ae8fcb2f\") " pod="openstack/glance-default-external-api-0" Dec 06 03:58:42 crc kubenswrapper[4802]: I1206 03:58:42.508925 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/63be777b-838a-499d-a04c-c522ae8fcb2f-logs\") pod \"glance-default-external-api-0\" (UID: \"63be777b-838a-499d-a04c-c522ae8fcb2f\") " pod="openstack/glance-default-external-api-0" Dec 06 03:58:42 crc kubenswrapper[4802]: I1206 03:58:42.516398 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63be777b-838a-499d-a04c-c522ae8fcb2f-config-data\") pod \"glance-default-external-api-0\" (UID: \"63be777b-838a-499d-a04c-c522ae8fcb2f\") " pod="openstack/glance-default-external-api-0" Dec 06 03:58:42 crc kubenswrapper[4802]: I1206 03:58:42.516575 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63be777b-838a-499d-a04c-c522ae8fcb2f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"63be777b-838a-499d-a04c-c522ae8fcb2f\") " pod="openstack/glance-default-external-api-0" Dec 06 03:58:42 crc kubenswrapper[4802]: I1206 03:58:42.516826 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63be777b-838a-499d-a04c-c522ae8fcb2f-scripts\") pod \"glance-default-external-api-0\" (UID: \"63be777b-838a-499d-a04c-c522ae8fcb2f\") " pod="openstack/glance-default-external-api-0" Dec 06 03:58:42 crc kubenswrapper[4802]: I1206 03:58:42.523607 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqt6l\" (UniqueName: \"kubernetes.io/projected/63be777b-838a-499d-a04c-c522ae8fcb2f-kube-api-access-lqt6l\") pod \"glance-default-external-api-0\" (UID: \"63be777b-838a-499d-a04c-c522ae8fcb2f\") " pod="openstack/glance-default-external-api-0" Dec 06 03:58:42 crc kubenswrapper[4802]: I1206 03:58:42.542583 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"63be777b-838a-499d-a04c-c522ae8fcb2f\") " pod="openstack/glance-default-external-api-0" Dec 06 03:58:42 crc kubenswrapper[4802]: I1206 03:58:42.648873 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 03:58:45 crc kubenswrapper[4802]: I1206 03:58:43.122658 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ec8cedbe-d81f-470c-a964-83e5a47a9db2","Type":"ContainerStarted","Data":"8d7e7a38230465761a6750596f1cb9ce45999172bf18816143ec857e49d26fc1"} Dec 06 03:58:45 crc kubenswrapper[4802]: I1206 03:58:43.283574 4802 patch_prober.go:28] interesting pod/machine-config-daemon-zpxxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 03:58:45 crc kubenswrapper[4802]: I1206 03:58:43.283622 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 03:58:45 crc kubenswrapper[4802]: I1206 03:58:43.400918 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 03:58:45 crc kubenswrapper[4802]: I1206 03:58:43.467270 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21cae3ac-9ae8-4c31-972d-7a3c25e8bd95" path="/var/lib/kubelet/pods/21cae3ac-9ae8-4c31-972d-7a3c25e8bd95/volumes" Dec 06 03:58:45 crc kubenswrapper[4802]: I1206 03:58:43.468251 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4" path="/var/lib/kubelet/pods/32ee40e7-4fe2-42c3-b4c9-0dc894eb17e4/volumes" Dec 06 03:58:45 crc kubenswrapper[4802]: I1206 03:58:43.468782 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb49a4b-5eee-498e-a5ba-584e6ee6d79a" path="/var/lib/kubelet/pods/3cb49a4b-5eee-498e-a5ba-584e6ee6d79a/volumes" Dec 06 03:58:45 crc kubenswrapper[4802]: I1206 03:58:44.147823 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ec8cedbe-d81f-470c-a964-83e5a47a9db2","Type":"ContainerStarted","Data":"8495b1cad08bd9b8cc419ef32705e16ac146ac9bcbe2464df1d2712f1a9e4db9"} Dec 06 03:58:45 crc kubenswrapper[4802]: I1206 03:58:44.147944 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="ec8cedbe-d81f-470c-a964-83e5a47a9db2" containerName="glance-log" containerID="cri-o://8d7e7a38230465761a6750596f1cb9ce45999172bf18816143ec857e49d26fc1" gracePeriod=30 Dec 06 03:58:45 crc kubenswrapper[4802]: I1206 03:58:44.147963 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="ec8cedbe-d81f-470c-a964-83e5a47a9db2" containerName="glance-httpd" containerID="cri-o://8495b1cad08bd9b8cc419ef32705e16ac146ac9bcbe2464df1d2712f1a9e4db9" gracePeriod=30 Dec 06 03:58:45 crc kubenswrapper[4802]: I1206 03:58:44.149627 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"63be777b-838a-499d-a04c-c522ae8fcb2f","Type":"ContainerStarted","Data":"73faaf4188dcb6c2b7516e1d6c3e449cfa4d3935d23f5695a286f1f60ad2ad72"} Dec 06 03:58:45 crc kubenswrapper[4802]: I1206 03:58:44.168842 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.168824533 podStartE2EDuration="5.168824533s" podCreationTimestamp="2025-12-06 03:58:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:58:44.163880938 +0000 UTC m=+1117.035790090" watchObservedRunningTime="2025-12-06 03:58:44.168824533 +0000 UTC m=+1117.040733685" Dec 06 03:58:45 crc kubenswrapper[4802]: I1206 03:58:45.163989 4802 generic.go:334] "Generic (PLEG): container finished" podID="ec8cedbe-d81f-470c-a964-83e5a47a9db2" containerID="8d7e7a38230465761a6750596f1cb9ce45999172bf18816143ec857e49d26fc1" exitCode=143 Dec 06 03:58:45 crc kubenswrapper[4802]: I1206 03:58:45.164070 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ec8cedbe-d81f-470c-a964-83e5a47a9db2","Type":"ContainerDied","Data":"8d7e7a38230465761a6750596f1cb9ce45999172bf18816143ec857e49d26fc1"} Dec 06 03:58:45 crc kubenswrapper[4802]: I1206 03:58:45.166729 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"63be777b-838a-499d-a04c-c522ae8fcb2f","Type":"ContainerStarted","Data":"19f6f568835b95f0c21c6e9c9ae4da59f444ec6a8fb7f93955d67e6748442c04"} Dec 06 03:58:46 crc kubenswrapper[4802]: I1206 03:58:46.175999 4802 generic.go:334] "Generic (PLEG): container finished" podID="ec8cedbe-d81f-470c-a964-83e5a47a9db2" containerID="8495b1cad08bd9b8cc419ef32705e16ac146ac9bcbe2464df1d2712f1a9e4db9" exitCode=0 Dec 06 03:58:46 crc kubenswrapper[4802]: I1206 03:58:46.176261 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ec8cedbe-d81f-470c-a964-83e5a47a9db2","Type":"ContainerDied","Data":"8495b1cad08bd9b8cc419ef32705e16ac146ac9bcbe2464df1d2712f1a9e4db9"} Dec 06 03:58:47 crc kubenswrapper[4802]: I1206 03:58:47.186075 4802 generic.go:334] "Generic (PLEG): container finished" podID="b24d756b-e69d-42c8-8f6c-0ea2a3d57d00" containerID="0f19de9d3ffb77620b7ecce801747092e84300ff9f114eff03946ce3f3ce3405" exitCode=0 Dec 06 03:58:47 crc kubenswrapper[4802]: I1206 03:58:47.186151 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-ddx8q" event={"ID":"b24d756b-e69d-42c8-8f6c-0ea2a3d57d00","Type":"ContainerDied","Data":"0f19de9d3ffb77620b7ecce801747092e84300ff9f114eff03946ce3f3ce3405"} Dec 06 03:58:47 crc kubenswrapper[4802]: I1206 03:58:47.615783 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 03:58:47 crc kubenswrapper[4802]: I1206 03:58:47.651422 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec8cedbe-d81f-470c-a964-83e5a47a9db2-logs\") pod \"ec8cedbe-d81f-470c-a964-83e5a47a9db2\" (UID: \"ec8cedbe-d81f-470c-a964-83e5a47a9db2\") " Dec 06 03:58:47 crc kubenswrapper[4802]: I1206 03:58:47.651503 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ec8cedbe-d81f-470c-a964-83e5a47a9db2-httpd-run\") pod \"ec8cedbe-d81f-470c-a964-83e5a47a9db2\" (UID: \"ec8cedbe-d81f-470c-a964-83e5a47a9db2\") " Dec 06 03:58:47 crc kubenswrapper[4802]: I1206 03:58:47.651538 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec8cedbe-d81f-470c-a964-83e5a47a9db2-scripts\") pod \"ec8cedbe-d81f-470c-a964-83e5a47a9db2\" (UID: \"ec8cedbe-d81f-470c-a964-83e5a47a9db2\") " Dec 06 03:58:47 crc kubenswrapper[4802]: I1206 03:58:47.651578 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t8b4t\" (UniqueName: \"kubernetes.io/projected/ec8cedbe-d81f-470c-a964-83e5a47a9db2-kube-api-access-t8b4t\") pod \"ec8cedbe-d81f-470c-a964-83e5a47a9db2\" (UID: \"ec8cedbe-d81f-470c-a964-83e5a47a9db2\") " Dec 06 03:58:47 crc kubenswrapper[4802]: I1206 03:58:47.651611 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec8cedbe-d81f-470c-a964-83e5a47a9db2-combined-ca-bundle\") pod \"ec8cedbe-d81f-470c-a964-83e5a47a9db2\" (UID: \"ec8cedbe-d81f-470c-a964-83e5a47a9db2\") " Dec 06 03:58:47 crc kubenswrapper[4802]: I1206 03:58:47.651684 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec8cedbe-d81f-470c-a964-83e5a47a9db2-config-data\") pod \"ec8cedbe-d81f-470c-a964-83e5a47a9db2\" (UID: \"ec8cedbe-d81f-470c-a964-83e5a47a9db2\") " Dec 06 03:58:47 crc kubenswrapper[4802]: I1206 03:58:47.651744 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ec8cedbe-d81f-470c-a964-83e5a47a9db2\" (UID: \"ec8cedbe-d81f-470c-a964-83e5a47a9db2\") " Dec 06 03:58:47 crc kubenswrapper[4802]: I1206 03:58:47.658103 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec8cedbe-d81f-470c-a964-83e5a47a9db2-logs" (OuterVolumeSpecName: "logs") pod "ec8cedbe-d81f-470c-a964-83e5a47a9db2" (UID: "ec8cedbe-d81f-470c-a964-83e5a47a9db2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:58:47 crc kubenswrapper[4802]: I1206 03:58:47.658173 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "ec8cedbe-d81f-470c-a964-83e5a47a9db2" (UID: "ec8cedbe-d81f-470c-a964-83e5a47a9db2"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 03:58:47 crc kubenswrapper[4802]: I1206 03:58:47.658446 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec8cedbe-d81f-470c-a964-83e5a47a9db2-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "ec8cedbe-d81f-470c-a964-83e5a47a9db2" (UID: "ec8cedbe-d81f-470c-a964-83e5a47a9db2"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:58:47 crc kubenswrapper[4802]: I1206 03:58:47.665931 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec8cedbe-d81f-470c-a964-83e5a47a9db2-scripts" (OuterVolumeSpecName: "scripts") pod "ec8cedbe-d81f-470c-a964-83e5a47a9db2" (UID: "ec8cedbe-d81f-470c-a964-83e5a47a9db2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:58:47 crc kubenswrapper[4802]: I1206 03:58:47.668514 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec8cedbe-d81f-470c-a964-83e5a47a9db2-kube-api-access-t8b4t" (OuterVolumeSpecName: "kube-api-access-t8b4t") pod "ec8cedbe-d81f-470c-a964-83e5a47a9db2" (UID: "ec8cedbe-d81f-470c-a964-83e5a47a9db2"). InnerVolumeSpecName "kube-api-access-t8b4t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:58:47 crc kubenswrapper[4802]: I1206 03:58:47.694082 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec8cedbe-d81f-470c-a964-83e5a47a9db2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ec8cedbe-d81f-470c-a964-83e5a47a9db2" (UID: "ec8cedbe-d81f-470c-a964-83e5a47a9db2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:58:47 crc kubenswrapper[4802]: I1206 03:58:47.708960 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec8cedbe-d81f-470c-a964-83e5a47a9db2-config-data" (OuterVolumeSpecName: "config-data") pod "ec8cedbe-d81f-470c-a964-83e5a47a9db2" (UID: "ec8cedbe-d81f-470c-a964-83e5a47a9db2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:58:47 crc kubenswrapper[4802]: I1206 03:58:47.758965 4802 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec8cedbe-d81f-470c-a964-83e5a47a9db2-logs\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:47 crc kubenswrapper[4802]: I1206 03:58:47.759017 4802 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ec8cedbe-d81f-470c-a964-83e5a47a9db2-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:47 crc kubenswrapper[4802]: I1206 03:58:47.759030 4802 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec8cedbe-d81f-470c-a964-83e5a47a9db2-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:47 crc kubenswrapper[4802]: I1206 03:58:47.759041 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t8b4t\" (UniqueName: \"kubernetes.io/projected/ec8cedbe-d81f-470c-a964-83e5a47a9db2-kube-api-access-t8b4t\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:47 crc kubenswrapper[4802]: I1206 03:58:47.759054 4802 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec8cedbe-d81f-470c-a964-83e5a47a9db2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:47 crc kubenswrapper[4802]: I1206 03:58:47.759065 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec8cedbe-d81f-470c-a964-83e5a47a9db2-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:47 crc kubenswrapper[4802]: I1206 03:58:47.759108 4802 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Dec 06 03:58:47 crc kubenswrapper[4802]: I1206 03:58:47.780547 4802 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Dec 06 03:58:47 crc kubenswrapper[4802]: I1206 03:58:47.861185 4802 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:48 crc kubenswrapper[4802]: I1206 03:58:48.205570 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 03:58:48 crc kubenswrapper[4802]: I1206 03:58:48.209429 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ec8cedbe-d81f-470c-a964-83e5a47a9db2","Type":"ContainerDied","Data":"4f34db756bb1581336b9b9813aff7202e5d250ae710d7ad6961772015e92c7d2"} Dec 06 03:58:48 crc kubenswrapper[4802]: I1206 03:58:48.209496 4802 scope.go:117] "RemoveContainer" containerID="8495b1cad08bd9b8cc419ef32705e16ac146ac9bcbe2464df1d2712f1a9e4db9" Dec 06 03:58:48 crc kubenswrapper[4802]: I1206 03:58:48.293804 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 03:58:48 crc kubenswrapper[4802]: I1206 03:58:48.317803 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 03:58:48 crc kubenswrapper[4802]: I1206 03:58:48.337599 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 03:58:48 crc kubenswrapper[4802]: E1206 03:58:48.338039 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec8cedbe-d81f-470c-a964-83e5a47a9db2" containerName="glance-httpd" Dec 06 03:58:48 crc kubenswrapper[4802]: I1206 03:58:48.338055 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec8cedbe-d81f-470c-a964-83e5a47a9db2" containerName="glance-httpd" Dec 06 03:58:48 crc kubenswrapper[4802]: E1206 03:58:48.338074 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec8cedbe-d81f-470c-a964-83e5a47a9db2" containerName="glance-log" Dec 06 03:58:48 crc kubenswrapper[4802]: I1206 03:58:48.338080 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec8cedbe-d81f-470c-a964-83e5a47a9db2" containerName="glance-log" Dec 06 03:58:48 crc kubenswrapper[4802]: I1206 03:58:48.338259 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec8cedbe-d81f-470c-a964-83e5a47a9db2" containerName="glance-log" Dec 06 03:58:48 crc kubenswrapper[4802]: I1206 03:58:48.338278 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec8cedbe-d81f-470c-a964-83e5a47a9db2" containerName="glance-httpd" Dec 06 03:58:48 crc kubenswrapper[4802]: I1206 03:58:48.339121 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 03:58:48 crc kubenswrapper[4802]: I1206 03:58:48.343012 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 06 03:58:48 crc kubenswrapper[4802]: I1206 03:58:48.354728 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 03:58:48 crc kubenswrapper[4802]: I1206 03:58:48.373157 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sczlv\" (UniqueName: \"kubernetes.io/projected/dd66ded5-a0ce-42a0-86a6-03a727128fb9-kube-api-access-sczlv\") pod \"glance-default-internal-api-0\" (UID: \"dd66ded5-a0ce-42a0-86a6-03a727128fb9\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:58:48 crc kubenswrapper[4802]: I1206 03:58:48.373205 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd66ded5-a0ce-42a0-86a6-03a727128fb9-logs\") pod \"glance-default-internal-api-0\" (UID: \"dd66ded5-a0ce-42a0-86a6-03a727128fb9\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:58:48 crc kubenswrapper[4802]: I1206 03:58:48.373249 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"dd66ded5-a0ce-42a0-86a6-03a727128fb9\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:58:48 crc kubenswrapper[4802]: I1206 03:58:48.373268 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd66ded5-a0ce-42a0-86a6-03a727128fb9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"dd66ded5-a0ce-42a0-86a6-03a727128fb9\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:58:48 crc kubenswrapper[4802]: I1206 03:58:48.373292 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/dd66ded5-a0ce-42a0-86a6-03a727128fb9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"dd66ded5-a0ce-42a0-86a6-03a727128fb9\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:58:48 crc kubenswrapper[4802]: I1206 03:58:48.373362 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dd66ded5-a0ce-42a0-86a6-03a727128fb9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"dd66ded5-a0ce-42a0-86a6-03a727128fb9\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:58:48 crc kubenswrapper[4802]: I1206 03:58:48.373399 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd66ded5-a0ce-42a0-86a6-03a727128fb9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"dd66ded5-a0ce-42a0-86a6-03a727128fb9\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:58:48 crc kubenswrapper[4802]: I1206 03:58:48.475293 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"dd66ded5-a0ce-42a0-86a6-03a727128fb9\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:58:48 crc kubenswrapper[4802]: I1206 03:58:48.475600 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd66ded5-a0ce-42a0-86a6-03a727128fb9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"dd66ded5-a0ce-42a0-86a6-03a727128fb9\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:58:48 crc kubenswrapper[4802]: I1206 03:58:48.475631 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/dd66ded5-a0ce-42a0-86a6-03a727128fb9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"dd66ded5-a0ce-42a0-86a6-03a727128fb9\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:58:48 crc kubenswrapper[4802]: I1206 03:58:48.475710 4802 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"dd66ded5-a0ce-42a0-86a6-03a727128fb9\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-internal-api-0" Dec 06 03:58:48 crc kubenswrapper[4802]: I1206 03:58:48.475715 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dd66ded5-a0ce-42a0-86a6-03a727128fb9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"dd66ded5-a0ce-42a0-86a6-03a727128fb9\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:58:48 crc kubenswrapper[4802]: I1206 03:58:48.475919 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd66ded5-a0ce-42a0-86a6-03a727128fb9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"dd66ded5-a0ce-42a0-86a6-03a727128fb9\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:58:48 crc kubenswrapper[4802]: I1206 03:58:48.476009 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sczlv\" (UniqueName: \"kubernetes.io/projected/dd66ded5-a0ce-42a0-86a6-03a727128fb9-kube-api-access-sczlv\") pod \"glance-default-internal-api-0\" (UID: \"dd66ded5-a0ce-42a0-86a6-03a727128fb9\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:58:48 crc kubenswrapper[4802]: I1206 03:58:48.476038 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd66ded5-a0ce-42a0-86a6-03a727128fb9-logs\") pod \"glance-default-internal-api-0\" (UID: \"dd66ded5-a0ce-42a0-86a6-03a727128fb9\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:58:48 crc kubenswrapper[4802]: I1206 03:58:48.476174 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/dd66ded5-a0ce-42a0-86a6-03a727128fb9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"dd66ded5-a0ce-42a0-86a6-03a727128fb9\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:58:48 crc kubenswrapper[4802]: I1206 03:58:48.476424 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd66ded5-a0ce-42a0-86a6-03a727128fb9-logs\") pod \"glance-default-internal-api-0\" (UID: \"dd66ded5-a0ce-42a0-86a6-03a727128fb9\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:58:48 crc kubenswrapper[4802]: I1206 03:58:48.480667 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dd66ded5-a0ce-42a0-86a6-03a727128fb9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"dd66ded5-a0ce-42a0-86a6-03a727128fb9\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:58:48 crc kubenswrapper[4802]: I1206 03:58:48.481870 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd66ded5-a0ce-42a0-86a6-03a727128fb9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"dd66ded5-a0ce-42a0-86a6-03a727128fb9\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:58:48 crc kubenswrapper[4802]: I1206 03:58:48.482511 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd66ded5-a0ce-42a0-86a6-03a727128fb9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"dd66ded5-a0ce-42a0-86a6-03a727128fb9\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:58:48 crc kubenswrapper[4802]: I1206 03:58:48.502348 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sczlv\" (UniqueName: \"kubernetes.io/projected/dd66ded5-a0ce-42a0-86a6-03a727128fb9-kube-api-access-sczlv\") pod \"glance-default-internal-api-0\" (UID: \"dd66ded5-a0ce-42a0-86a6-03a727128fb9\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:58:48 crc kubenswrapper[4802]: I1206 03:58:48.528998 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"dd66ded5-a0ce-42a0-86a6-03a727128fb9\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:58:48 crc kubenswrapper[4802]: I1206 03:58:48.542050 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 03:58:48 crc kubenswrapper[4802]: I1206 03:58:48.607899 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 03:58:48 crc kubenswrapper[4802]: I1206 03:58:48.608490 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 03:58:49 crc kubenswrapper[4802]: I1206 03:58:49.466061 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec8cedbe-d81f-470c-a964-83e5a47a9db2" path="/var/lib/kubelet/pods/ec8cedbe-d81f-470c-a964-83e5a47a9db2/volumes" Dec 06 03:58:49 crc kubenswrapper[4802]: I1206 03:58:49.610989 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-56df8fb6b7-z4x9c" Dec 06 03:58:49 crc kubenswrapper[4802]: I1206 03:58:49.684355 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-77frt"] Dec 06 03:58:49 crc kubenswrapper[4802]: I1206 03:58:49.684596 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b8fbc5445-77frt" podUID="5cf08e89-5a37-4f63-8067-e8c161437deb" containerName="dnsmasq-dns" containerID="cri-o://c83ceab3315891cab0cb020370c47d360ceda11dd02b35e3f9db75790ec648da" gracePeriod=10 Dec 06 03:58:50 crc kubenswrapper[4802]: I1206 03:58:50.223080 4802 generic.go:334] "Generic (PLEG): container finished" podID="5cf08e89-5a37-4f63-8067-e8c161437deb" containerID="c83ceab3315891cab0cb020370c47d360ceda11dd02b35e3f9db75790ec648da" exitCode=0 Dec 06 03:58:50 crc kubenswrapper[4802]: I1206 03:58:50.223131 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-77frt" event={"ID":"5cf08e89-5a37-4f63-8067-e8c161437deb","Type":"ContainerDied","Data":"c83ceab3315891cab0cb020370c47d360ceda11dd02b35e3f9db75790ec648da"} Dec 06 03:58:52 crc kubenswrapper[4802]: I1206 03:58:52.659229 4802 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-77frt" podUID="5cf08e89-5a37-4f63-8067-e8c161437deb" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.118:5353: connect: connection refused" Dec 06 03:58:57 crc kubenswrapper[4802]: I1206 03:58:57.659663 4802 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-77frt" podUID="5cf08e89-5a37-4f63-8067-e8c161437deb" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.118:5353: connect: connection refused" Dec 06 03:58:59 crc kubenswrapper[4802]: I1206 03:58:59.301666 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-ddx8q" event={"ID":"b24d756b-e69d-42c8-8f6c-0ea2a3d57d00","Type":"ContainerDied","Data":"bb2592161c9b1b7b1be4815a96ef98534ba2a1228433ba4eeaeb74cab39d3fba"} Dec 06 03:58:59 crc kubenswrapper[4802]: I1206 03:58:59.301740 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bb2592161c9b1b7b1be4815a96ef98534ba2a1228433ba4eeaeb74cab39d3fba" Dec 06 03:58:59 crc kubenswrapper[4802]: I1206 03:58:59.350973 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-ddx8q" Dec 06 03:58:59 crc kubenswrapper[4802]: I1206 03:58:59.391654 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b24d756b-e69d-42c8-8f6c-0ea2a3d57d00-fernet-keys\") pod \"b24d756b-e69d-42c8-8f6c-0ea2a3d57d00\" (UID: \"b24d756b-e69d-42c8-8f6c-0ea2a3d57d00\") " Dec 06 03:58:59 crc kubenswrapper[4802]: I1206 03:58:59.391701 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cr2zr\" (UniqueName: \"kubernetes.io/projected/b24d756b-e69d-42c8-8f6c-0ea2a3d57d00-kube-api-access-cr2zr\") pod \"b24d756b-e69d-42c8-8f6c-0ea2a3d57d00\" (UID: \"b24d756b-e69d-42c8-8f6c-0ea2a3d57d00\") " Dec 06 03:58:59 crc kubenswrapper[4802]: I1206 03:58:59.391735 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b24d756b-e69d-42c8-8f6c-0ea2a3d57d00-scripts\") pod \"b24d756b-e69d-42c8-8f6c-0ea2a3d57d00\" (UID: \"b24d756b-e69d-42c8-8f6c-0ea2a3d57d00\") " Dec 06 03:58:59 crc kubenswrapper[4802]: I1206 03:58:59.391917 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b24d756b-e69d-42c8-8f6c-0ea2a3d57d00-credential-keys\") pod \"b24d756b-e69d-42c8-8f6c-0ea2a3d57d00\" (UID: \"b24d756b-e69d-42c8-8f6c-0ea2a3d57d00\") " Dec 06 03:58:59 crc kubenswrapper[4802]: I1206 03:58:59.391938 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b24d756b-e69d-42c8-8f6c-0ea2a3d57d00-config-data\") pod \"b24d756b-e69d-42c8-8f6c-0ea2a3d57d00\" (UID: \"b24d756b-e69d-42c8-8f6c-0ea2a3d57d00\") " Dec 06 03:58:59 crc kubenswrapper[4802]: I1206 03:58:59.391959 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b24d756b-e69d-42c8-8f6c-0ea2a3d57d00-combined-ca-bundle\") pod \"b24d756b-e69d-42c8-8f6c-0ea2a3d57d00\" (UID: \"b24d756b-e69d-42c8-8f6c-0ea2a3d57d00\") " Dec 06 03:58:59 crc kubenswrapper[4802]: I1206 03:58:59.398640 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b24d756b-e69d-42c8-8f6c-0ea2a3d57d00-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "b24d756b-e69d-42c8-8f6c-0ea2a3d57d00" (UID: "b24d756b-e69d-42c8-8f6c-0ea2a3d57d00"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:58:59 crc kubenswrapper[4802]: I1206 03:58:59.402498 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b24d756b-e69d-42c8-8f6c-0ea2a3d57d00-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "b24d756b-e69d-42c8-8f6c-0ea2a3d57d00" (UID: "b24d756b-e69d-42c8-8f6c-0ea2a3d57d00"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:58:59 crc kubenswrapper[4802]: I1206 03:58:59.415698 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b24d756b-e69d-42c8-8f6c-0ea2a3d57d00-kube-api-access-cr2zr" (OuterVolumeSpecName: "kube-api-access-cr2zr") pod "b24d756b-e69d-42c8-8f6c-0ea2a3d57d00" (UID: "b24d756b-e69d-42c8-8f6c-0ea2a3d57d00"). InnerVolumeSpecName "kube-api-access-cr2zr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:58:59 crc kubenswrapper[4802]: I1206 03:58:59.419020 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b24d756b-e69d-42c8-8f6c-0ea2a3d57d00-config-data" (OuterVolumeSpecName: "config-data") pod "b24d756b-e69d-42c8-8f6c-0ea2a3d57d00" (UID: "b24d756b-e69d-42c8-8f6c-0ea2a3d57d00"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:58:59 crc kubenswrapper[4802]: I1206 03:58:59.420997 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b24d756b-e69d-42c8-8f6c-0ea2a3d57d00-scripts" (OuterVolumeSpecName: "scripts") pod "b24d756b-e69d-42c8-8f6c-0ea2a3d57d00" (UID: "b24d756b-e69d-42c8-8f6c-0ea2a3d57d00"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:58:59 crc kubenswrapper[4802]: I1206 03:58:59.453731 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b24d756b-e69d-42c8-8f6c-0ea2a3d57d00-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b24d756b-e69d-42c8-8f6c-0ea2a3d57d00" (UID: "b24d756b-e69d-42c8-8f6c-0ea2a3d57d00"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:58:59 crc kubenswrapper[4802]: I1206 03:58:59.494524 4802 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b24d756b-e69d-42c8-8f6c-0ea2a3d57d00-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:59 crc kubenswrapper[4802]: I1206 03:58:59.494919 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cr2zr\" (UniqueName: \"kubernetes.io/projected/b24d756b-e69d-42c8-8f6c-0ea2a3d57d00-kube-api-access-cr2zr\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:59 crc kubenswrapper[4802]: I1206 03:58:59.494948 4802 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b24d756b-e69d-42c8-8f6c-0ea2a3d57d00-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:59 crc kubenswrapper[4802]: I1206 03:58:59.494962 4802 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b24d756b-e69d-42c8-8f6c-0ea2a3d57d00-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:59 crc kubenswrapper[4802]: I1206 03:58:59.494977 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b24d756b-e69d-42c8-8f6c-0ea2a3d57d00-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 03:58:59 crc kubenswrapper[4802]: I1206 03:58:59.494989 4802 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b24d756b-e69d-42c8-8f6c-0ea2a3d57d00-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:00 crc kubenswrapper[4802]: I1206 03:59:00.307882 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-ddx8q" Dec 06 03:59:00 crc kubenswrapper[4802]: I1206 03:59:00.420456 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-ddx8q"] Dec 06 03:59:00 crc kubenswrapper[4802]: I1206 03:59:00.432171 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-ddx8q"] Dec 06 03:59:00 crc kubenswrapper[4802]: I1206 03:59:00.536412 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-twnc4"] Dec 06 03:59:00 crc kubenswrapper[4802]: E1206 03:59:00.537394 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b24d756b-e69d-42c8-8f6c-0ea2a3d57d00" containerName="keystone-bootstrap" Dec 06 03:59:00 crc kubenswrapper[4802]: I1206 03:59:00.537486 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="b24d756b-e69d-42c8-8f6c-0ea2a3d57d00" containerName="keystone-bootstrap" Dec 06 03:59:00 crc kubenswrapper[4802]: I1206 03:59:00.537713 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="b24d756b-e69d-42c8-8f6c-0ea2a3d57d00" containerName="keystone-bootstrap" Dec 06 03:59:00 crc kubenswrapper[4802]: I1206 03:59:00.538428 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-twnc4" Dec 06 03:59:00 crc kubenswrapper[4802]: I1206 03:59:00.540874 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 06 03:59:00 crc kubenswrapper[4802]: I1206 03:59:00.541173 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 06 03:59:00 crc kubenswrapper[4802]: I1206 03:59:00.541508 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 06 03:59:00 crc kubenswrapper[4802]: I1206 03:59:00.541736 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 06 03:59:00 crc kubenswrapper[4802]: I1206 03:59:00.542743 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-zk4kp" Dec 06 03:59:00 crc kubenswrapper[4802]: I1206 03:59:00.547939 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-twnc4"] Dec 06 03:59:00 crc kubenswrapper[4802]: I1206 03:59:00.613841 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f33c46c6-6721-4c7a-a602-48351824838a-scripts\") pod \"keystone-bootstrap-twnc4\" (UID: \"f33c46c6-6721-4c7a-a602-48351824838a\") " pod="openstack/keystone-bootstrap-twnc4" Dec 06 03:59:00 crc kubenswrapper[4802]: I1206 03:59:00.614169 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f33c46c6-6721-4c7a-a602-48351824838a-config-data\") pod \"keystone-bootstrap-twnc4\" (UID: \"f33c46c6-6721-4c7a-a602-48351824838a\") " pod="openstack/keystone-bootstrap-twnc4" Dec 06 03:59:00 crc kubenswrapper[4802]: I1206 03:59:00.614288 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f33c46c6-6721-4c7a-a602-48351824838a-credential-keys\") pod \"keystone-bootstrap-twnc4\" (UID: \"f33c46c6-6721-4c7a-a602-48351824838a\") " pod="openstack/keystone-bootstrap-twnc4" Dec 06 03:59:00 crc kubenswrapper[4802]: I1206 03:59:00.614488 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntxvn\" (UniqueName: \"kubernetes.io/projected/f33c46c6-6721-4c7a-a602-48351824838a-kube-api-access-ntxvn\") pod \"keystone-bootstrap-twnc4\" (UID: \"f33c46c6-6721-4c7a-a602-48351824838a\") " pod="openstack/keystone-bootstrap-twnc4" Dec 06 03:59:00 crc kubenswrapper[4802]: I1206 03:59:00.614614 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f33c46c6-6721-4c7a-a602-48351824838a-fernet-keys\") pod \"keystone-bootstrap-twnc4\" (UID: \"f33c46c6-6721-4c7a-a602-48351824838a\") " pod="openstack/keystone-bootstrap-twnc4" Dec 06 03:59:00 crc kubenswrapper[4802]: I1206 03:59:00.614647 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f33c46c6-6721-4c7a-a602-48351824838a-combined-ca-bundle\") pod \"keystone-bootstrap-twnc4\" (UID: \"f33c46c6-6721-4c7a-a602-48351824838a\") " pod="openstack/keystone-bootstrap-twnc4" Dec 06 03:59:00 crc kubenswrapper[4802]: I1206 03:59:00.715593 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntxvn\" (UniqueName: \"kubernetes.io/projected/f33c46c6-6721-4c7a-a602-48351824838a-kube-api-access-ntxvn\") pod \"keystone-bootstrap-twnc4\" (UID: \"f33c46c6-6721-4c7a-a602-48351824838a\") " pod="openstack/keystone-bootstrap-twnc4" Dec 06 03:59:00 crc kubenswrapper[4802]: I1206 03:59:00.715652 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f33c46c6-6721-4c7a-a602-48351824838a-fernet-keys\") pod \"keystone-bootstrap-twnc4\" (UID: \"f33c46c6-6721-4c7a-a602-48351824838a\") " pod="openstack/keystone-bootstrap-twnc4" Dec 06 03:59:00 crc kubenswrapper[4802]: I1206 03:59:00.715676 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f33c46c6-6721-4c7a-a602-48351824838a-combined-ca-bundle\") pod \"keystone-bootstrap-twnc4\" (UID: \"f33c46c6-6721-4c7a-a602-48351824838a\") " pod="openstack/keystone-bootstrap-twnc4" Dec 06 03:59:00 crc kubenswrapper[4802]: I1206 03:59:00.715734 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f33c46c6-6721-4c7a-a602-48351824838a-scripts\") pod \"keystone-bootstrap-twnc4\" (UID: \"f33c46c6-6721-4c7a-a602-48351824838a\") " pod="openstack/keystone-bootstrap-twnc4" Dec 06 03:59:00 crc kubenswrapper[4802]: I1206 03:59:00.715784 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f33c46c6-6721-4c7a-a602-48351824838a-config-data\") pod \"keystone-bootstrap-twnc4\" (UID: \"f33c46c6-6721-4c7a-a602-48351824838a\") " pod="openstack/keystone-bootstrap-twnc4" Dec 06 03:59:00 crc kubenswrapper[4802]: I1206 03:59:00.715811 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f33c46c6-6721-4c7a-a602-48351824838a-credential-keys\") pod \"keystone-bootstrap-twnc4\" (UID: \"f33c46c6-6721-4c7a-a602-48351824838a\") " pod="openstack/keystone-bootstrap-twnc4" Dec 06 03:59:00 crc kubenswrapper[4802]: I1206 03:59:00.721331 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f33c46c6-6721-4c7a-a602-48351824838a-credential-keys\") pod \"keystone-bootstrap-twnc4\" (UID: \"f33c46c6-6721-4c7a-a602-48351824838a\") " pod="openstack/keystone-bootstrap-twnc4" Dec 06 03:59:00 crc kubenswrapper[4802]: I1206 03:59:00.722064 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f33c46c6-6721-4c7a-a602-48351824838a-config-data\") pod \"keystone-bootstrap-twnc4\" (UID: \"f33c46c6-6721-4c7a-a602-48351824838a\") " pod="openstack/keystone-bootstrap-twnc4" Dec 06 03:59:00 crc kubenswrapper[4802]: I1206 03:59:00.723152 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f33c46c6-6721-4c7a-a602-48351824838a-combined-ca-bundle\") pod \"keystone-bootstrap-twnc4\" (UID: \"f33c46c6-6721-4c7a-a602-48351824838a\") " pod="openstack/keystone-bootstrap-twnc4" Dec 06 03:59:00 crc kubenswrapper[4802]: I1206 03:59:00.724062 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f33c46c6-6721-4c7a-a602-48351824838a-fernet-keys\") pod \"keystone-bootstrap-twnc4\" (UID: \"f33c46c6-6721-4c7a-a602-48351824838a\") " pod="openstack/keystone-bootstrap-twnc4" Dec 06 03:59:00 crc kubenswrapper[4802]: I1206 03:59:00.724966 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f33c46c6-6721-4c7a-a602-48351824838a-scripts\") pod \"keystone-bootstrap-twnc4\" (UID: \"f33c46c6-6721-4c7a-a602-48351824838a\") " pod="openstack/keystone-bootstrap-twnc4" Dec 06 03:59:00 crc kubenswrapper[4802]: I1206 03:59:00.731334 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntxvn\" (UniqueName: \"kubernetes.io/projected/f33c46c6-6721-4c7a-a602-48351824838a-kube-api-access-ntxvn\") pod \"keystone-bootstrap-twnc4\" (UID: \"f33c46c6-6721-4c7a-a602-48351824838a\") " pod="openstack/keystone-bootstrap-twnc4" Dec 06 03:59:00 crc kubenswrapper[4802]: I1206 03:59:00.862441 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-twnc4" Dec 06 03:59:01 crc kubenswrapper[4802]: I1206 03:59:01.470454 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b24d756b-e69d-42c8-8f6c-0ea2a3d57d00" path="/var/lib/kubelet/pods/b24d756b-e69d-42c8-8f6c-0ea2a3d57d00/volumes" Dec 06 03:59:07 crc kubenswrapper[4802]: I1206 03:59:07.659717 4802 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-77frt" podUID="5cf08e89-5a37-4f63-8067-e8c161437deb" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.118:5353: i/o timeout" Dec 06 03:59:07 crc kubenswrapper[4802]: I1206 03:59:07.660536 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b8fbc5445-77frt" Dec 06 03:59:08 crc kubenswrapper[4802]: E1206 03:59:08.136926 4802 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Dec 06 03:59:08 crc kubenswrapper[4802]: E1206 03:59:08.137307 4802 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n694h655h67ch67chc5h664hb5hcbh645h57fh5fch77h9bhc8hc6h574h9dh559h67ch57ch9ch5fbh67fh6dh65dh96h64h597h59chb7h54dh7bq,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fjnp6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(4d4f4fff-a6ec-4fef-82b4-6ba392fd297c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 03:59:08 crc kubenswrapper[4802]: I1206 03:59:08.992461 4802 scope.go:117] "RemoveContainer" containerID="8d7e7a38230465761a6750596f1cb9ce45999172bf18816143ec857e49d26fc1" Dec 06 03:59:09 crc kubenswrapper[4802]: E1206 03:59:09.036504 4802 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified" Dec 06 03:59:09 crc kubenswrapper[4802]: E1206 03:59:09.036773 4802 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-db-sync,Image:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,Command:[/bin/bash],Args:[-c /usr/bin/heat-manage --config-dir /etc/heat/heat.conf.d db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/heat/heat.conf.d/00-default.conf,SubPath:00-default.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/heat/heat.conf.d/01-custom.conf,SubPath:01-custom.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6ssjb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-db-sync-2hzlt_openstack(2a639667-be04-429f-a2ce-b20a011425f8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 03:59:09 crc kubenswrapper[4802]: E1206 03:59:09.039785 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/heat-db-sync-2hzlt" podUID="2a639667-be04-429f-a2ce-b20a011425f8" Dec 06 03:59:09 crc kubenswrapper[4802]: I1206 03:59:09.125007 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-77frt" Dec 06 03:59:09 crc kubenswrapper[4802]: I1206 03:59:09.168549 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-44thn\" (UniqueName: \"kubernetes.io/projected/5cf08e89-5a37-4f63-8067-e8c161437deb-kube-api-access-44thn\") pod \"5cf08e89-5a37-4f63-8067-e8c161437deb\" (UID: \"5cf08e89-5a37-4f63-8067-e8c161437deb\") " Dec 06 03:59:09 crc kubenswrapper[4802]: I1206 03:59:09.168871 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5cf08e89-5a37-4f63-8067-e8c161437deb-dns-svc\") pod \"5cf08e89-5a37-4f63-8067-e8c161437deb\" (UID: \"5cf08e89-5a37-4f63-8067-e8c161437deb\") " Dec 06 03:59:09 crc kubenswrapper[4802]: I1206 03:59:09.168958 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5cf08e89-5a37-4f63-8067-e8c161437deb-config\") pod \"5cf08e89-5a37-4f63-8067-e8c161437deb\" (UID: \"5cf08e89-5a37-4f63-8067-e8c161437deb\") " Dec 06 03:59:09 crc kubenswrapper[4802]: I1206 03:59:09.169027 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5cf08e89-5a37-4f63-8067-e8c161437deb-ovsdbserver-sb\") pod \"5cf08e89-5a37-4f63-8067-e8c161437deb\" (UID: \"5cf08e89-5a37-4f63-8067-e8c161437deb\") " Dec 06 03:59:09 crc kubenswrapper[4802]: I1206 03:59:09.169155 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5cf08e89-5a37-4f63-8067-e8c161437deb-ovsdbserver-nb\") pod \"5cf08e89-5a37-4f63-8067-e8c161437deb\" (UID: \"5cf08e89-5a37-4f63-8067-e8c161437deb\") " Dec 06 03:59:09 crc kubenswrapper[4802]: I1206 03:59:09.197538 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cf08e89-5a37-4f63-8067-e8c161437deb-kube-api-access-44thn" (OuterVolumeSpecName: "kube-api-access-44thn") pod "5cf08e89-5a37-4f63-8067-e8c161437deb" (UID: "5cf08e89-5a37-4f63-8067-e8c161437deb"). InnerVolumeSpecName "kube-api-access-44thn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:59:09 crc kubenswrapper[4802]: I1206 03:59:09.215096 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5cf08e89-5a37-4f63-8067-e8c161437deb-config" (OuterVolumeSpecName: "config") pod "5cf08e89-5a37-4f63-8067-e8c161437deb" (UID: "5cf08e89-5a37-4f63-8067-e8c161437deb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:59:09 crc kubenswrapper[4802]: I1206 03:59:09.229794 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5cf08e89-5a37-4f63-8067-e8c161437deb-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5cf08e89-5a37-4f63-8067-e8c161437deb" (UID: "5cf08e89-5a37-4f63-8067-e8c161437deb"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:59:09 crc kubenswrapper[4802]: I1206 03:59:09.233684 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5cf08e89-5a37-4f63-8067-e8c161437deb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5cf08e89-5a37-4f63-8067-e8c161437deb" (UID: "5cf08e89-5a37-4f63-8067-e8c161437deb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:59:09 crc kubenswrapper[4802]: I1206 03:59:09.234648 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5cf08e89-5a37-4f63-8067-e8c161437deb-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5cf08e89-5a37-4f63-8067-e8c161437deb" (UID: "5cf08e89-5a37-4f63-8067-e8c161437deb"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:59:09 crc kubenswrapper[4802]: I1206 03:59:09.271677 4802 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5cf08e89-5a37-4f63-8067-e8c161437deb-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:09 crc kubenswrapper[4802]: I1206 03:59:09.272282 4802 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5cf08e89-5a37-4f63-8067-e8c161437deb-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:09 crc kubenswrapper[4802]: I1206 03:59:09.272502 4802 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5cf08e89-5a37-4f63-8067-e8c161437deb-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:09 crc kubenswrapper[4802]: I1206 03:59:09.272614 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-44thn\" (UniqueName: \"kubernetes.io/projected/5cf08e89-5a37-4f63-8067-e8c161437deb-kube-api-access-44thn\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:09 crc kubenswrapper[4802]: I1206 03:59:09.272679 4802 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5cf08e89-5a37-4f63-8067-e8c161437deb-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:09 crc kubenswrapper[4802]: I1206 03:59:09.385076 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-77frt" event={"ID":"5cf08e89-5a37-4f63-8067-e8c161437deb","Type":"ContainerDied","Data":"6b964fa18fa81f1dd9eea2df8466f8fb04be19ebac233f2616c463b64af9ddc9"} Dec 06 03:59:09 crc kubenswrapper[4802]: I1206 03:59:09.385164 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-77frt" Dec 06 03:59:09 crc kubenswrapper[4802]: E1206 03:59:09.389991 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified\\\"\"" pod="openstack/heat-db-sync-2hzlt" podUID="2a639667-be04-429f-a2ce-b20a011425f8" Dec 06 03:59:09 crc kubenswrapper[4802]: I1206 03:59:09.443450 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-77frt"] Dec 06 03:59:09 crc kubenswrapper[4802]: I1206 03:59:09.481352 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-77frt"] Dec 06 03:59:10 crc kubenswrapper[4802]: E1206 03:59:10.186012 4802 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Dec 06 03:59:10 crc kubenswrapper[4802]: E1206 03:59:10.186206 4802 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pln97,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-5hj5p_openstack(f1644907-b4b1-4e23-a273-19d535bd7f88): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 03:59:10 crc kubenswrapper[4802]: E1206 03:59:10.187570 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-5hj5p" podUID="f1644907-b4b1-4e23-a273-19d535bd7f88" Dec 06 03:59:10 crc kubenswrapper[4802]: I1206 03:59:10.196284 4802 scope.go:117] "RemoveContainer" containerID="c83ceab3315891cab0cb020370c47d360ceda11dd02b35e3f9db75790ec648da" Dec 06 03:59:10 crc kubenswrapper[4802]: E1206 03:59:10.402095 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-5hj5p" podUID="f1644907-b4b1-4e23-a273-19d535bd7f88" Dec 06 03:59:10 crc kubenswrapper[4802]: I1206 03:59:10.632814 4802 scope.go:117] "RemoveContainer" containerID="850ac4aa9d4757681aa9d6b1c01963349575b45bbbfc69dbd6dd97dae052e093" Dec 06 03:59:10 crc kubenswrapper[4802]: I1206 03:59:10.686815 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-twnc4"] Dec 06 03:59:10 crc kubenswrapper[4802]: W1206 03:59:10.700410 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf33c46c6_6721_4c7a_a602_48351824838a.slice/crio-810aabdb5375ba82926e111c0fc51eb3841821b137bba5d436272c2f9998e254 WatchSource:0}: Error finding container 810aabdb5375ba82926e111c0fc51eb3841821b137bba5d436272c2f9998e254: Status 404 returned error can't find the container with id 810aabdb5375ba82926e111c0fc51eb3841821b137bba5d436272c2f9998e254 Dec 06 03:59:10 crc kubenswrapper[4802]: I1206 03:59:10.716145 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 06 03:59:10 crc kubenswrapper[4802]: I1206 03:59:10.819953 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 03:59:10 crc kubenswrapper[4802]: W1206 03:59:10.832275 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddd66ded5_a0ce_42a0_86a6_03a727128fb9.slice/crio-32cbbe3a0e2261985649e89df717505ed5f88ac73df0888b2d2e96c71507fa9c WatchSource:0}: Error finding container 32cbbe3a0e2261985649e89df717505ed5f88ac73df0888b2d2e96c71507fa9c: Status 404 returned error can't find the container with id 32cbbe3a0e2261985649e89df717505ed5f88ac73df0888b2d2e96c71507fa9c Dec 06 03:59:11 crc kubenswrapper[4802]: I1206 03:59:11.411019 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4d4f4fff-a6ec-4fef-82b4-6ba392fd297c","Type":"ContainerStarted","Data":"fa037e55af9d2c4256d307c1f7e958d5e921062fe281b0182ef8ec24b83b1e30"} Dec 06 03:59:11 crc kubenswrapper[4802]: I1206 03:59:11.415735 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-qtg4k" event={"ID":"5c636931-d435-47ec-b770-80dc1ee60756","Type":"ContainerStarted","Data":"860845ba16dd5a1de8338b6605815539607b7f5f02448fa1f96a5b3f02ee9a9e"} Dec 06 03:59:11 crc kubenswrapper[4802]: I1206 03:59:11.423712 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-dxvz6" event={"ID":"4abd9ed3-07e6-46a6-8539-85fc5c624145","Type":"ContainerStarted","Data":"9d2031d53e371d2056994227b30d5a30e7440a814f6f693c65b33b77e8a0aefe"} Dec 06 03:59:11 crc kubenswrapper[4802]: I1206 03:59:11.425984 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-twnc4" event={"ID":"f33c46c6-6721-4c7a-a602-48351824838a","Type":"ContainerStarted","Data":"3d3dd772d0f77aac781b02d536ba30f54b9ce0a6d82391bd065c0446c975682b"} Dec 06 03:59:11 crc kubenswrapper[4802]: I1206 03:59:11.426007 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-twnc4" event={"ID":"f33c46c6-6721-4c7a-a602-48351824838a","Type":"ContainerStarted","Data":"810aabdb5375ba82926e111c0fc51eb3841821b137bba5d436272c2f9998e254"} Dec 06 03:59:11 crc kubenswrapper[4802]: I1206 03:59:11.428234 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"63be777b-838a-499d-a04c-c522ae8fcb2f","Type":"ContainerStarted","Data":"5bcabe61d91a57fa2987503a17088816bc046571632f430727623baf04f9d776"} Dec 06 03:59:11 crc kubenswrapper[4802]: I1206 03:59:11.428346 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="63be777b-838a-499d-a04c-c522ae8fcb2f" containerName="glance-log" containerID="cri-o://19f6f568835b95f0c21c6e9c9ae4da59f444ec6a8fb7f93955d67e6748442c04" gracePeriod=30 Dec 06 03:59:11 crc kubenswrapper[4802]: I1206 03:59:11.428672 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="63be777b-838a-499d-a04c-c522ae8fcb2f" containerName="glance-httpd" containerID="cri-o://5bcabe61d91a57fa2987503a17088816bc046571632f430727623baf04f9d776" gracePeriod=30 Dec 06 03:59:11 crc kubenswrapper[4802]: I1206 03:59:11.434788 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"dd66ded5-a0ce-42a0-86a6-03a727128fb9","Type":"ContainerStarted","Data":"859a4d9ebe14f01da56fed5fba4fec3b96524d4d4188c5dcd62f5b895a5cca25"} Dec 06 03:59:11 crc kubenswrapper[4802]: I1206 03:59:11.434840 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"dd66ded5-a0ce-42a0-86a6-03a727128fb9","Type":"ContainerStarted","Data":"32cbbe3a0e2261985649e89df717505ed5f88ac73df0888b2d2e96c71507fa9c"} Dec 06 03:59:11 crc kubenswrapper[4802]: I1206 03:59:11.441505 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-qtg4k" podStartSLOduration=3.511306213 podStartE2EDuration="33.441486385s" podCreationTimestamp="2025-12-06 03:58:38 +0000 UTC" firstStartedPulling="2025-12-06 03:58:40.208960503 +0000 UTC m=+1113.080869665" lastFinishedPulling="2025-12-06 03:59:10.139140675 +0000 UTC m=+1143.011049837" observedRunningTime="2025-12-06 03:59:11.440783226 +0000 UTC m=+1144.312692388" watchObservedRunningTime="2025-12-06 03:59:11.441486385 +0000 UTC m=+1144.313395537" Dec 06 03:59:11 crc kubenswrapper[4802]: I1206 03:59:11.467495 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-dxvz6" podStartSLOduration=3.638016832 podStartE2EDuration="33.467477403s" podCreationTimestamp="2025-12-06 03:58:38 +0000 UTC" firstStartedPulling="2025-12-06 03:58:40.280572381 +0000 UTC m=+1113.152481533" lastFinishedPulling="2025-12-06 03:59:10.110032952 +0000 UTC m=+1142.981942104" observedRunningTime="2025-12-06 03:59:11.45892939 +0000 UTC m=+1144.330838552" watchObservedRunningTime="2025-12-06 03:59:11.467477403 +0000 UTC m=+1144.339386555" Dec 06 03:59:11 crc kubenswrapper[4802]: I1206 03:59:11.489800 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=29.48978592 podStartE2EDuration="29.48978592s" podCreationTimestamp="2025-12-06 03:58:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:59:11.483979202 +0000 UTC m=+1144.355888364" watchObservedRunningTime="2025-12-06 03:59:11.48978592 +0000 UTC m=+1144.361695072" Dec 06 03:59:11 crc kubenswrapper[4802]: I1206 03:59:11.507741 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-twnc4" podStartSLOduration=11.507720548 podStartE2EDuration="11.507720548s" podCreationTimestamp="2025-12-06 03:59:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:59:11.503911904 +0000 UTC m=+1144.375821076" watchObservedRunningTime="2025-12-06 03:59:11.507720548 +0000 UTC m=+1144.379629700" Dec 06 03:59:11 crc kubenswrapper[4802]: I1206 03:59:11.518965 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5cf08e89-5a37-4f63-8067-e8c161437deb" path="/var/lib/kubelet/pods/5cf08e89-5a37-4f63-8067-e8c161437deb/volumes" Dec 06 03:59:11 crc kubenswrapper[4802]: I1206 03:59:11.983648 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.127381 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/63be777b-838a-499d-a04c-c522ae8fcb2f-logs\") pod \"63be777b-838a-499d-a04c-c522ae8fcb2f\" (UID: \"63be777b-838a-499d-a04c-c522ae8fcb2f\") " Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.127451 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/63be777b-838a-499d-a04c-c522ae8fcb2f-httpd-run\") pod \"63be777b-838a-499d-a04c-c522ae8fcb2f\" (UID: \"63be777b-838a-499d-a04c-c522ae8fcb2f\") " Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.127506 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63be777b-838a-499d-a04c-c522ae8fcb2f-config-data\") pod \"63be777b-838a-499d-a04c-c522ae8fcb2f\" (UID: \"63be777b-838a-499d-a04c-c522ae8fcb2f\") " Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.128060 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63be777b-838a-499d-a04c-c522ae8fcb2f-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "63be777b-838a-499d-a04c-c522ae8fcb2f" (UID: "63be777b-838a-499d-a04c-c522ae8fcb2f"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.127532 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63be777b-838a-499d-a04c-c522ae8fcb2f-combined-ca-bundle\") pod \"63be777b-838a-499d-a04c-c522ae8fcb2f\" (UID: \"63be777b-838a-499d-a04c-c522ae8fcb2f\") " Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.128123 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63be777b-838a-499d-a04c-c522ae8fcb2f-logs" (OuterVolumeSpecName: "logs") pod "63be777b-838a-499d-a04c-c522ae8fcb2f" (UID: "63be777b-838a-499d-a04c-c522ae8fcb2f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.128184 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"63be777b-838a-499d-a04c-c522ae8fcb2f\" (UID: \"63be777b-838a-499d-a04c-c522ae8fcb2f\") " Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.128304 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lqt6l\" (UniqueName: \"kubernetes.io/projected/63be777b-838a-499d-a04c-c522ae8fcb2f-kube-api-access-lqt6l\") pod \"63be777b-838a-499d-a04c-c522ae8fcb2f\" (UID: \"63be777b-838a-499d-a04c-c522ae8fcb2f\") " Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.128388 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63be777b-838a-499d-a04c-c522ae8fcb2f-scripts\") pod \"63be777b-838a-499d-a04c-c522ae8fcb2f\" (UID: \"63be777b-838a-499d-a04c-c522ae8fcb2f\") " Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.128970 4802 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/63be777b-838a-499d-a04c-c522ae8fcb2f-logs\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.128991 4802 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/63be777b-838a-499d-a04c-c522ae8fcb2f-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.133985 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63be777b-838a-499d-a04c-c522ae8fcb2f-scripts" (OuterVolumeSpecName: "scripts") pod "63be777b-838a-499d-a04c-c522ae8fcb2f" (UID: "63be777b-838a-499d-a04c-c522ae8fcb2f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.134994 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63be777b-838a-499d-a04c-c522ae8fcb2f-kube-api-access-lqt6l" (OuterVolumeSpecName: "kube-api-access-lqt6l") pod "63be777b-838a-499d-a04c-c522ae8fcb2f" (UID: "63be777b-838a-499d-a04c-c522ae8fcb2f"). InnerVolumeSpecName "kube-api-access-lqt6l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.136835 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "63be777b-838a-499d-a04c-c522ae8fcb2f" (UID: "63be777b-838a-499d-a04c-c522ae8fcb2f"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.153798 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63be777b-838a-499d-a04c-c522ae8fcb2f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "63be777b-838a-499d-a04c-c522ae8fcb2f" (UID: "63be777b-838a-499d-a04c-c522ae8fcb2f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.210670 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63be777b-838a-499d-a04c-c522ae8fcb2f-config-data" (OuterVolumeSpecName: "config-data") pod "63be777b-838a-499d-a04c-c522ae8fcb2f" (UID: "63be777b-838a-499d-a04c-c522ae8fcb2f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.230161 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63be777b-838a-499d-a04c-c522ae8fcb2f-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.230194 4802 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63be777b-838a-499d-a04c-c522ae8fcb2f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.230233 4802 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.230246 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lqt6l\" (UniqueName: \"kubernetes.io/projected/63be777b-838a-499d-a04c-c522ae8fcb2f-kube-api-access-lqt6l\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.230257 4802 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63be777b-838a-499d-a04c-c522ae8fcb2f-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.252582 4802 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.331657 4802 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.447674 4802 generic.go:334] "Generic (PLEG): container finished" podID="63be777b-838a-499d-a04c-c522ae8fcb2f" containerID="5bcabe61d91a57fa2987503a17088816bc046571632f430727623baf04f9d776" exitCode=0 Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.447705 4802 generic.go:334] "Generic (PLEG): container finished" podID="63be777b-838a-499d-a04c-c522ae8fcb2f" containerID="19f6f568835b95f0c21c6e9c9ae4da59f444ec6a8fb7f93955d67e6748442c04" exitCode=143 Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.447820 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"63be777b-838a-499d-a04c-c522ae8fcb2f","Type":"ContainerDied","Data":"5bcabe61d91a57fa2987503a17088816bc046571632f430727623baf04f9d776"} Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.447847 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"63be777b-838a-499d-a04c-c522ae8fcb2f","Type":"ContainerDied","Data":"19f6f568835b95f0c21c6e9c9ae4da59f444ec6a8fb7f93955d67e6748442c04"} Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.447858 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"63be777b-838a-499d-a04c-c522ae8fcb2f","Type":"ContainerDied","Data":"73faaf4188dcb6c2b7516e1d6c3e449cfa4d3935d23f5695a286f1f60ad2ad72"} Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.447873 4802 scope.go:117] "RemoveContainer" containerID="5bcabe61d91a57fa2987503a17088816bc046571632f430727623baf04f9d776" Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.448067 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.455227 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"dd66ded5-a0ce-42a0-86a6-03a727128fb9","Type":"ContainerStarted","Data":"b96e7a241b3adc317a0b265ed0bda6ce6a17415000fa541766e249e42f817a16"} Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.455735 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="dd66ded5-a0ce-42a0-86a6-03a727128fb9" containerName="glance-log" containerID="cri-o://859a4d9ebe14f01da56fed5fba4fec3b96524d4d4188c5dcd62f5b895a5cca25" gracePeriod=30 Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.455959 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="dd66ded5-a0ce-42a0-86a6-03a727128fb9" containerName="glance-httpd" containerID="cri-o://b96e7a241b3adc317a0b265ed0bda6ce6a17415000fa541766e249e42f817a16" gracePeriod=30 Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.481863 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=24.481843524 podStartE2EDuration="24.481843524s" podCreationTimestamp="2025-12-06 03:58:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:59:12.477909138 +0000 UTC m=+1145.349818290" watchObservedRunningTime="2025-12-06 03:59:12.481843524 +0000 UTC m=+1145.353752676" Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.514492 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.521805 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.526646 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 03:59:12 crc kubenswrapper[4802]: E1206 03:59:12.526963 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cf08e89-5a37-4f63-8067-e8c161437deb" containerName="dnsmasq-dns" Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.526978 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cf08e89-5a37-4f63-8067-e8c161437deb" containerName="dnsmasq-dns" Dec 06 03:59:12 crc kubenswrapper[4802]: E1206 03:59:12.526991 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63be777b-838a-499d-a04c-c522ae8fcb2f" containerName="glance-log" Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.526998 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="63be777b-838a-499d-a04c-c522ae8fcb2f" containerName="glance-log" Dec 06 03:59:12 crc kubenswrapper[4802]: E1206 03:59:12.527012 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63be777b-838a-499d-a04c-c522ae8fcb2f" containerName="glance-httpd" Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.527018 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="63be777b-838a-499d-a04c-c522ae8fcb2f" containerName="glance-httpd" Dec 06 03:59:12 crc kubenswrapper[4802]: E1206 03:59:12.527042 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cf08e89-5a37-4f63-8067-e8c161437deb" containerName="init" Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.527048 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cf08e89-5a37-4f63-8067-e8c161437deb" containerName="init" Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.527192 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="63be777b-838a-499d-a04c-c522ae8fcb2f" containerName="glance-log" Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.527209 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="63be777b-838a-499d-a04c-c522ae8fcb2f" containerName="glance-httpd" Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.527224 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cf08e89-5a37-4f63-8067-e8c161437deb" containerName="dnsmasq-dns" Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.528255 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.534421 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.534686 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.539639 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.636228 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8c7901e-3653-4d82-a15f-dd12cf4ab8c5-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f8c7901e-3653-4d82-a15f-dd12cf4ab8c5\") " pod="openstack/glance-default-external-api-0" Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.636571 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8c7901e-3653-4d82-a15f-dd12cf4ab8c5-config-data\") pod \"glance-default-external-api-0\" (UID: \"f8c7901e-3653-4d82-a15f-dd12cf4ab8c5\") " pod="openstack/glance-default-external-api-0" Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.636622 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8c7901e-3653-4d82-a15f-dd12cf4ab8c5-scripts\") pod \"glance-default-external-api-0\" (UID: \"f8c7901e-3653-4d82-a15f-dd12cf4ab8c5\") " pod="openstack/glance-default-external-api-0" Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.636647 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbkgw\" (UniqueName: \"kubernetes.io/projected/f8c7901e-3653-4d82-a15f-dd12cf4ab8c5-kube-api-access-hbkgw\") pod \"glance-default-external-api-0\" (UID: \"f8c7901e-3653-4d82-a15f-dd12cf4ab8c5\") " pod="openstack/glance-default-external-api-0" Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.636707 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"f8c7901e-3653-4d82-a15f-dd12cf4ab8c5\") " pod="openstack/glance-default-external-api-0" Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.636824 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f8c7901e-3653-4d82-a15f-dd12cf4ab8c5-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f8c7901e-3653-4d82-a15f-dd12cf4ab8c5\") " pod="openstack/glance-default-external-api-0" Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.636967 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8c7901e-3653-4d82-a15f-dd12cf4ab8c5-logs\") pod \"glance-default-external-api-0\" (UID: \"f8c7901e-3653-4d82-a15f-dd12cf4ab8c5\") " pod="openstack/glance-default-external-api-0" Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.637909 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8c7901e-3653-4d82-a15f-dd12cf4ab8c5-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f8c7901e-3653-4d82-a15f-dd12cf4ab8c5\") " pod="openstack/glance-default-external-api-0" Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.660997 4802 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-77frt" podUID="5cf08e89-5a37-4f63-8067-e8c161437deb" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.118:5353: i/o timeout" Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.739863 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"f8c7901e-3653-4d82-a15f-dd12cf4ab8c5\") " pod="openstack/glance-default-external-api-0" Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.740281 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f8c7901e-3653-4d82-a15f-dd12cf4ab8c5-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f8c7901e-3653-4d82-a15f-dd12cf4ab8c5\") " pod="openstack/glance-default-external-api-0" Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.740234 4802 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"f8c7901e-3653-4d82-a15f-dd12cf4ab8c5\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-external-api-0" Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.740481 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8c7901e-3653-4d82-a15f-dd12cf4ab8c5-logs\") pod \"glance-default-external-api-0\" (UID: \"f8c7901e-3653-4d82-a15f-dd12cf4ab8c5\") " pod="openstack/glance-default-external-api-0" Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.740524 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8c7901e-3653-4d82-a15f-dd12cf4ab8c5-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f8c7901e-3653-4d82-a15f-dd12cf4ab8c5\") " pod="openstack/glance-default-external-api-0" Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.740574 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8c7901e-3653-4d82-a15f-dd12cf4ab8c5-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f8c7901e-3653-4d82-a15f-dd12cf4ab8c5\") " pod="openstack/glance-default-external-api-0" Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.740603 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8c7901e-3653-4d82-a15f-dd12cf4ab8c5-config-data\") pod \"glance-default-external-api-0\" (UID: \"f8c7901e-3653-4d82-a15f-dd12cf4ab8c5\") " pod="openstack/glance-default-external-api-0" Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.740630 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8c7901e-3653-4d82-a15f-dd12cf4ab8c5-scripts\") pod \"glance-default-external-api-0\" (UID: \"f8c7901e-3653-4d82-a15f-dd12cf4ab8c5\") " pod="openstack/glance-default-external-api-0" Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.740647 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbkgw\" (UniqueName: \"kubernetes.io/projected/f8c7901e-3653-4d82-a15f-dd12cf4ab8c5-kube-api-access-hbkgw\") pod \"glance-default-external-api-0\" (UID: \"f8c7901e-3653-4d82-a15f-dd12cf4ab8c5\") " pod="openstack/glance-default-external-api-0" Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.740966 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f8c7901e-3653-4d82-a15f-dd12cf4ab8c5-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f8c7901e-3653-4d82-a15f-dd12cf4ab8c5\") " pod="openstack/glance-default-external-api-0" Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.741005 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8c7901e-3653-4d82-a15f-dd12cf4ab8c5-logs\") pod \"glance-default-external-api-0\" (UID: \"f8c7901e-3653-4d82-a15f-dd12cf4ab8c5\") " pod="openstack/glance-default-external-api-0" Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.746194 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8c7901e-3653-4d82-a15f-dd12cf4ab8c5-config-data\") pod \"glance-default-external-api-0\" (UID: \"f8c7901e-3653-4d82-a15f-dd12cf4ab8c5\") " pod="openstack/glance-default-external-api-0" Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.766525 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8c7901e-3653-4d82-a15f-dd12cf4ab8c5-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f8c7901e-3653-4d82-a15f-dd12cf4ab8c5\") " pod="openstack/glance-default-external-api-0" Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.766643 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8c7901e-3653-4d82-a15f-dd12cf4ab8c5-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f8c7901e-3653-4d82-a15f-dd12cf4ab8c5\") " pod="openstack/glance-default-external-api-0" Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.770655 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8c7901e-3653-4d82-a15f-dd12cf4ab8c5-scripts\") pod \"glance-default-external-api-0\" (UID: \"f8c7901e-3653-4d82-a15f-dd12cf4ab8c5\") " pod="openstack/glance-default-external-api-0" Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.770965 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbkgw\" (UniqueName: \"kubernetes.io/projected/f8c7901e-3653-4d82-a15f-dd12cf4ab8c5-kube-api-access-hbkgw\") pod \"glance-default-external-api-0\" (UID: \"f8c7901e-3653-4d82-a15f-dd12cf4ab8c5\") " pod="openstack/glance-default-external-api-0" Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.777458 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"f8c7901e-3653-4d82-a15f-dd12cf4ab8c5\") " pod="openstack/glance-default-external-api-0" Dec 06 03:59:12 crc kubenswrapper[4802]: I1206 03:59:12.857607 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 03:59:13 crc kubenswrapper[4802]: I1206 03:59:13.283720 4802 patch_prober.go:28] interesting pod/machine-config-daemon-zpxxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 03:59:13 crc kubenswrapper[4802]: I1206 03:59:13.283796 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 03:59:13 crc kubenswrapper[4802]: I1206 03:59:13.471784 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63be777b-838a-499d-a04c-c522ae8fcb2f" path="/var/lib/kubelet/pods/63be777b-838a-499d-a04c-c522ae8fcb2f/volumes" Dec 06 03:59:13 crc kubenswrapper[4802]: I1206 03:59:13.479344 4802 generic.go:334] "Generic (PLEG): container finished" podID="dd66ded5-a0ce-42a0-86a6-03a727128fb9" containerID="b96e7a241b3adc317a0b265ed0bda6ce6a17415000fa541766e249e42f817a16" exitCode=0 Dec 06 03:59:13 crc kubenswrapper[4802]: I1206 03:59:13.479378 4802 generic.go:334] "Generic (PLEG): container finished" podID="dd66ded5-a0ce-42a0-86a6-03a727128fb9" containerID="859a4d9ebe14f01da56fed5fba4fec3b96524d4d4188c5dcd62f5b895a5cca25" exitCode=143 Dec 06 03:59:13 crc kubenswrapper[4802]: I1206 03:59:13.479394 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"dd66ded5-a0ce-42a0-86a6-03a727128fb9","Type":"ContainerDied","Data":"b96e7a241b3adc317a0b265ed0bda6ce6a17415000fa541766e249e42f817a16"} Dec 06 03:59:13 crc kubenswrapper[4802]: I1206 03:59:13.479417 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"dd66ded5-a0ce-42a0-86a6-03a727128fb9","Type":"ContainerDied","Data":"859a4d9ebe14f01da56fed5fba4fec3b96524d4d4188c5dcd62f5b895a5cca25"} Dec 06 03:59:14 crc kubenswrapper[4802]: I1206 03:59:14.260099 4802 scope.go:117] "RemoveContainer" containerID="19f6f568835b95f0c21c6e9c9ae4da59f444ec6a8fb7f93955d67e6748442c04" Dec 06 03:59:14 crc kubenswrapper[4802]: I1206 03:59:14.389017 4802 scope.go:117] "RemoveContainer" containerID="5bcabe61d91a57fa2987503a17088816bc046571632f430727623baf04f9d776" Dec 06 03:59:14 crc kubenswrapper[4802]: E1206 03:59:14.389660 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5bcabe61d91a57fa2987503a17088816bc046571632f430727623baf04f9d776\": container with ID starting with 5bcabe61d91a57fa2987503a17088816bc046571632f430727623baf04f9d776 not found: ID does not exist" containerID="5bcabe61d91a57fa2987503a17088816bc046571632f430727623baf04f9d776" Dec 06 03:59:14 crc kubenswrapper[4802]: I1206 03:59:14.389700 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5bcabe61d91a57fa2987503a17088816bc046571632f430727623baf04f9d776"} err="failed to get container status \"5bcabe61d91a57fa2987503a17088816bc046571632f430727623baf04f9d776\": rpc error: code = NotFound desc = could not find container \"5bcabe61d91a57fa2987503a17088816bc046571632f430727623baf04f9d776\": container with ID starting with 5bcabe61d91a57fa2987503a17088816bc046571632f430727623baf04f9d776 not found: ID does not exist" Dec 06 03:59:14 crc kubenswrapper[4802]: I1206 03:59:14.389723 4802 scope.go:117] "RemoveContainer" containerID="19f6f568835b95f0c21c6e9c9ae4da59f444ec6a8fb7f93955d67e6748442c04" Dec 06 03:59:14 crc kubenswrapper[4802]: E1206 03:59:14.390200 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19f6f568835b95f0c21c6e9c9ae4da59f444ec6a8fb7f93955d67e6748442c04\": container with ID starting with 19f6f568835b95f0c21c6e9c9ae4da59f444ec6a8fb7f93955d67e6748442c04 not found: ID does not exist" containerID="19f6f568835b95f0c21c6e9c9ae4da59f444ec6a8fb7f93955d67e6748442c04" Dec 06 03:59:14 crc kubenswrapper[4802]: I1206 03:59:14.390250 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19f6f568835b95f0c21c6e9c9ae4da59f444ec6a8fb7f93955d67e6748442c04"} err="failed to get container status \"19f6f568835b95f0c21c6e9c9ae4da59f444ec6a8fb7f93955d67e6748442c04\": rpc error: code = NotFound desc = could not find container \"19f6f568835b95f0c21c6e9c9ae4da59f444ec6a8fb7f93955d67e6748442c04\": container with ID starting with 19f6f568835b95f0c21c6e9c9ae4da59f444ec6a8fb7f93955d67e6748442c04 not found: ID does not exist" Dec 06 03:59:14 crc kubenswrapper[4802]: I1206 03:59:14.390265 4802 scope.go:117] "RemoveContainer" containerID="5bcabe61d91a57fa2987503a17088816bc046571632f430727623baf04f9d776" Dec 06 03:59:14 crc kubenswrapper[4802]: I1206 03:59:14.390526 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5bcabe61d91a57fa2987503a17088816bc046571632f430727623baf04f9d776"} err="failed to get container status \"5bcabe61d91a57fa2987503a17088816bc046571632f430727623baf04f9d776\": rpc error: code = NotFound desc = could not find container \"5bcabe61d91a57fa2987503a17088816bc046571632f430727623baf04f9d776\": container with ID starting with 5bcabe61d91a57fa2987503a17088816bc046571632f430727623baf04f9d776 not found: ID does not exist" Dec 06 03:59:14 crc kubenswrapper[4802]: I1206 03:59:14.390581 4802 scope.go:117] "RemoveContainer" containerID="19f6f568835b95f0c21c6e9c9ae4da59f444ec6a8fb7f93955d67e6748442c04" Dec 06 03:59:14 crc kubenswrapper[4802]: I1206 03:59:14.390888 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19f6f568835b95f0c21c6e9c9ae4da59f444ec6a8fb7f93955d67e6748442c04"} err="failed to get container status \"19f6f568835b95f0c21c6e9c9ae4da59f444ec6a8fb7f93955d67e6748442c04\": rpc error: code = NotFound desc = could not find container \"19f6f568835b95f0c21c6e9c9ae4da59f444ec6a8fb7f93955d67e6748442c04\": container with ID starting with 19f6f568835b95f0c21c6e9c9ae4da59f444ec6a8fb7f93955d67e6748442c04 not found: ID does not exist" Dec 06 03:59:15 crc kubenswrapper[4802]: W1206 03:59:15.187543 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf8c7901e_3653_4d82_a15f_dd12cf4ab8c5.slice/crio-92c412ca7d508b88eea090e0fed09145e674f09e003ef82d0e103b23d4b190b6 WatchSource:0}: Error finding container 92c412ca7d508b88eea090e0fed09145e674f09e003ef82d0e103b23d4b190b6: Status 404 returned error can't find the container with id 92c412ca7d508b88eea090e0fed09145e674f09e003ef82d0e103b23d4b190b6 Dec 06 03:59:15 crc kubenswrapper[4802]: I1206 03:59:15.208910 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 03:59:15 crc kubenswrapper[4802]: I1206 03:59:15.557716 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f8c7901e-3653-4d82-a15f-dd12cf4ab8c5","Type":"ContainerStarted","Data":"92c412ca7d508b88eea090e0fed09145e674f09e003ef82d0e103b23d4b190b6"} Dec 06 03:59:16 crc kubenswrapper[4802]: I1206 03:59:16.569107 4802 generic.go:334] "Generic (PLEG): container finished" podID="f33c46c6-6721-4c7a-a602-48351824838a" containerID="3d3dd772d0f77aac781b02d536ba30f54b9ce0a6d82391bd065c0446c975682b" exitCode=0 Dec 06 03:59:16 crc kubenswrapper[4802]: I1206 03:59:16.569193 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-twnc4" event={"ID":"f33c46c6-6721-4c7a-a602-48351824838a","Type":"ContainerDied","Data":"3d3dd772d0f77aac781b02d536ba30f54b9ce0a6d82391bd065c0446c975682b"} Dec 06 03:59:16 crc kubenswrapper[4802]: I1206 03:59:16.571984 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f8c7901e-3653-4d82-a15f-dd12cf4ab8c5","Type":"ContainerStarted","Data":"bd7d4fa9d3441f11c0fea392f92f249b999c25be443db3a969902839362960b2"} Dec 06 03:59:17 crc kubenswrapper[4802]: I1206 03:59:17.581554 4802 generic.go:334] "Generic (PLEG): container finished" podID="4abd9ed3-07e6-46a6-8539-85fc5c624145" containerID="9d2031d53e371d2056994227b30d5a30e7440a814f6f693c65b33b77e8a0aefe" exitCode=0 Dec 06 03:59:17 crc kubenswrapper[4802]: I1206 03:59:17.581649 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-dxvz6" event={"ID":"4abd9ed3-07e6-46a6-8539-85fc5c624145","Type":"ContainerDied","Data":"9d2031d53e371d2056994227b30d5a30e7440a814f6f693c65b33b77e8a0aefe"} Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.243106 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.249362 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-twnc4" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.340143 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f33c46c6-6721-4c7a-a602-48351824838a-fernet-keys\") pod \"f33c46c6-6721-4c7a-a602-48351824838a\" (UID: \"f33c46c6-6721-4c7a-a602-48351824838a\") " Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.340206 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dd66ded5-a0ce-42a0-86a6-03a727128fb9-scripts\") pod \"dd66ded5-a0ce-42a0-86a6-03a727128fb9\" (UID: \"dd66ded5-a0ce-42a0-86a6-03a727128fb9\") " Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.340226 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f33c46c6-6721-4c7a-a602-48351824838a-config-data\") pod \"f33c46c6-6721-4c7a-a602-48351824838a\" (UID: \"f33c46c6-6721-4c7a-a602-48351824838a\") " Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.340253 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/dd66ded5-a0ce-42a0-86a6-03a727128fb9-httpd-run\") pod \"dd66ded5-a0ce-42a0-86a6-03a727128fb9\" (UID: \"dd66ded5-a0ce-42a0-86a6-03a727128fb9\") " Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.340316 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f33c46c6-6721-4c7a-a602-48351824838a-combined-ca-bundle\") pod \"f33c46c6-6721-4c7a-a602-48351824838a\" (UID: \"f33c46c6-6721-4c7a-a602-48351824838a\") " Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.340343 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd66ded5-a0ce-42a0-86a6-03a727128fb9-logs\") pod \"dd66ded5-a0ce-42a0-86a6-03a727128fb9\" (UID: \"dd66ded5-a0ce-42a0-86a6-03a727128fb9\") " Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.340916 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f33c46c6-6721-4c7a-a602-48351824838a-scripts\") pod \"f33c46c6-6721-4c7a-a602-48351824838a\" (UID: \"f33c46c6-6721-4c7a-a602-48351824838a\") " Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.340974 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ntxvn\" (UniqueName: \"kubernetes.io/projected/f33c46c6-6721-4c7a-a602-48351824838a-kube-api-access-ntxvn\") pod \"f33c46c6-6721-4c7a-a602-48351824838a\" (UID: \"f33c46c6-6721-4c7a-a602-48351824838a\") " Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.340996 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd66ded5-a0ce-42a0-86a6-03a727128fb9-combined-ca-bundle\") pod \"dd66ded5-a0ce-42a0-86a6-03a727128fb9\" (UID: \"dd66ded5-a0ce-42a0-86a6-03a727128fb9\") " Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.341052 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"dd66ded5-a0ce-42a0-86a6-03a727128fb9\" (UID: \"dd66ded5-a0ce-42a0-86a6-03a727128fb9\") " Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.341093 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f33c46c6-6721-4c7a-a602-48351824838a-credential-keys\") pod \"f33c46c6-6721-4c7a-a602-48351824838a\" (UID: \"f33c46c6-6721-4c7a-a602-48351824838a\") " Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.341125 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd66ded5-a0ce-42a0-86a6-03a727128fb9-config-data\") pod \"dd66ded5-a0ce-42a0-86a6-03a727128fb9\" (UID: \"dd66ded5-a0ce-42a0-86a6-03a727128fb9\") " Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.341170 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sczlv\" (UniqueName: \"kubernetes.io/projected/dd66ded5-a0ce-42a0-86a6-03a727128fb9-kube-api-access-sczlv\") pod \"dd66ded5-a0ce-42a0-86a6-03a727128fb9\" (UID: \"dd66ded5-a0ce-42a0-86a6-03a727128fb9\") " Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.341840 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd66ded5-a0ce-42a0-86a6-03a727128fb9-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "dd66ded5-a0ce-42a0-86a6-03a727128fb9" (UID: "dd66ded5-a0ce-42a0-86a6-03a727128fb9"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.343817 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f33c46c6-6721-4c7a-a602-48351824838a-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "f33c46c6-6721-4c7a-a602-48351824838a" (UID: "f33c46c6-6721-4c7a-a602-48351824838a"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.346702 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd66ded5-a0ce-42a0-86a6-03a727128fb9-scripts" (OuterVolumeSpecName: "scripts") pod "dd66ded5-a0ce-42a0-86a6-03a727128fb9" (UID: "dd66ded5-a0ce-42a0-86a6-03a727128fb9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.347395 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f33c46c6-6721-4c7a-a602-48351824838a-scripts" (OuterVolumeSpecName: "scripts") pod "f33c46c6-6721-4c7a-a602-48351824838a" (UID: "f33c46c6-6721-4c7a-a602-48351824838a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.347593 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd66ded5-a0ce-42a0-86a6-03a727128fb9-logs" (OuterVolumeSpecName: "logs") pod "dd66ded5-a0ce-42a0-86a6-03a727128fb9" (UID: "dd66ded5-a0ce-42a0-86a6-03a727128fb9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.347742 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f33c46c6-6721-4c7a-a602-48351824838a-kube-api-access-ntxvn" (OuterVolumeSpecName: "kube-api-access-ntxvn") pod "f33c46c6-6721-4c7a-a602-48351824838a" (UID: "f33c46c6-6721-4c7a-a602-48351824838a"). InnerVolumeSpecName "kube-api-access-ntxvn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.348942 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f33c46c6-6721-4c7a-a602-48351824838a-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "f33c46c6-6721-4c7a-a602-48351824838a" (UID: "f33c46c6-6721-4c7a-a602-48351824838a"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.351202 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd66ded5-a0ce-42a0-86a6-03a727128fb9-kube-api-access-sczlv" (OuterVolumeSpecName: "kube-api-access-sczlv") pod "dd66ded5-a0ce-42a0-86a6-03a727128fb9" (UID: "dd66ded5-a0ce-42a0-86a6-03a727128fb9"). InnerVolumeSpecName "kube-api-access-sczlv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.363081 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "dd66ded5-a0ce-42a0-86a6-03a727128fb9" (UID: "dd66ded5-a0ce-42a0-86a6-03a727128fb9"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.371959 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd66ded5-a0ce-42a0-86a6-03a727128fb9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dd66ded5-a0ce-42a0-86a6-03a727128fb9" (UID: "dd66ded5-a0ce-42a0-86a6-03a727128fb9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.372217 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f33c46c6-6721-4c7a-a602-48351824838a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f33c46c6-6721-4c7a-a602-48351824838a" (UID: "f33c46c6-6721-4c7a-a602-48351824838a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.381428 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f33c46c6-6721-4c7a-a602-48351824838a-config-data" (OuterVolumeSpecName: "config-data") pod "f33c46c6-6721-4c7a-a602-48351824838a" (UID: "f33c46c6-6721-4c7a-a602-48351824838a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.399021 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd66ded5-a0ce-42a0-86a6-03a727128fb9-config-data" (OuterVolumeSpecName: "config-data") pod "dd66ded5-a0ce-42a0-86a6-03a727128fb9" (UID: "dd66ded5-a0ce-42a0-86a6-03a727128fb9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.443473 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sczlv\" (UniqueName: \"kubernetes.io/projected/dd66ded5-a0ce-42a0-86a6-03a727128fb9-kube-api-access-sczlv\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.443508 4802 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f33c46c6-6721-4c7a-a602-48351824838a-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.443521 4802 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dd66ded5-a0ce-42a0-86a6-03a727128fb9-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.443530 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f33c46c6-6721-4c7a-a602-48351824838a-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.443539 4802 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/dd66ded5-a0ce-42a0-86a6-03a727128fb9-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.443548 4802 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f33c46c6-6721-4c7a-a602-48351824838a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.443556 4802 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd66ded5-a0ce-42a0-86a6-03a727128fb9-logs\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.443564 4802 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f33c46c6-6721-4c7a-a602-48351824838a-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.443571 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ntxvn\" (UniqueName: \"kubernetes.io/projected/f33c46c6-6721-4c7a-a602-48351824838a-kube-api-access-ntxvn\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.443579 4802 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd66ded5-a0ce-42a0-86a6-03a727128fb9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.443609 4802 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.443618 4802 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f33c46c6-6721-4c7a-a602-48351824838a-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.443625 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd66ded5-a0ce-42a0-86a6-03a727128fb9-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.465310 4802 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.545941 4802 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.594515 4802 generic.go:334] "Generic (PLEG): container finished" podID="5c636931-d435-47ec-b770-80dc1ee60756" containerID="860845ba16dd5a1de8338b6605815539607b7f5f02448fa1f96a5b3f02ee9a9e" exitCode=0 Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.594883 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-qtg4k" event={"ID":"5c636931-d435-47ec-b770-80dc1ee60756","Type":"ContainerDied","Data":"860845ba16dd5a1de8338b6605815539607b7f5f02448fa1f96a5b3f02ee9a9e"} Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.597119 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-twnc4" event={"ID":"f33c46c6-6721-4c7a-a602-48351824838a","Type":"ContainerDied","Data":"810aabdb5375ba82926e111c0fc51eb3841821b137bba5d436272c2f9998e254"} Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.597165 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="810aabdb5375ba82926e111c0fc51eb3841821b137bba5d436272c2f9998e254" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.597920 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-twnc4" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.600699 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"dd66ded5-a0ce-42a0-86a6-03a727128fb9","Type":"ContainerDied","Data":"32cbbe3a0e2261985649e89df717505ed5f88ac73df0888b2d2e96c71507fa9c"} Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.600827 4802 scope.go:117] "RemoveContainer" containerID="b96e7a241b3adc317a0b265ed0bda6ce6a17415000fa541766e249e42f817a16" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.601025 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.603738 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4d4f4fff-a6ec-4fef-82b4-6ba392fd297c","Type":"ContainerStarted","Data":"fff16873781a4a16ca4ed5fa3a576de4e59fcca96bb97e804e8f0e742df14984"} Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.647572 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.653344 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.654291 4802 scope.go:117] "RemoveContainer" containerID="859a4d9ebe14f01da56fed5fba4fec3b96524d4d4188c5dcd62f5b895a5cca25" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.675399 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 03:59:18 crc kubenswrapper[4802]: E1206 03:59:18.675781 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd66ded5-a0ce-42a0-86a6-03a727128fb9" containerName="glance-httpd" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.675797 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd66ded5-a0ce-42a0-86a6-03a727128fb9" containerName="glance-httpd" Dec 06 03:59:18 crc kubenswrapper[4802]: E1206 03:59:18.675817 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd66ded5-a0ce-42a0-86a6-03a727128fb9" containerName="glance-log" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.675823 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd66ded5-a0ce-42a0-86a6-03a727128fb9" containerName="glance-log" Dec 06 03:59:18 crc kubenswrapper[4802]: E1206 03:59:18.675843 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f33c46c6-6721-4c7a-a602-48351824838a" containerName="keystone-bootstrap" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.675850 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="f33c46c6-6721-4c7a-a602-48351824838a" containerName="keystone-bootstrap" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.676085 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd66ded5-a0ce-42a0-86a6-03a727128fb9" containerName="glance-log" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.676109 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd66ded5-a0ce-42a0-86a6-03a727128fb9" containerName="glance-httpd" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.676121 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="f33c46c6-6721-4c7a-a602-48351824838a" containerName="keystone-bootstrap" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.677052 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.686607 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.686878 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.701628 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.722081 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-75dc574896-4vrbs"] Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.723101 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-75dc574896-4vrbs" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.727639 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.727951 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.728079 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.728177 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.728318 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.728482 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-zk4kp" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.750576 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-75dc574896-4vrbs"] Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.751432 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/22294441-50c6-4dc0-80f1-4b14c1fc62a9-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"22294441-50c6-4dc0-80f1-4b14c1fc62a9\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.751493 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22294441-50c6-4dc0-80f1-4b14c1fc62a9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"22294441-50c6-4dc0-80f1-4b14c1fc62a9\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.751510 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/22294441-50c6-4dc0-80f1-4b14c1fc62a9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"22294441-50c6-4dc0-80f1-4b14c1fc62a9\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.751550 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9csvl\" (UniqueName: \"kubernetes.io/projected/22294441-50c6-4dc0-80f1-4b14c1fc62a9-kube-api-access-9csvl\") pod \"glance-default-internal-api-0\" (UID: \"22294441-50c6-4dc0-80f1-4b14c1fc62a9\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.751567 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22294441-50c6-4dc0-80f1-4b14c1fc62a9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"22294441-50c6-4dc0-80f1-4b14c1fc62a9\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.751584 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/22294441-50c6-4dc0-80f1-4b14c1fc62a9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"22294441-50c6-4dc0-80f1-4b14c1fc62a9\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.751632 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"22294441-50c6-4dc0-80f1-4b14c1fc62a9\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.751656 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22294441-50c6-4dc0-80f1-4b14c1fc62a9-logs\") pod \"glance-default-internal-api-0\" (UID: \"22294441-50c6-4dc0-80f1-4b14c1fc62a9\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.852944 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bdd2c50b-481a-4ab7-90ce-86cb69efd255-scripts\") pod \"keystone-75dc574896-4vrbs\" (UID: \"bdd2c50b-481a-4ab7-90ce-86cb69efd255\") " pod="openstack/keystone-75dc574896-4vrbs" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.853007 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22294441-50c6-4dc0-80f1-4b14c1fc62a9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"22294441-50c6-4dc0-80f1-4b14c1fc62a9\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.853027 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/22294441-50c6-4dc0-80f1-4b14c1fc62a9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"22294441-50c6-4dc0-80f1-4b14c1fc62a9\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.853057 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdd2c50b-481a-4ab7-90ce-86cb69efd255-combined-ca-bundle\") pod \"keystone-75dc574896-4vrbs\" (UID: \"bdd2c50b-481a-4ab7-90ce-86cb69efd255\") " pod="openstack/keystone-75dc574896-4vrbs" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.853097 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9csvl\" (UniqueName: \"kubernetes.io/projected/22294441-50c6-4dc0-80f1-4b14c1fc62a9-kube-api-access-9csvl\") pod \"glance-default-internal-api-0\" (UID: \"22294441-50c6-4dc0-80f1-4b14c1fc62a9\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.853114 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22294441-50c6-4dc0-80f1-4b14c1fc62a9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"22294441-50c6-4dc0-80f1-4b14c1fc62a9\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.853132 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/22294441-50c6-4dc0-80f1-4b14c1fc62a9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"22294441-50c6-4dc0-80f1-4b14c1fc62a9\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.853155 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdd2c50b-481a-4ab7-90ce-86cb69efd255-config-data\") pod \"keystone-75dc574896-4vrbs\" (UID: \"bdd2c50b-481a-4ab7-90ce-86cb69efd255\") " pod="openstack/keystone-75dc574896-4vrbs" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.853192 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/bdd2c50b-481a-4ab7-90ce-86cb69efd255-credential-keys\") pod \"keystone-75dc574896-4vrbs\" (UID: \"bdd2c50b-481a-4ab7-90ce-86cb69efd255\") " pod="openstack/keystone-75dc574896-4vrbs" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.853217 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdd2c50b-481a-4ab7-90ce-86cb69efd255-internal-tls-certs\") pod \"keystone-75dc574896-4vrbs\" (UID: \"bdd2c50b-481a-4ab7-90ce-86cb69efd255\") " pod="openstack/keystone-75dc574896-4vrbs" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.853233 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bdd2c50b-481a-4ab7-90ce-86cb69efd255-fernet-keys\") pod \"keystone-75dc574896-4vrbs\" (UID: \"bdd2c50b-481a-4ab7-90ce-86cb69efd255\") " pod="openstack/keystone-75dc574896-4vrbs" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.853258 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"22294441-50c6-4dc0-80f1-4b14c1fc62a9\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.853277 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t457b\" (UniqueName: \"kubernetes.io/projected/bdd2c50b-481a-4ab7-90ce-86cb69efd255-kube-api-access-t457b\") pod \"keystone-75dc574896-4vrbs\" (UID: \"bdd2c50b-481a-4ab7-90ce-86cb69efd255\") " pod="openstack/keystone-75dc574896-4vrbs" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.853303 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22294441-50c6-4dc0-80f1-4b14c1fc62a9-logs\") pod \"glance-default-internal-api-0\" (UID: \"22294441-50c6-4dc0-80f1-4b14c1fc62a9\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.853336 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/22294441-50c6-4dc0-80f1-4b14c1fc62a9-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"22294441-50c6-4dc0-80f1-4b14c1fc62a9\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.853367 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdd2c50b-481a-4ab7-90ce-86cb69efd255-public-tls-certs\") pod \"keystone-75dc574896-4vrbs\" (UID: \"bdd2c50b-481a-4ab7-90ce-86cb69efd255\") " pod="openstack/keystone-75dc574896-4vrbs" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.854708 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22294441-50c6-4dc0-80f1-4b14c1fc62a9-logs\") pod \"glance-default-internal-api-0\" (UID: \"22294441-50c6-4dc0-80f1-4b14c1fc62a9\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.855021 4802 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"22294441-50c6-4dc0-80f1-4b14c1fc62a9\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-internal-api-0" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.855539 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/22294441-50c6-4dc0-80f1-4b14c1fc62a9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"22294441-50c6-4dc0-80f1-4b14c1fc62a9\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.858442 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22294441-50c6-4dc0-80f1-4b14c1fc62a9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"22294441-50c6-4dc0-80f1-4b14c1fc62a9\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.859011 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/22294441-50c6-4dc0-80f1-4b14c1fc62a9-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"22294441-50c6-4dc0-80f1-4b14c1fc62a9\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.860333 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/22294441-50c6-4dc0-80f1-4b14c1fc62a9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"22294441-50c6-4dc0-80f1-4b14c1fc62a9\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.864707 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22294441-50c6-4dc0-80f1-4b14c1fc62a9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"22294441-50c6-4dc0-80f1-4b14c1fc62a9\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.873398 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9csvl\" (UniqueName: \"kubernetes.io/projected/22294441-50c6-4dc0-80f1-4b14c1fc62a9-kube-api-access-9csvl\") pod \"glance-default-internal-api-0\" (UID: \"22294441-50c6-4dc0-80f1-4b14c1fc62a9\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.888620 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"22294441-50c6-4dc0-80f1-4b14c1fc62a9\") " pod="openstack/glance-default-internal-api-0" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.948732 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-dxvz6" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.954343 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/bdd2c50b-481a-4ab7-90ce-86cb69efd255-credential-keys\") pod \"keystone-75dc574896-4vrbs\" (UID: \"bdd2c50b-481a-4ab7-90ce-86cb69efd255\") " pod="openstack/keystone-75dc574896-4vrbs" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.954411 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdd2c50b-481a-4ab7-90ce-86cb69efd255-internal-tls-certs\") pod \"keystone-75dc574896-4vrbs\" (UID: \"bdd2c50b-481a-4ab7-90ce-86cb69efd255\") " pod="openstack/keystone-75dc574896-4vrbs" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.954437 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bdd2c50b-481a-4ab7-90ce-86cb69efd255-fernet-keys\") pod \"keystone-75dc574896-4vrbs\" (UID: \"bdd2c50b-481a-4ab7-90ce-86cb69efd255\") " pod="openstack/keystone-75dc574896-4vrbs" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.954475 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t457b\" (UniqueName: \"kubernetes.io/projected/bdd2c50b-481a-4ab7-90ce-86cb69efd255-kube-api-access-t457b\") pod \"keystone-75dc574896-4vrbs\" (UID: \"bdd2c50b-481a-4ab7-90ce-86cb69efd255\") " pod="openstack/keystone-75dc574896-4vrbs" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.954554 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdd2c50b-481a-4ab7-90ce-86cb69efd255-public-tls-certs\") pod \"keystone-75dc574896-4vrbs\" (UID: \"bdd2c50b-481a-4ab7-90ce-86cb69efd255\") " pod="openstack/keystone-75dc574896-4vrbs" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.954592 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bdd2c50b-481a-4ab7-90ce-86cb69efd255-scripts\") pod \"keystone-75dc574896-4vrbs\" (UID: \"bdd2c50b-481a-4ab7-90ce-86cb69efd255\") " pod="openstack/keystone-75dc574896-4vrbs" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.954635 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdd2c50b-481a-4ab7-90ce-86cb69efd255-combined-ca-bundle\") pod \"keystone-75dc574896-4vrbs\" (UID: \"bdd2c50b-481a-4ab7-90ce-86cb69efd255\") " pod="openstack/keystone-75dc574896-4vrbs" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.954687 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdd2c50b-481a-4ab7-90ce-86cb69efd255-config-data\") pod \"keystone-75dc574896-4vrbs\" (UID: \"bdd2c50b-481a-4ab7-90ce-86cb69efd255\") " pod="openstack/keystone-75dc574896-4vrbs" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.958429 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/bdd2c50b-481a-4ab7-90ce-86cb69efd255-credential-keys\") pod \"keystone-75dc574896-4vrbs\" (UID: \"bdd2c50b-481a-4ab7-90ce-86cb69efd255\") " pod="openstack/keystone-75dc574896-4vrbs" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.958965 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdd2c50b-481a-4ab7-90ce-86cb69efd255-combined-ca-bundle\") pod \"keystone-75dc574896-4vrbs\" (UID: \"bdd2c50b-481a-4ab7-90ce-86cb69efd255\") " pod="openstack/keystone-75dc574896-4vrbs" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.959614 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdd2c50b-481a-4ab7-90ce-86cb69efd255-config-data\") pod \"keystone-75dc574896-4vrbs\" (UID: \"bdd2c50b-481a-4ab7-90ce-86cb69efd255\") " pod="openstack/keystone-75dc574896-4vrbs" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.960293 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bdd2c50b-481a-4ab7-90ce-86cb69efd255-scripts\") pod \"keystone-75dc574896-4vrbs\" (UID: \"bdd2c50b-481a-4ab7-90ce-86cb69efd255\") " pod="openstack/keystone-75dc574896-4vrbs" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.960998 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdd2c50b-481a-4ab7-90ce-86cb69efd255-internal-tls-certs\") pod \"keystone-75dc574896-4vrbs\" (UID: \"bdd2c50b-481a-4ab7-90ce-86cb69efd255\") " pod="openstack/keystone-75dc574896-4vrbs" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.962734 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdd2c50b-481a-4ab7-90ce-86cb69efd255-public-tls-certs\") pod \"keystone-75dc574896-4vrbs\" (UID: \"bdd2c50b-481a-4ab7-90ce-86cb69efd255\") " pod="openstack/keystone-75dc574896-4vrbs" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.963525 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bdd2c50b-481a-4ab7-90ce-86cb69efd255-fernet-keys\") pod \"keystone-75dc574896-4vrbs\" (UID: \"bdd2c50b-481a-4ab7-90ce-86cb69efd255\") " pod="openstack/keystone-75dc574896-4vrbs" Dec 06 03:59:18 crc kubenswrapper[4802]: I1206 03:59:18.970609 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t457b\" (UniqueName: \"kubernetes.io/projected/bdd2c50b-481a-4ab7-90ce-86cb69efd255-kube-api-access-t457b\") pod \"keystone-75dc574896-4vrbs\" (UID: \"bdd2c50b-481a-4ab7-90ce-86cb69efd255\") " pod="openstack/keystone-75dc574896-4vrbs" Dec 06 03:59:19 crc kubenswrapper[4802]: I1206 03:59:19.054201 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 03:59:19 crc kubenswrapper[4802]: I1206 03:59:19.055493 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4abd9ed3-07e6-46a6-8539-85fc5c624145-logs\") pod \"4abd9ed3-07e6-46a6-8539-85fc5c624145\" (UID: \"4abd9ed3-07e6-46a6-8539-85fc5c624145\") " Dec 06 03:59:19 crc kubenswrapper[4802]: I1206 03:59:19.055569 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4abd9ed3-07e6-46a6-8539-85fc5c624145-scripts\") pod \"4abd9ed3-07e6-46a6-8539-85fc5c624145\" (UID: \"4abd9ed3-07e6-46a6-8539-85fc5c624145\") " Dec 06 03:59:19 crc kubenswrapper[4802]: I1206 03:59:19.055595 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h57kj\" (UniqueName: \"kubernetes.io/projected/4abd9ed3-07e6-46a6-8539-85fc5c624145-kube-api-access-h57kj\") pod \"4abd9ed3-07e6-46a6-8539-85fc5c624145\" (UID: \"4abd9ed3-07e6-46a6-8539-85fc5c624145\") " Dec 06 03:59:19 crc kubenswrapper[4802]: I1206 03:59:19.055709 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4abd9ed3-07e6-46a6-8539-85fc5c624145-config-data\") pod \"4abd9ed3-07e6-46a6-8539-85fc5c624145\" (UID: \"4abd9ed3-07e6-46a6-8539-85fc5c624145\") " Dec 06 03:59:19 crc kubenswrapper[4802]: I1206 03:59:19.055813 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4abd9ed3-07e6-46a6-8539-85fc5c624145-combined-ca-bundle\") pod \"4abd9ed3-07e6-46a6-8539-85fc5c624145\" (UID: \"4abd9ed3-07e6-46a6-8539-85fc5c624145\") " Dec 06 03:59:19 crc kubenswrapper[4802]: I1206 03:59:19.056278 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4abd9ed3-07e6-46a6-8539-85fc5c624145-logs" (OuterVolumeSpecName: "logs") pod "4abd9ed3-07e6-46a6-8539-85fc5c624145" (UID: "4abd9ed3-07e6-46a6-8539-85fc5c624145"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:59:19 crc kubenswrapper[4802]: I1206 03:59:19.058928 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4abd9ed3-07e6-46a6-8539-85fc5c624145-scripts" (OuterVolumeSpecName: "scripts") pod "4abd9ed3-07e6-46a6-8539-85fc5c624145" (UID: "4abd9ed3-07e6-46a6-8539-85fc5c624145"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:59:19 crc kubenswrapper[4802]: I1206 03:59:19.062147 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4abd9ed3-07e6-46a6-8539-85fc5c624145-kube-api-access-h57kj" (OuterVolumeSpecName: "kube-api-access-h57kj") pod "4abd9ed3-07e6-46a6-8539-85fc5c624145" (UID: "4abd9ed3-07e6-46a6-8539-85fc5c624145"). InnerVolumeSpecName "kube-api-access-h57kj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:59:19 crc kubenswrapper[4802]: I1206 03:59:19.065464 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-75dc574896-4vrbs" Dec 06 03:59:19 crc kubenswrapper[4802]: I1206 03:59:19.077873 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4abd9ed3-07e6-46a6-8539-85fc5c624145-config-data" (OuterVolumeSpecName: "config-data") pod "4abd9ed3-07e6-46a6-8539-85fc5c624145" (UID: "4abd9ed3-07e6-46a6-8539-85fc5c624145"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:59:19 crc kubenswrapper[4802]: I1206 03:59:19.088053 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4abd9ed3-07e6-46a6-8539-85fc5c624145-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4abd9ed3-07e6-46a6-8539-85fc5c624145" (UID: "4abd9ed3-07e6-46a6-8539-85fc5c624145"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:59:19 crc kubenswrapper[4802]: I1206 03:59:19.158808 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4abd9ed3-07e6-46a6-8539-85fc5c624145-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:19 crc kubenswrapper[4802]: I1206 03:59:19.158839 4802 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4abd9ed3-07e6-46a6-8539-85fc5c624145-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:19 crc kubenswrapper[4802]: I1206 03:59:19.158849 4802 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4abd9ed3-07e6-46a6-8539-85fc5c624145-logs\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:19 crc kubenswrapper[4802]: I1206 03:59:19.158857 4802 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4abd9ed3-07e6-46a6-8539-85fc5c624145-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:19 crc kubenswrapper[4802]: I1206 03:59:19.158866 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h57kj\" (UniqueName: \"kubernetes.io/projected/4abd9ed3-07e6-46a6-8539-85fc5c624145-kube-api-access-h57kj\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:19 crc kubenswrapper[4802]: I1206 03:59:19.464184 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd66ded5-a0ce-42a0-86a6-03a727128fb9" path="/var/lib/kubelet/pods/dd66ded5-a0ce-42a0-86a6-03a727128fb9/volumes" Dec 06 03:59:19 crc kubenswrapper[4802]: I1206 03:59:19.613383 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f8c7901e-3653-4d82-a15f-dd12cf4ab8c5","Type":"ContainerStarted","Data":"43393bb092a35409606f84f7092e94a2e0eec77e324a88aed2126f8d32b543b1"} Dec 06 03:59:19 crc kubenswrapper[4802]: I1206 03:59:19.616400 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-dxvz6" event={"ID":"4abd9ed3-07e6-46a6-8539-85fc5c624145","Type":"ContainerDied","Data":"91add2d6376f69256916a934641c5d1403c2313455ee03e98ad61e269d1bff21"} Dec 06 03:59:19 crc kubenswrapper[4802]: I1206 03:59:19.616428 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-dxvz6" Dec 06 03:59:19 crc kubenswrapper[4802]: I1206 03:59:19.616451 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="91add2d6376f69256916a934641c5d1403c2313455ee03e98ad61e269d1bff21" Dec 06 03:59:19 crc kubenswrapper[4802]: I1206 03:59:19.630476 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-75dc574896-4vrbs"] Dec 06 03:59:19 crc kubenswrapper[4802]: W1206 03:59:19.639035 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbdd2c50b_481a_4ab7_90ce_86cb69efd255.slice/crio-0e2ccce3e097ae73526c17b07324c3e3c1d679ac791f6c1d0a0bceb76a0359af WatchSource:0}: Error finding container 0e2ccce3e097ae73526c17b07324c3e3c1d679ac791f6c1d0a0bceb76a0359af: Status 404 returned error can't find the container with id 0e2ccce3e097ae73526c17b07324c3e3c1d679ac791f6c1d0a0bceb76a0359af Dec 06 03:59:19 crc kubenswrapper[4802]: I1206 03:59:19.669593 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=7.66956819 podStartE2EDuration="7.66956819s" podCreationTimestamp="2025-12-06 03:59:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:59:19.650528861 +0000 UTC m=+1152.522438033" watchObservedRunningTime="2025-12-06 03:59:19.66956819 +0000 UTC m=+1152.541477342" Dec 06 03:59:19 crc kubenswrapper[4802]: I1206 03:59:19.709228 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 03:59:19 crc kubenswrapper[4802]: I1206 03:59:19.734280 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-788d784b-hjc7w"] Dec 06 03:59:19 crc kubenswrapper[4802]: E1206 03:59:19.734961 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4abd9ed3-07e6-46a6-8539-85fc5c624145" containerName="placement-db-sync" Dec 06 03:59:19 crc kubenswrapper[4802]: I1206 03:59:19.734976 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="4abd9ed3-07e6-46a6-8539-85fc5c624145" containerName="placement-db-sync" Dec 06 03:59:19 crc kubenswrapper[4802]: I1206 03:59:19.735164 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="4abd9ed3-07e6-46a6-8539-85fc5c624145" containerName="placement-db-sync" Dec 06 03:59:19 crc kubenswrapper[4802]: I1206 03:59:19.736083 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-788d784b-hjc7w" Dec 06 03:59:19 crc kubenswrapper[4802]: I1206 03:59:19.741424 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 06 03:59:19 crc kubenswrapper[4802]: I1206 03:59:19.741678 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 06 03:59:19 crc kubenswrapper[4802]: I1206 03:59:19.741832 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-xvxk2" Dec 06 03:59:19 crc kubenswrapper[4802]: I1206 03:59:19.741905 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 06 03:59:19 crc kubenswrapper[4802]: I1206 03:59:19.742010 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 06 03:59:19 crc kubenswrapper[4802]: I1206 03:59:19.759176 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-788d784b-hjc7w"] Dec 06 03:59:19 crc kubenswrapper[4802]: I1206 03:59:19.880882 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5800f6d2-346a-4e9a-8585-e839ddfd035f-config-data\") pod \"placement-788d784b-hjc7w\" (UID: \"5800f6d2-346a-4e9a-8585-e839ddfd035f\") " pod="openstack/placement-788d784b-hjc7w" Dec 06 03:59:19 crc kubenswrapper[4802]: I1206 03:59:19.881081 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wxfc\" (UniqueName: \"kubernetes.io/projected/5800f6d2-346a-4e9a-8585-e839ddfd035f-kube-api-access-2wxfc\") pod \"placement-788d784b-hjc7w\" (UID: \"5800f6d2-346a-4e9a-8585-e839ddfd035f\") " pod="openstack/placement-788d784b-hjc7w" Dec 06 03:59:19 crc kubenswrapper[4802]: I1206 03:59:19.881245 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5800f6d2-346a-4e9a-8585-e839ddfd035f-combined-ca-bundle\") pod \"placement-788d784b-hjc7w\" (UID: \"5800f6d2-346a-4e9a-8585-e839ddfd035f\") " pod="openstack/placement-788d784b-hjc7w" Dec 06 03:59:19 crc kubenswrapper[4802]: I1206 03:59:19.881321 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5800f6d2-346a-4e9a-8585-e839ddfd035f-internal-tls-certs\") pod \"placement-788d784b-hjc7w\" (UID: \"5800f6d2-346a-4e9a-8585-e839ddfd035f\") " pod="openstack/placement-788d784b-hjc7w" Dec 06 03:59:19 crc kubenswrapper[4802]: I1206 03:59:19.881446 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5800f6d2-346a-4e9a-8585-e839ddfd035f-logs\") pod \"placement-788d784b-hjc7w\" (UID: \"5800f6d2-346a-4e9a-8585-e839ddfd035f\") " pod="openstack/placement-788d784b-hjc7w" Dec 06 03:59:19 crc kubenswrapper[4802]: I1206 03:59:19.881501 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5800f6d2-346a-4e9a-8585-e839ddfd035f-scripts\") pod \"placement-788d784b-hjc7w\" (UID: \"5800f6d2-346a-4e9a-8585-e839ddfd035f\") " pod="openstack/placement-788d784b-hjc7w" Dec 06 03:59:19 crc kubenswrapper[4802]: I1206 03:59:19.881527 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5800f6d2-346a-4e9a-8585-e839ddfd035f-public-tls-certs\") pod \"placement-788d784b-hjc7w\" (UID: \"5800f6d2-346a-4e9a-8585-e839ddfd035f\") " pod="openstack/placement-788d784b-hjc7w" Dec 06 03:59:19 crc kubenswrapper[4802]: I1206 03:59:19.891238 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-qtg4k" Dec 06 03:59:19 crc kubenswrapper[4802]: I1206 03:59:19.983472 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wxfc\" (UniqueName: \"kubernetes.io/projected/5800f6d2-346a-4e9a-8585-e839ddfd035f-kube-api-access-2wxfc\") pod \"placement-788d784b-hjc7w\" (UID: \"5800f6d2-346a-4e9a-8585-e839ddfd035f\") " pod="openstack/placement-788d784b-hjc7w" Dec 06 03:59:19 crc kubenswrapper[4802]: I1206 03:59:19.983553 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5800f6d2-346a-4e9a-8585-e839ddfd035f-combined-ca-bundle\") pod \"placement-788d784b-hjc7w\" (UID: \"5800f6d2-346a-4e9a-8585-e839ddfd035f\") " pod="openstack/placement-788d784b-hjc7w" Dec 06 03:59:19 crc kubenswrapper[4802]: I1206 03:59:19.983582 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5800f6d2-346a-4e9a-8585-e839ddfd035f-internal-tls-certs\") pod \"placement-788d784b-hjc7w\" (UID: \"5800f6d2-346a-4e9a-8585-e839ddfd035f\") " pod="openstack/placement-788d784b-hjc7w" Dec 06 03:59:19 crc kubenswrapper[4802]: I1206 03:59:19.983628 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5800f6d2-346a-4e9a-8585-e839ddfd035f-logs\") pod \"placement-788d784b-hjc7w\" (UID: \"5800f6d2-346a-4e9a-8585-e839ddfd035f\") " pod="openstack/placement-788d784b-hjc7w" Dec 06 03:59:19 crc kubenswrapper[4802]: I1206 03:59:19.983662 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5800f6d2-346a-4e9a-8585-e839ddfd035f-scripts\") pod \"placement-788d784b-hjc7w\" (UID: \"5800f6d2-346a-4e9a-8585-e839ddfd035f\") " pod="openstack/placement-788d784b-hjc7w" Dec 06 03:59:19 crc kubenswrapper[4802]: I1206 03:59:19.983691 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5800f6d2-346a-4e9a-8585-e839ddfd035f-public-tls-certs\") pod \"placement-788d784b-hjc7w\" (UID: \"5800f6d2-346a-4e9a-8585-e839ddfd035f\") " pod="openstack/placement-788d784b-hjc7w" Dec 06 03:59:19 crc kubenswrapper[4802]: I1206 03:59:19.983778 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5800f6d2-346a-4e9a-8585-e839ddfd035f-config-data\") pod \"placement-788d784b-hjc7w\" (UID: \"5800f6d2-346a-4e9a-8585-e839ddfd035f\") " pod="openstack/placement-788d784b-hjc7w" Dec 06 03:59:19 crc kubenswrapper[4802]: I1206 03:59:19.984537 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5800f6d2-346a-4e9a-8585-e839ddfd035f-logs\") pod \"placement-788d784b-hjc7w\" (UID: \"5800f6d2-346a-4e9a-8585-e839ddfd035f\") " pod="openstack/placement-788d784b-hjc7w" Dec 06 03:59:19 crc kubenswrapper[4802]: I1206 03:59:19.988613 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5800f6d2-346a-4e9a-8585-e839ddfd035f-internal-tls-certs\") pod \"placement-788d784b-hjc7w\" (UID: \"5800f6d2-346a-4e9a-8585-e839ddfd035f\") " pod="openstack/placement-788d784b-hjc7w" Dec 06 03:59:19 crc kubenswrapper[4802]: I1206 03:59:19.989490 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5800f6d2-346a-4e9a-8585-e839ddfd035f-scripts\") pod \"placement-788d784b-hjc7w\" (UID: \"5800f6d2-346a-4e9a-8585-e839ddfd035f\") " pod="openstack/placement-788d784b-hjc7w" Dec 06 03:59:19 crc kubenswrapper[4802]: I1206 03:59:19.990429 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5800f6d2-346a-4e9a-8585-e839ddfd035f-config-data\") pod \"placement-788d784b-hjc7w\" (UID: \"5800f6d2-346a-4e9a-8585-e839ddfd035f\") " pod="openstack/placement-788d784b-hjc7w" Dec 06 03:59:19 crc kubenswrapper[4802]: I1206 03:59:19.990732 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5800f6d2-346a-4e9a-8585-e839ddfd035f-combined-ca-bundle\") pod \"placement-788d784b-hjc7w\" (UID: \"5800f6d2-346a-4e9a-8585-e839ddfd035f\") " pod="openstack/placement-788d784b-hjc7w" Dec 06 03:59:19 crc kubenswrapper[4802]: I1206 03:59:19.993076 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5800f6d2-346a-4e9a-8585-e839ddfd035f-public-tls-certs\") pod \"placement-788d784b-hjc7w\" (UID: \"5800f6d2-346a-4e9a-8585-e839ddfd035f\") " pod="openstack/placement-788d784b-hjc7w" Dec 06 03:59:20 crc kubenswrapper[4802]: I1206 03:59:20.001831 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wxfc\" (UniqueName: \"kubernetes.io/projected/5800f6d2-346a-4e9a-8585-e839ddfd035f-kube-api-access-2wxfc\") pod \"placement-788d784b-hjc7w\" (UID: \"5800f6d2-346a-4e9a-8585-e839ddfd035f\") " pod="openstack/placement-788d784b-hjc7w" Dec 06 03:59:20 crc kubenswrapper[4802]: I1206 03:59:20.085327 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5c636931-d435-47ec-b770-80dc1ee60756-db-sync-config-data\") pod \"5c636931-d435-47ec-b770-80dc1ee60756\" (UID: \"5c636931-d435-47ec-b770-80dc1ee60756\") " Dec 06 03:59:20 crc kubenswrapper[4802]: I1206 03:59:20.085523 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c636931-d435-47ec-b770-80dc1ee60756-combined-ca-bundle\") pod \"5c636931-d435-47ec-b770-80dc1ee60756\" (UID: \"5c636931-d435-47ec-b770-80dc1ee60756\") " Dec 06 03:59:20 crc kubenswrapper[4802]: I1206 03:59:20.085656 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4fzfx\" (UniqueName: \"kubernetes.io/projected/5c636931-d435-47ec-b770-80dc1ee60756-kube-api-access-4fzfx\") pod \"5c636931-d435-47ec-b770-80dc1ee60756\" (UID: \"5c636931-d435-47ec-b770-80dc1ee60756\") " Dec 06 03:59:20 crc kubenswrapper[4802]: I1206 03:59:20.090635 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c636931-d435-47ec-b770-80dc1ee60756-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "5c636931-d435-47ec-b770-80dc1ee60756" (UID: "5c636931-d435-47ec-b770-80dc1ee60756"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:59:20 crc kubenswrapper[4802]: I1206 03:59:20.092785 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c636931-d435-47ec-b770-80dc1ee60756-kube-api-access-4fzfx" (OuterVolumeSpecName: "kube-api-access-4fzfx") pod "5c636931-d435-47ec-b770-80dc1ee60756" (UID: "5c636931-d435-47ec-b770-80dc1ee60756"). InnerVolumeSpecName "kube-api-access-4fzfx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:59:20 crc kubenswrapper[4802]: I1206 03:59:20.112637 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-788d784b-hjc7w" Dec 06 03:59:20 crc kubenswrapper[4802]: I1206 03:59:20.116409 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c636931-d435-47ec-b770-80dc1ee60756-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5c636931-d435-47ec-b770-80dc1ee60756" (UID: "5c636931-d435-47ec-b770-80dc1ee60756"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:59:20 crc kubenswrapper[4802]: I1206 03:59:20.187809 4802 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c636931-d435-47ec-b770-80dc1ee60756-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:20 crc kubenswrapper[4802]: I1206 03:59:20.188146 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4fzfx\" (UniqueName: \"kubernetes.io/projected/5c636931-d435-47ec-b770-80dc1ee60756-kube-api-access-4fzfx\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:20 crc kubenswrapper[4802]: I1206 03:59:20.188161 4802 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5c636931-d435-47ec-b770-80dc1ee60756-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:20 crc kubenswrapper[4802]: I1206 03:59:20.555041 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-788d784b-hjc7w"] Dec 06 03:59:20 crc kubenswrapper[4802]: W1206 03:59:20.562946 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5800f6d2_346a_4e9a_8585_e839ddfd035f.slice/crio-4648972eab2fd111fc318045add56811c2e1de34f049a6905c41657bf501e0be WatchSource:0}: Error finding container 4648972eab2fd111fc318045add56811c2e1de34f049a6905c41657bf501e0be: Status 404 returned error can't find the container with id 4648972eab2fd111fc318045add56811c2e1de34f049a6905c41657bf501e0be Dec 06 03:59:20 crc kubenswrapper[4802]: I1206 03:59:20.645410 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-75dc574896-4vrbs" event={"ID":"bdd2c50b-481a-4ab7-90ce-86cb69efd255","Type":"ContainerStarted","Data":"0e78ef584f8deaae67ecd3e413ec314bb1fd2f80448ba64dfd87f4cd8dfc0d2e"} Dec 06 03:59:20 crc kubenswrapper[4802]: I1206 03:59:20.645456 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-75dc574896-4vrbs" event={"ID":"bdd2c50b-481a-4ab7-90ce-86cb69efd255","Type":"ContainerStarted","Data":"0e2ccce3e097ae73526c17b07324c3e3c1d679ac791f6c1d0a0bceb76a0359af"} Dec 06 03:59:20 crc kubenswrapper[4802]: I1206 03:59:20.645487 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-75dc574896-4vrbs" Dec 06 03:59:20 crc kubenswrapper[4802]: I1206 03:59:20.648743 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"22294441-50c6-4dc0-80f1-4b14c1fc62a9","Type":"ContainerStarted","Data":"58648fb9db1348bd1d6f78c8f2f582f88632a5f63e2e397444c22fc545cb0cf0"} Dec 06 03:59:20 crc kubenswrapper[4802]: I1206 03:59:20.649542 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-788d784b-hjc7w" event={"ID":"5800f6d2-346a-4e9a-8585-e839ddfd035f","Type":"ContainerStarted","Data":"4648972eab2fd111fc318045add56811c2e1de34f049a6905c41657bf501e0be"} Dec 06 03:59:20 crc kubenswrapper[4802]: I1206 03:59:20.654429 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-qtg4k" Dec 06 03:59:20 crc kubenswrapper[4802]: I1206 03:59:20.654463 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-qtg4k" event={"ID":"5c636931-d435-47ec-b770-80dc1ee60756","Type":"ContainerDied","Data":"8c51e9bbf9a382e680ba73c38c4086dfd7247c8036b228086da905f8a953b122"} Dec 06 03:59:20 crc kubenswrapper[4802]: I1206 03:59:20.654483 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8c51e9bbf9a382e680ba73c38c4086dfd7247c8036b228086da905f8a953b122" Dec 06 03:59:20 crc kubenswrapper[4802]: I1206 03:59:20.668004 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-75dc574896-4vrbs" podStartSLOduration=2.667989067 podStartE2EDuration="2.667989067s" podCreationTimestamp="2025-12-06 03:59:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:59:20.665722825 +0000 UTC m=+1153.537631997" watchObservedRunningTime="2025-12-06 03:59:20.667989067 +0000 UTC m=+1153.539898219" Dec 06 03:59:20 crc kubenswrapper[4802]: I1206 03:59:20.779461 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-68998b57f5-6qgkl"] Dec 06 03:59:20 crc kubenswrapper[4802]: E1206 03:59:20.785424 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c636931-d435-47ec-b770-80dc1ee60756" containerName="barbican-db-sync" Dec 06 03:59:20 crc kubenswrapper[4802]: I1206 03:59:20.785463 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c636931-d435-47ec-b770-80dc1ee60756" containerName="barbican-db-sync" Dec 06 03:59:20 crc kubenswrapper[4802]: I1206 03:59:20.785720 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c636931-d435-47ec-b770-80dc1ee60756" containerName="barbican-db-sync" Dec 06 03:59:20 crc kubenswrapper[4802]: I1206 03:59:20.786585 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-68998b57f5-6qgkl" Dec 06 03:59:20 crc kubenswrapper[4802]: I1206 03:59:20.798121 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 06 03:59:20 crc kubenswrapper[4802]: I1206 03:59:20.798832 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-qfhm9" Dec 06 03:59:20 crc kubenswrapper[4802]: I1206 03:59:20.799009 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 06 03:59:20 crc kubenswrapper[4802]: I1206 03:59:20.811443 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-68998b57f5-6qgkl"] Dec 06 03:59:20 crc kubenswrapper[4802]: I1206 03:59:20.838145 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-554f6fb55d-hdgjt"] Dec 06 03:59:20 crc kubenswrapper[4802]: I1206 03:59:20.841000 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-554f6fb55d-hdgjt" Dec 06 03:59:20 crc kubenswrapper[4802]: I1206 03:59:20.845136 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 06 03:59:20 crc kubenswrapper[4802]: I1206 03:59:20.860714 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-554f6fb55d-hdgjt"] Dec 06 03:59:20 crc kubenswrapper[4802]: I1206 03:59:20.893774 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7c67bffd47-mrlr7"] Dec 06 03:59:20 crc kubenswrapper[4802]: I1206 03:59:20.895160 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c67bffd47-mrlr7" Dec 06 03:59:20 crc kubenswrapper[4802]: I1206 03:59:20.901175 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcvz5\" (UniqueName: \"kubernetes.io/projected/57a20a46-fdfb-410f-8a13-0b58fa5751da-kube-api-access-rcvz5\") pod \"barbican-worker-68998b57f5-6qgkl\" (UID: \"57a20a46-fdfb-410f-8a13-0b58fa5751da\") " pod="openstack/barbican-worker-68998b57f5-6qgkl" Dec 06 03:59:20 crc kubenswrapper[4802]: I1206 03:59:20.901257 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57a20a46-fdfb-410f-8a13-0b58fa5751da-combined-ca-bundle\") pod \"barbican-worker-68998b57f5-6qgkl\" (UID: \"57a20a46-fdfb-410f-8a13-0b58fa5751da\") " pod="openstack/barbican-worker-68998b57f5-6qgkl" Dec 06 03:59:20 crc kubenswrapper[4802]: I1206 03:59:20.901326 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57a20a46-fdfb-410f-8a13-0b58fa5751da-config-data\") pod \"barbican-worker-68998b57f5-6qgkl\" (UID: \"57a20a46-fdfb-410f-8a13-0b58fa5751da\") " pod="openstack/barbican-worker-68998b57f5-6qgkl" Dec 06 03:59:20 crc kubenswrapper[4802]: I1206 03:59:20.901347 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/57a20a46-fdfb-410f-8a13-0b58fa5751da-config-data-custom\") pod \"barbican-worker-68998b57f5-6qgkl\" (UID: \"57a20a46-fdfb-410f-8a13-0b58fa5751da\") " pod="openstack/barbican-worker-68998b57f5-6qgkl" Dec 06 03:59:20 crc kubenswrapper[4802]: I1206 03:59:20.901383 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/57a20a46-fdfb-410f-8a13-0b58fa5751da-logs\") pod \"barbican-worker-68998b57f5-6qgkl\" (UID: \"57a20a46-fdfb-410f-8a13-0b58fa5751da\") " pod="openstack/barbican-worker-68998b57f5-6qgkl" Dec 06 03:59:20 crc kubenswrapper[4802]: I1206 03:59:20.902184 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c67bffd47-mrlr7"] Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.002447 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57a20a46-fdfb-410f-8a13-0b58fa5751da-combined-ca-bundle\") pod \"barbican-worker-68998b57f5-6qgkl\" (UID: \"57a20a46-fdfb-410f-8a13-0b58fa5751da\") " pod="openstack/barbican-worker-68998b57f5-6qgkl" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.002499 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57a20a46-fdfb-410f-8a13-0b58fa5751da-config-data\") pod \"barbican-worker-68998b57f5-6qgkl\" (UID: \"57a20a46-fdfb-410f-8a13-0b58fa5751da\") " pod="openstack/barbican-worker-68998b57f5-6qgkl" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.002518 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/57a20a46-fdfb-410f-8a13-0b58fa5751da-config-data-custom\") pod \"barbican-worker-68998b57f5-6qgkl\" (UID: \"57a20a46-fdfb-410f-8a13-0b58fa5751da\") " pod="openstack/barbican-worker-68998b57f5-6qgkl" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.002545 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b443b1b9-ec2c-499d-ae4d-1937c6f156f7-config-data-custom\") pod \"barbican-keystone-listener-554f6fb55d-hdgjt\" (UID: \"b443b1b9-ec2c-499d-ae4d-1937c6f156f7\") " pod="openstack/barbican-keystone-listener-554f6fb55d-hdgjt" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.002577 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b443b1b9-ec2c-499d-ae4d-1937c6f156f7-logs\") pod \"barbican-keystone-listener-554f6fb55d-hdgjt\" (UID: \"b443b1b9-ec2c-499d-ae4d-1937c6f156f7\") " pod="openstack/barbican-keystone-listener-554f6fb55d-hdgjt" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.002593 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/57a20a46-fdfb-410f-8a13-0b58fa5751da-logs\") pod \"barbican-worker-68998b57f5-6qgkl\" (UID: \"57a20a46-fdfb-410f-8a13-0b58fa5751da\") " pod="openstack/barbican-worker-68998b57f5-6qgkl" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.002615 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c-config\") pod \"dnsmasq-dns-7c67bffd47-mrlr7\" (UID: \"a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c\") " pod="openstack/dnsmasq-dns-7c67bffd47-mrlr7" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.002637 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c-dns-svc\") pod \"dnsmasq-dns-7c67bffd47-mrlr7\" (UID: \"a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c\") " pod="openstack/dnsmasq-dns-7c67bffd47-mrlr7" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.002664 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b443b1b9-ec2c-499d-ae4d-1937c6f156f7-config-data\") pod \"barbican-keystone-listener-554f6fb55d-hdgjt\" (UID: \"b443b1b9-ec2c-499d-ae4d-1937c6f156f7\") " pod="openstack/barbican-keystone-listener-554f6fb55d-hdgjt" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.002680 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b443b1b9-ec2c-499d-ae4d-1937c6f156f7-combined-ca-bundle\") pod \"barbican-keystone-listener-554f6fb55d-hdgjt\" (UID: \"b443b1b9-ec2c-499d-ae4d-1937c6f156f7\") " pod="openstack/barbican-keystone-listener-554f6fb55d-hdgjt" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.002696 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kzf2\" (UniqueName: \"kubernetes.io/projected/b443b1b9-ec2c-499d-ae4d-1937c6f156f7-kube-api-access-4kzf2\") pod \"barbican-keystone-listener-554f6fb55d-hdgjt\" (UID: \"b443b1b9-ec2c-499d-ae4d-1937c6f156f7\") " pod="openstack/barbican-keystone-listener-554f6fb55d-hdgjt" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.002728 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c-dns-swift-storage-0\") pod \"dnsmasq-dns-7c67bffd47-mrlr7\" (UID: \"a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c\") " pod="openstack/dnsmasq-dns-7c67bffd47-mrlr7" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.002767 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcvz5\" (UniqueName: \"kubernetes.io/projected/57a20a46-fdfb-410f-8a13-0b58fa5751da-kube-api-access-rcvz5\") pod \"barbican-worker-68998b57f5-6qgkl\" (UID: \"57a20a46-fdfb-410f-8a13-0b58fa5751da\") " pod="openstack/barbican-worker-68998b57f5-6qgkl" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.002788 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c-ovsdbserver-nb\") pod \"dnsmasq-dns-7c67bffd47-mrlr7\" (UID: \"a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c\") " pod="openstack/dnsmasq-dns-7c67bffd47-mrlr7" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.007165 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c-ovsdbserver-sb\") pod \"dnsmasq-dns-7c67bffd47-mrlr7\" (UID: \"a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c\") " pod="openstack/dnsmasq-dns-7c67bffd47-mrlr7" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.007431 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85s6x\" (UniqueName: \"kubernetes.io/projected/a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c-kube-api-access-85s6x\") pod \"dnsmasq-dns-7c67bffd47-mrlr7\" (UID: \"a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c\") " pod="openstack/dnsmasq-dns-7c67bffd47-mrlr7" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.007653 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/57a20a46-fdfb-410f-8a13-0b58fa5751da-logs\") pod \"barbican-worker-68998b57f5-6qgkl\" (UID: \"57a20a46-fdfb-410f-8a13-0b58fa5751da\") " pod="openstack/barbican-worker-68998b57f5-6qgkl" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.009473 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57a20a46-fdfb-410f-8a13-0b58fa5751da-config-data\") pod \"barbican-worker-68998b57f5-6qgkl\" (UID: \"57a20a46-fdfb-410f-8a13-0b58fa5751da\") " pod="openstack/barbican-worker-68998b57f5-6qgkl" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.026833 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6f6b9c98b6-8ndrw"] Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.028134 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6f6b9c98b6-8ndrw" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.029505 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/57a20a46-fdfb-410f-8a13-0b58fa5751da-config-data-custom\") pod \"barbican-worker-68998b57f5-6qgkl\" (UID: \"57a20a46-fdfb-410f-8a13-0b58fa5751da\") " pod="openstack/barbican-worker-68998b57f5-6qgkl" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.030745 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.030998 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rcvz5\" (UniqueName: \"kubernetes.io/projected/57a20a46-fdfb-410f-8a13-0b58fa5751da-kube-api-access-rcvz5\") pod \"barbican-worker-68998b57f5-6qgkl\" (UID: \"57a20a46-fdfb-410f-8a13-0b58fa5751da\") " pod="openstack/barbican-worker-68998b57f5-6qgkl" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.037151 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6f6b9c98b6-8ndrw"] Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.046974 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57a20a46-fdfb-410f-8a13-0b58fa5751da-combined-ca-bundle\") pod \"barbican-worker-68998b57f5-6qgkl\" (UID: \"57a20a46-fdfb-410f-8a13-0b58fa5751da\") " pod="openstack/barbican-worker-68998b57f5-6qgkl" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.108528 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b443b1b9-ec2c-499d-ae4d-1937c6f156f7-config-data\") pod \"barbican-keystone-listener-554f6fb55d-hdgjt\" (UID: \"b443b1b9-ec2c-499d-ae4d-1937c6f156f7\") " pod="openstack/barbican-keystone-listener-554f6fb55d-hdgjt" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.108577 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b443b1b9-ec2c-499d-ae4d-1937c6f156f7-combined-ca-bundle\") pod \"barbican-keystone-listener-554f6fb55d-hdgjt\" (UID: \"b443b1b9-ec2c-499d-ae4d-1937c6f156f7\") " pod="openstack/barbican-keystone-listener-554f6fb55d-hdgjt" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.108602 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kzf2\" (UniqueName: \"kubernetes.io/projected/b443b1b9-ec2c-499d-ae4d-1937c6f156f7-kube-api-access-4kzf2\") pod \"barbican-keystone-listener-554f6fb55d-hdgjt\" (UID: \"b443b1b9-ec2c-499d-ae4d-1937c6f156f7\") " pod="openstack/barbican-keystone-listener-554f6fb55d-hdgjt" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.108717 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c-dns-swift-storage-0\") pod \"dnsmasq-dns-7c67bffd47-mrlr7\" (UID: \"a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c\") " pod="openstack/dnsmasq-dns-7c67bffd47-mrlr7" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.108794 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c-ovsdbserver-nb\") pod \"dnsmasq-dns-7c67bffd47-mrlr7\" (UID: \"a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c\") " pod="openstack/dnsmasq-dns-7c67bffd47-mrlr7" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.108830 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c-ovsdbserver-sb\") pod \"dnsmasq-dns-7c67bffd47-mrlr7\" (UID: \"a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c\") " pod="openstack/dnsmasq-dns-7c67bffd47-mrlr7" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.108849 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85s6x\" (UniqueName: \"kubernetes.io/projected/a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c-kube-api-access-85s6x\") pod \"dnsmasq-dns-7c67bffd47-mrlr7\" (UID: \"a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c\") " pod="openstack/dnsmasq-dns-7c67bffd47-mrlr7" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.108913 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b443b1b9-ec2c-499d-ae4d-1937c6f156f7-config-data-custom\") pod \"barbican-keystone-listener-554f6fb55d-hdgjt\" (UID: \"b443b1b9-ec2c-499d-ae4d-1937c6f156f7\") " pod="openstack/barbican-keystone-listener-554f6fb55d-hdgjt" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.108946 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b443b1b9-ec2c-499d-ae4d-1937c6f156f7-logs\") pod \"barbican-keystone-listener-554f6fb55d-hdgjt\" (UID: \"b443b1b9-ec2c-499d-ae4d-1937c6f156f7\") " pod="openstack/barbican-keystone-listener-554f6fb55d-hdgjt" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.108975 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c-config\") pod \"dnsmasq-dns-7c67bffd47-mrlr7\" (UID: \"a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c\") " pod="openstack/dnsmasq-dns-7c67bffd47-mrlr7" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.109002 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c-dns-svc\") pod \"dnsmasq-dns-7c67bffd47-mrlr7\" (UID: \"a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c\") " pod="openstack/dnsmasq-dns-7c67bffd47-mrlr7" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.109969 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c-ovsdbserver-sb\") pod \"dnsmasq-dns-7c67bffd47-mrlr7\" (UID: \"a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c\") " pod="openstack/dnsmasq-dns-7c67bffd47-mrlr7" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.110072 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b443b1b9-ec2c-499d-ae4d-1937c6f156f7-logs\") pod \"barbican-keystone-listener-554f6fb55d-hdgjt\" (UID: \"b443b1b9-ec2c-499d-ae4d-1937c6f156f7\") " pod="openstack/barbican-keystone-listener-554f6fb55d-hdgjt" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.110469 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c-ovsdbserver-nb\") pod \"dnsmasq-dns-7c67bffd47-mrlr7\" (UID: \"a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c\") " pod="openstack/dnsmasq-dns-7c67bffd47-mrlr7" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.110967 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-68998b57f5-6qgkl" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.110987 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c-dns-svc\") pod \"dnsmasq-dns-7c67bffd47-mrlr7\" (UID: \"a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c\") " pod="openstack/dnsmasq-dns-7c67bffd47-mrlr7" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.111587 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c-config\") pod \"dnsmasq-dns-7c67bffd47-mrlr7\" (UID: \"a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c\") " pod="openstack/dnsmasq-dns-7c67bffd47-mrlr7" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.112104 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c-dns-swift-storage-0\") pod \"dnsmasq-dns-7c67bffd47-mrlr7\" (UID: \"a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c\") " pod="openstack/dnsmasq-dns-7c67bffd47-mrlr7" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.113514 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b443b1b9-ec2c-499d-ae4d-1937c6f156f7-config-data-custom\") pod \"barbican-keystone-listener-554f6fb55d-hdgjt\" (UID: \"b443b1b9-ec2c-499d-ae4d-1937c6f156f7\") " pod="openstack/barbican-keystone-listener-554f6fb55d-hdgjt" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.114655 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b443b1b9-ec2c-499d-ae4d-1937c6f156f7-combined-ca-bundle\") pod \"barbican-keystone-listener-554f6fb55d-hdgjt\" (UID: \"b443b1b9-ec2c-499d-ae4d-1937c6f156f7\") " pod="openstack/barbican-keystone-listener-554f6fb55d-hdgjt" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.116350 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b443b1b9-ec2c-499d-ae4d-1937c6f156f7-config-data\") pod \"barbican-keystone-listener-554f6fb55d-hdgjt\" (UID: \"b443b1b9-ec2c-499d-ae4d-1937c6f156f7\") " pod="openstack/barbican-keystone-listener-554f6fb55d-hdgjt" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.122783 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kzf2\" (UniqueName: \"kubernetes.io/projected/b443b1b9-ec2c-499d-ae4d-1937c6f156f7-kube-api-access-4kzf2\") pod \"barbican-keystone-listener-554f6fb55d-hdgjt\" (UID: \"b443b1b9-ec2c-499d-ae4d-1937c6f156f7\") " pod="openstack/barbican-keystone-listener-554f6fb55d-hdgjt" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.125791 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85s6x\" (UniqueName: \"kubernetes.io/projected/a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c-kube-api-access-85s6x\") pod \"dnsmasq-dns-7c67bffd47-mrlr7\" (UID: \"a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c\") " pod="openstack/dnsmasq-dns-7c67bffd47-mrlr7" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.170055 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-554f6fb55d-hdgjt" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.210790 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d098e8a2-f622-44f7-997f-51c6ff7eceaa-config-data-custom\") pod \"barbican-api-6f6b9c98b6-8ndrw\" (UID: \"d098e8a2-f622-44f7-997f-51c6ff7eceaa\") " pod="openstack/barbican-api-6f6b9c98b6-8ndrw" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.210870 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fr2d\" (UniqueName: \"kubernetes.io/projected/d098e8a2-f622-44f7-997f-51c6ff7eceaa-kube-api-access-4fr2d\") pod \"barbican-api-6f6b9c98b6-8ndrw\" (UID: \"d098e8a2-f622-44f7-997f-51c6ff7eceaa\") " pod="openstack/barbican-api-6f6b9c98b6-8ndrw" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.210890 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d098e8a2-f622-44f7-997f-51c6ff7eceaa-logs\") pod \"barbican-api-6f6b9c98b6-8ndrw\" (UID: \"d098e8a2-f622-44f7-997f-51c6ff7eceaa\") " pod="openstack/barbican-api-6f6b9c98b6-8ndrw" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.210907 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d098e8a2-f622-44f7-997f-51c6ff7eceaa-combined-ca-bundle\") pod \"barbican-api-6f6b9c98b6-8ndrw\" (UID: \"d098e8a2-f622-44f7-997f-51c6ff7eceaa\") " pod="openstack/barbican-api-6f6b9c98b6-8ndrw" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.210935 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d098e8a2-f622-44f7-997f-51c6ff7eceaa-config-data\") pod \"barbican-api-6f6b9c98b6-8ndrw\" (UID: \"d098e8a2-f622-44f7-997f-51c6ff7eceaa\") " pod="openstack/barbican-api-6f6b9c98b6-8ndrw" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.232263 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c67bffd47-mrlr7" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.312024 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fr2d\" (UniqueName: \"kubernetes.io/projected/d098e8a2-f622-44f7-997f-51c6ff7eceaa-kube-api-access-4fr2d\") pod \"barbican-api-6f6b9c98b6-8ndrw\" (UID: \"d098e8a2-f622-44f7-997f-51c6ff7eceaa\") " pod="openstack/barbican-api-6f6b9c98b6-8ndrw" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.313908 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d098e8a2-f622-44f7-997f-51c6ff7eceaa-logs\") pod \"barbican-api-6f6b9c98b6-8ndrw\" (UID: \"d098e8a2-f622-44f7-997f-51c6ff7eceaa\") " pod="openstack/barbican-api-6f6b9c98b6-8ndrw" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.313960 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d098e8a2-f622-44f7-997f-51c6ff7eceaa-combined-ca-bundle\") pod \"barbican-api-6f6b9c98b6-8ndrw\" (UID: \"d098e8a2-f622-44f7-997f-51c6ff7eceaa\") " pod="openstack/barbican-api-6f6b9c98b6-8ndrw" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.314026 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d098e8a2-f622-44f7-997f-51c6ff7eceaa-config-data\") pod \"barbican-api-6f6b9c98b6-8ndrw\" (UID: \"d098e8a2-f622-44f7-997f-51c6ff7eceaa\") " pod="openstack/barbican-api-6f6b9c98b6-8ndrw" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.314290 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d098e8a2-f622-44f7-997f-51c6ff7eceaa-config-data-custom\") pod \"barbican-api-6f6b9c98b6-8ndrw\" (UID: \"d098e8a2-f622-44f7-997f-51c6ff7eceaa\") " pod="openstack/barbican-api-6f6b9c98b6-8ndrw" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.318292 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d098e8a2-f622-44f7-997f-51c6ff7eceaa-logs\") pod \"barbican-api-6f6b9c98b6-8ndrw\" (UID: \"d098e8a2-f622-44f7-997f-51c6ff7eceaa\") " pod="openstack/barbican-api-6f6b9c98b6-8ndrw" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.319892 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d098e8a2-f622-44f7-997f-51c6ff7eceaa-config-data-custom\") pod \"barbican-api-6f6b9c98b6-8ndrw\" (UID: \"d098e8a2-f622-44f7-997f-51c6ff7eceaa\") " pod="openstack/barbican-api-6f6b9c98b6-8ndrw" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.322582 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d098e8a2-f622-44f7-997f-51c6ff7eceaa-config-data\") pod \"barbican-api-6f6b9c98b6-8ndrw\" (UID: \"d098e8a2-f622-44f7-997f-51c6ff7eceaa\") " pod="openstack/barbican-api-6f6b9c98b6-8ndrw" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.328203 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d098e8a2-f622-44f7-997f-51c6ff7eceaa-combined-ca-bundle\") pod \"barbican-api-6f6b9c98b6-8ndrw\" (UID: \"d098e8a2-f622-44f7-997f-51c6ff7eceaa\") " pod="openstack/barbican-api-6f6b9c98b6-8ndrw" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.329762 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fr2d\" (UniqueName: \"kubernetes.io/projected/d098e8a2-f622-44f7-997f-51c6ff7eceaa-kube-api-access-4fr2d\") pod \"barbican-api-6f6b9c98b6-8ndrw\" (UID: \"d098e8a2-f622-44f7-997f-51c6ff7eceaa\") " pod="openstack/barbican-api-6f6b9c98b6-8ndrw" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.366421 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6f6b9c98b6-8ndrw" Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.647125 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-68998b57f5-6qgkl"] Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.667166 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-788d784b-hjc7w" event={"ID":"5800f6d2-346a-4e9a-8585-e839ddfd035f","Type":"ContainerStarted","Data":"5127d36516c5e6c3a6f8b545961a343f93df426a91ca8474f5b17dc33791a9a7"} Dec 06 03:59:21 crc kubenswrapper[4802]: W1206 03:59:21.668218 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod57a20a46_fdfb_410f_8a13_0b58fa5751da.slice/crio-0a93dd470383c30a0a86f3f8d0390493cef05f5d8268c3fec76e1d0967ca6331 WatchSource:0}: Error finding container 0a93dd470383c30a0a86f3f8d0390493cef05f5d8268c3fec76e1d0967ca6331: Status 404 returned error can't find the container with id 0a93dd470383c30a0a86f3f8d0390493cef05f5d8268c3fec76e1d0967ca6331 Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.669783 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"22294441-50c6-4dc0-80f1-4b14c1fc62a9","Type":"ContainerStarted","Data":"c6a619cd1b3bf4fef4f5bd45c896b6e8b20d75cfc21ef84bac1fb69e9b572e9a"} Dec 06 03:59:21 crc kubenswrapper[4802]: W1206 03:59:21.753946 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb443b1b9_ec2c_499d_ae4d_1937c6f156f7.slice/crio-161e0100abf441acbf515c8c8c61a6879fe4daeb6354effce90a943bf9379c77 WatchSource:0}: Error finding container 161e0100abf441acbf515c8c8c61a6879fe4daeb6354effce90a943bf9379c77: Status 404 returned error can't find the container with id 161e0100abf441acbf515c8c8c61a6879fe4daeb6354effce90a943bf9379c77 Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.770769 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-554f6fb55d-hdgjt"] Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.783033 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c67bffd47-mrlr7"] Dec 06 03:59:21 crc kubenswrapper[4802]: I1206 03:59:21.864479 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6f6b9c98b6-8ndrw"] Dec 06 03:59:21 crc kubenswrapper[4802]: W1206 03:59:21.873504 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd098e8a2_f622_44f7_997f_51c6ff7eceaa.slice/crio-eccf0d8cb38ce9260b1f4b144eb77a94a99e70245a85f95655db5812158f3d30 WatchSource:0}: Error finding container eccf0d8cb38ce9260b1f4b144eb77a94a99e70245a85f95655db5812158f3d30: Status 404 returned error can't find the container with id eccf0d8cb38ce9260b1f4b144eb77a94a99e70245a85f95655db5812158f3d30 Dec 06 03:59:22 crc kubenswrapper[4802]: I1206 03:59:22.679479 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-554f6fb55d-hdgjt" event={"ID":"b443b1b9-ec2c-499d-ae4d-1937c6f156f7","Type":"ContainerStarted","Data":"161e0100abf441acbf515c8c8c61a6879fe4daeb6354effce90a943bf9379c77"} Dec 06 03:59:22 crc kubenswrapper[4802]: I1206 03:59:22.680512 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6f6b9c98b6-8ndrw" event={"ID":"d098e8a2-f622-44f7-997f-51c6ff7eceaa","Type":"ContainerStarted","Data":"eccf0d8cb38ce9260b1f4b144eb77a94a99e70245a85f95655db5812158f3d30"} Dec 06 03:59:22 crc kubenswrapper[4802]: I1206 03:59:22.681557 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c67bffd47-mrlr7" event={"ID":"a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c","Type":"ContainerStarted","Data":"f09944a2b14a1af41f9b3ee21e231341ffe633826fe185e0eb96f81995808058"} Dec 06 03:59:22 crc kubenswrapper[4802]: I1206 03:59:22.683823 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-68998b57f5-6qgkl" event={"ID":"57a20a46-fdfb-410f-8a13-0b58fa5751da","Type":"ContainerStarted","Data":"0a93dd470383c30a0a86f3f8d0390493cef05f5d8268c3fec76e1d0967ca6331"} Dec 06 03:59:22 crc kubenswrapper[4802]: I1206 03:59:22.858260 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 06 03:59:22 crc kubenswrapper[4802]: I1206 03:59:22.858305 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 06 03:59:22 crc kubenswrapper[4802]: I1206 03:59:22.891087 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 06 03:59:22 crc kubenswrapper[4802]: I1206 03:59:22.897575 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 06 03:59:23 crc kubenswrapper[4802]: I1206 03:59:23.695089 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6f6b9c98b6-8ndrw" event={"ID":"d098e8a2-f622-44f7-997f-51c6ff7eceaa","Type":"ContainerStarted","Data":"dac72a8abf11cd63050015a15e0298021665dc56d9b2efceac05497ee84a7a29"} Dec 06 03:59:23 crc kubenswrapper[4802]: I1206 03:59:23.698051 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"22294441-50c6-4dc0-80f1-4b14c1fc62a9","Type":"ContainerStarted","Data":"90662e183b0e95670b64555522541d9155b259f309bcf2d1939c8051e4b9fb83"} Dec 06 03:59:23 crc kubenswrapper[4802]: I1206 03:59:23.700517 4802 generic.go:334] "Generic (PLEG): container finished" podID="a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c" containerID="970fda877b59771db20a7752f0768084b73e8e6c9cd7b12537a9dc08c89f9e30" exitCode=0 Dec 06 03:59:23 crc kubenswrapper[4802]: I1206 03:59:23.700568 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c67bffd47-mrlr7" event={"ID":"a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c","Type":"ContainerDied","Data":"970fda877b59771db20a7752f0768084b73e8e6c9cd7b12537a9dc08c89f9e30"} Dec 06 03:59:23 crc kubenswrapper[4802]: I1206 03:59:23.702951 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-788d784b-hjc7w" event={"ID":"5800f6d2-346a-4e9a-8585-e839ddfd035f","Type":"ContainerStarted","Data":"7619ca45f193ebe432b3e5f8ea05aab302a6f92eb21ab3a51306fa0bcbb29be7"} Dec 06 03:59:23 crc kubenswrapper[4802]: I1206 03:59:23.702993 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 06 03:59:23 crc kubenswrapper[4802]: I1206 03:59:23.703459 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 06 03:59:23 crc kubenswrapper[4802]: I1206 03:59:23.703538 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-788d784b-hjc7w" Dec 06 03:59:23 crc kubenswrapper[4802]: I1206 03:59:23.703557 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-788d784b-hjc7w" Dec 06 03:59:23 crc kubenswrapper[4802]: I1206 03:59:23.730120 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.73009453 podStartE2EDuration="5.73009453s" podCreationTimestamp="2025-12-06 03:59:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:59:23.720983112 +0000 UTC m=+1156.592892274" watchObservedRunningTime="2025-12-06 03:59:23.73009453 +0000 UTC m=+1156.602003682" Dec 06 03:59:23 crc kubenswrapper[4802]: I1206 03:59:23.779694 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-788d784b-hjc7w" podStartSLOduration=4.779659089 podStartE2EDuration="4.779659089s" podCreationTimestamp="2025-12-06 03:59:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:59:23.755186403 +0000 UTC m=+1156.627095565" watchObservedRunningTime="2025-12-06 03:59:23.779659089 +0000 UTC m=+1156.651568241" Dec 06 03:59:24 crc kubenswrapper[4802]: I1206 03:59:24.122264 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-7f79b98d58-l9qc5"] Dec 06 03:59:24 crc kubenswrapper[4802]: I1206 03:59:24.123987 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7f79b98d58-l9qc5" Dec 06 03:59:24 crc kubenswrapper[4802]: I1206 03:59:24.126298 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 06 03:59:24 crc kubenswrapper[4802]: I1206 03:59:24.126551 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 06 03:59:24 crc kubenswrapper[4802]: I1206 03:59:24.144688 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7f79b98d58-l9qc5"] Dec 06 03:59:24 crc kubenswrapper[4802]: I1206 03:59:24.276475 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmn7r\" (UniqueName: \"kubernetes.io/projected/21e91f8f-b3aa-4a45-85e6-5192a3f85a04-kube-api-access-kmn7r\") pod \"barbican-api-7f79b98d58-l9qc5\" (UID: \"21e91f8f-b3aa-4a45-85e6-5192a3f85a04\") " pod="openstack/barbican-api-7f79b98d58-l9qc5" Dec 06 03:59:24 crc kubenswrapper[4802]: I1206 03:59:24.276692 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21e91f8f-b3aa-4a45-85e6-5192a3f85a04-config-data\") pod \"barbican-api-7f79b98d58-l9qc5\" (UID: \"21e91f8f-b3aa-4a45-85e6-5192a3f85a04\") " pod="openstack/barbican-api-7f79b98d58-l9qc5" Dec 06 03:59:24 crc kubenswrapper[4802]: I1206 03:59:24.276920 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/21e91f8f-b3aa-4a45-85e6-5192a3f85a04-logs\") pod \"barbican-api-7f79b98d58-l9qc5\" (UID: \"21e91f8f-b3aa-4a45-85e6-5192a3f85a04\") " pod="openstack/barbican-api-7f79b98d58-l9qc5" Dec 06 03:59:24 crc kubenswrapper[4802]: I1206 03:59:24.277023 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21e91f8f-b3aa-4a45-85e6-5192a3f85a04-combined-ca-bundle\") pod \"barbican-api-7f79b98d58-l9qc5\" (UID: \"21e91f8f-b3aa-4a45-85e6-5192a3f85a04\") " pod="openstack/barbican-api-7f79b98d58-l9qc5" Dec 06 03:59:24 crc kubenswrapper[4802]: I1206 03:59:24.277063 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/21e91f8f-b3aa-4a45-85e6-5192a3f85a04-config-data-custom\") pod \"barbican-api-7f79b98d58-l9qc5\" (UID: \"21e91f8f-b3aa-4a45-85e6-5192a3f85a04\") " pod="openstack/barbican-api-7f79b98d58-l9qc5" Dec 06 03:59:24 crc kubenswrapper[4802]: I1206 03:59:24.277095 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/21e91f8f-b3aa-4a45-85e6-5192a3f85a04-public-tls-certs\") pod \"barbican-api-7f79b98d58-l9qc5\" (UID: \"21e91f8f-b3aa-4a45-85e6-5192a3f85a04\") " pod="openstack/barbican-api-7f79b98d58-l9qc5" Dec 06 03:59:24 crc kubenswrapper[4802]: I1206 03:59:24.277133 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/21e91f8f-b3aa-4a45-85e6-5192a3f85a04-internal-tls-certs\") pod \"barbican-api-7f79b98d58-l9qc5\" (UID: \"21e91f8f-b3aa-4a45-85e6-5192a3f85a04\") " pod="openstack/barbican-api-7f79b98d58-l9qc5" Dec 06 03:59:24 crc kubenswrapper[4802]: I1206 03:59:24.379239 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21e91f8f-b3aa-4a45-85e6-5192a3f85a04-config-data\") pod \"barbican-api-7f79b98d58-l9qc5\" (UID: \"21e91f8f-b3aa-4a45-85e6-5192a3f85a04\") " pod="openstack/barbican-api-7f79b98d58-l9qc5" Dec 06 03:59:24 crc kubenswrapper[4802]: I1206 03:59:24.379298 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/21e91f8f-b3aa-4a45-85e6-5192a3f85a04-logs\") pod \"barbican-api-7f79b98d58-l9qc5\" (UID: \"21e91f8f-b3aa-4a45-85e6-5192a3f85a04\") " pod="openstack/barbican-api-7f79b98d58-l9qc5" Dec 06 03:59:24 crc kubenswrapper[4802]: I1206 03:59:24.379327 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21e91f8f-b3aa-4a45-85e6-5192a3f85a04-combined-ca-bundle\") pod \"barbican-api-7f79b98d58-l9qc5\" (UID: \"21e91f8f-b3aa-4a45-85e6-5192a3f85a04\") " pod="openstack/barbican-api-7f79b98d58-l9qc5" Dec 06 03:59:24 crc kubenswrapper[4802]: I1206 03:59:24.379343 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/21e91f8f-b3aa-4a45-85e6-5192a3f85a04-config-data-custom\") pod \"barbican-api-7f79b98d58-l9qc5\" (UID: \"21e91f8f-b3aa-4a45-85e6-5192a3f85a04\") " pod="openstack/barbican-api-7f79b98d58-l9qc5" Dec 06 03:59:24 crc kubenswrapper[4802]: I1206 03:59:24.379361 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/21e91f8f-b3aa-4a45-85e6-5192a3f85a04-public-tls-certs\") pod \"barbican-api-7f79b98d58-l9qc5\" (UID: \"21e91f8f-b3aa-4a45-85e6-5192a3f85a04\") " pod="openstack/barbican-api-7f79b98d58-l9qc5" Dec 06 03:59:24 crc kubenswrapper[4802]: I1206 03:59:24.379380 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/21e91f8f-b3aa-4a45-85e6-5192a3f85a04-internal-tls-certs\") pod \"barbican-api-7f79b98d58-l9qc5\" (UID: \"21e91f8f-b3aa-4a45-85e6-5192a3f85a04\") " pod="openstack/barbican-api-7f79b98d58-l9qc5" Dec 06 03:59:24 crc kubenswrapper[4802]: I1206 03:59:24.379423 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmn7r\" (UniqueName: \"kubernetes.io/projected/21e91f8f-b3aa-4a45-85e6-5192a3f85a04-kube-api-access-kmn7r\") pod \"barbican-api-7f79b98d58-l9qc5\" (UID: \"21e91f8f-b3aa-4a45-85e6-5192a3f85a04\") " pod="openstack/barbican-api-7f79b98d58-l9qc5" Dec 06 03:59:24 crc kubenswrapper[4802]: I1206 03:59:24.379688 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/21e91f8f-b3aa-4a45-85e6-5192a3f85a04-logs\") pod \"barbican-api-7f79b98d58-l9qc5\" (UID: \"21e91f8f-b3aa-4a45-85e6-5192a3f85a04\") " pod="openstack/barbican-api-7f79b98d58-l9qc5" Dec 06 03:59:24 crc kubenswrapper[4802]: I1206 03:59:24.385460 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21e91f8f-b3aa-4a45-85e6-5192a3f85a04-combined-ca-bundle\") pod \"barbican-api-7f79b98d58-l9qc5\" (UID: \"21e91f8f-b3aa-4a45-85e6-5192a3f85a04\") " pod="openstack/barbican-api-7f79b98d58-l9qc5" Dec 06 03:59:24 crc kubenswrapper[4802]: I1206 03:59:24.385937 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/21e91f8f-b3aa-4a45-85e6-5192a3f85a04-config-data-custom\") pod \"barbican-api-7f79b98d58-l9qc5\" (UID: \"21e91f8f-b3aa-4a45-85e6-5192a3f85a04\") " pod="openstack/barbican-api-7f79b98d58-l9qc5" Dec 06 03:59:24 crc kubenswrapper[4802]: I1206 03:59:24.388124 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/21e91f8f-b3aa-4a45-85e6-5192a3f85a04-internal-tls-certs\") pod \"barbican-api-7f79b98d58-l9qc5\" (UID: \"21e91f8f-b3aa-4a45-85e6-5192a3f85a04\") " pod="openstack/barbican-api-7f79b98d58-l9qc5" Dec 06 03:59:24 crc kubenswrapper[4802]: I1206 03:59:24.389005 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21e91f8f-b3aa-4a45-85e6-5192a3f85a04-config-data\") pod \"barbican-api-7f79b98d58-l9qc5\" (UID: \"21e91f8f-b3aa-4a45-85e6-5192a3f85a04\") " pod="openstack/barbican-api-7f79b98d58-l9qc5" Dec 06 03:59:24 crc kubenswrapper[4802]: I1206 03:59:24.392350 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/21e91f8f-b3aa-4a45-85e6-5192a3f85a04-public-tls-certs\") pod \"barbican-api-7f79b98d58-l9qc5\" (UID: \"21e91f8f-b3aa-4a45-85e6-5192a3f85a04\") " pod="openstack/barbican-api-7f79b98d58-l9qc5" Dec 06 03:59:24 crc kubenswrapper[4802]: I1206 03:59:24.404468 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmn7r\" (UniqueName: \"kubernetes.io/projected/21e91f8f-b3aa-4a45-85e6-5192a3f85a04-kube-api-access-kmn7r\") pod \"barbican-api-7f79b98d58-l9qc5\" (UID: \"21e91f8f-b3aa-4a45-85e6-5192a3f85a04\") " pod="openstack/barbican-api-7f79b98d58-l9qc5" Dec 06 03:59:24 crc kubenswrapper[4802]: I1206 03:59:24.458110 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7f79b98d58-l9qc5" Dec 06 03:59:25 crc kubenswrapper[4802]: I1206 03:59:25.971138 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 06 03:59:25 crc kubenswrapper[4802]: I1206 03:59:25.971511 4802 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 03:59:25 crc kubenswrapper[4802]: I1206 03:59:25.975386 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 06 03:59:27 crc kubenswrapper[4802]: I1206 03:59:27.737354 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6f6b9c98b6-8ndrw" event={"ID":"d098e8a2-f622-44f7-997f-51c6ff7eceaa","Type":"ContainerStarted","Data":"748b82913c72c4ebf47e1088833f5e5da1132a05f680ef7cb0ff05f10fe522c5"} Dec 06 03:59:27 crc kubenswrapper[4802]: I1206 03:59:27.737956 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6f6b9c98b6-8ndrw" Dec 06 03:59:27 crc kubenswrapper[4802]: I1206 03:59:27.765863 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6f6b9c98b6-8ndrw" podStartSLOduration=7.765840797 podStartE2EDuration="7.765840797s" podCreationTimestamp="2025-12-06 03:59:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:59:27.759268857 +0000 UTC m=+1160.631178009" watchObservedRunningTime="2025-12-06 03:59:27.765840797 +0000 UTC m=+1160.637749949" Dec 06 03:59:28 crc kubenswrapper[4802]: I1206 03:59:28.752432 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6f6b9c98b6-8ndrw" Dec 06 03:59:29 crc kubenswrapper[4802]: I1206 03:59:29.054715 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 06 03:59:29 crc kubenswrapper[4802]: I1206 03:59:29.055230 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 06 03:59:29 crc kubenswrapper[4802]: I1206 03:59:29.119126 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 06 03:59:29 crc kubenswrapper[4802]: I1206 03:59:29.122239 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 06 03:59:29 crc kubenswrapper[4802]: I1206 03:59:29.768458 4802 generic.go:334] "Generic (PLEG): container finished" podID="8c25c327-e3cd-4f28-855a-249e4bb0c5f6" containerID="22256b01fbd251cd8845a65d030b3e708b70e6bb74755d572798823340fae615" exitCode=0 Dec 06 03:59:29 crc kubenswrapper[4802]: I1206 03:59:29.768570 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-qs8hv" event={"ID":"8c25c327-e3cd-4f28-855a-249e4bb0c5f6","Type":"ContainerDied","Data":"22256b01fbd251cd8845a65d030b3e708b70e6bb74755d572798823340fae615"} Dec 06 03:59:29 crc kubenswrapper[4802]: I1206 03:59:29.771296 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 06 03:59:29 crc kubenswrapper[4802]: I1206 03:59:29.771323 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 06 03:59:29 crc kubenswrapper[4802]: I1206 03:59:29.792570 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7f79b98d58-l9qc5"] Dec 06 03:59:30 crc kubenswrapper[4802]: W1206 03:59:30.236527 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod21e91f8f_b3aa_4a45_85e6_5192a3f85a04.slice/crio-f52e3cbdc50f5a2b42846545cb35e16675bc96fa3566d0d4b16ec6bd4e202ed8 WatchSource:0}: Error finding container f52e3cbdc50f5a2b42846545cb35e16675bc96fa3566d0d4b16ec6bd4e202ed8: Status 404 returned error can't find the container with id f52e3cbdc50f5a2b42846545cb35e16675bc96fa3566d0d4b16ec6bd4e202ed8 Dec 06 03:59:30 crc kubenswrapper[4802]: I1206 03:59:30.264357 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6f6b9c98b6-8ndrw" Dec 06 03:59:30 crc kubenswrapper[4802]: E1206 03:59:30.720048 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="4d4f4fff-a6ec-4fef-82b4-6ba392fd297c" Dec 06 03:59:30 crc kubenswrapper[4802]: I1206 03:59:30.803459 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-2hzlt" event={"ID":"2a639667-be04-429f-a2ce-b20a011425f8","Type":"ContainerStarted","Data":"285d42f44e31edd7beb22eac4c447d12107075be937e5c0e69d394dc90c0ac17"} Dec 06 03:59:30 crc kubenswrapper[4802]: I1206 03:59:30.811573 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4d4f4fff-a6ec-4fef-82b4-6ba392fd297c" containerName="ceilometer-notification-agent" containerID="cri-o://fa037e55af9d2c4256d307c1f7e958d5e921062fe281b0182ef8ec24b83b1e30" gracePeriod=30 Dec 06 03:59:30 crc kubenswrapper[4802]: I1206 03:59:30.811823 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4d4f4fff-a6ec-4fef-82b4-6ba392fd297c","Type":"ContainerStarted","Data":"ec333f217f947fe8e066593d0ae87d8fd22fa9027bc4f3bc00f5f9ae7a30b116"} Dec 06 03:59:30 crc kubenswrapper[4802]: I1206 03:59:30.811872 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4d4f4fff-a6ec-4fef-82b4-6ba392fd297c" containerName="proxy-httpd" containerID="cri-o://ec333f217f947fe8e066593d0ae87d8fd22fa9027bc4f3bc00f5f9ae7a30b116" gracePeriod=30 Dec 06 03:59:30 crc kubenswrapper[4802]: I1206 03:59:30.811892 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 06 03:59:30 crc kubenswrapper[4802]: I1206 03:59:30.811922 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4d4f4fff-a6ec-4fef-82b4-6ba392fd297c" containerName="sg-core" containerID="cri-o://fff16873781a4a16ca4ed5fa3a576de4e59fcca96bb97e804e8f0e742df14984" gracePeriod=30 Dec 06 03:59:30 crc kubenswrapper[4802]: I1206 03:59:30.819442 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-554f6fb55d-hdgjt" event={"ID":"b443b1b9-ec2c-499d-ae4d-1937c6f156f7","Type":"ContainerStarted","Data":"ff0b8f6758c95d9ef33a8f87aaa6f1055e2c24fccb2bd741669b2ecc4cb9677e"} Dec 06 03:59:30 crc kubenswrapper[4802]: I1206 03:59:30.827829 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-2hzlt" podStartSLOduration=2.266540669 podStartE2EDuration="52.827805776s" podCreationTimestamp="2025-12-06 03:58:38 +0000 UTC" firstStartedPulling="2025-12-06 03:58:39.740711446 +0000 UTC m=+1112.612620598" lastFinishedPulling="2025-12-06 03:59:30.301976553 +0000 UTC m=+1163.173885705" observedRunningTime="2025-12-06 03:59:30.820051945 +0000 UTC m=+1163.691961097" watchObservedRunningTime="2025-12-06 03:59:30.827805776 +0000 UTC m=+1163.699714928" Dec 06 03:59:30 crc kubenswrapper[4802]: I1206 03:59:30.831643 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c67bffd47-mrlr7" event={"ID":"a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c","Type":"ContainerStarted","Data":"5504caf2de012952cdbf63c5da40cb08607e111ca2ac64a80b98bbe613d99954"} Dec 06 03:59:30 crc kubenswrapper[4802]: I1206 03:59:30.831976 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7c67bffd47-mrlr7" Dec 06 03:59:30 crc kubenswrapper[4802]: I1206 03:59:30.833979 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7f79b98d58-l9qc5" event={"ID":"21e91f8f-b3aa-4a45-85e6-5192a3f85a04","Type":"ContainerStarted","Data":"ce8590c91cb4919940231805f5a80a5ded9c9f17dd31ff538fc8b71f6da62d07"} Dec 06 03:59:30 crc kubenswrapper[4802]: I1206 03:59:30.834007 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7f79b98d58-l9qc5" event={"ID":"21e91f8f-b3aa-4a45-85e6-5192a3f85a04","Type":"ContainerStarted","Data":"f52e3cbdc50f5a2b42846545cb35e16675bc96fa3566d0d4b16ec6bd4e202ed8"} Dec 06 03:59:30 crc kubenswrapper[4802]: I1206 03:59:30.837847 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-68998b57f5-6qgkl" event={"ID":"57a20a46-fdfb-410f-8a13-0b58fa5751da","Type":"ContainerStarted","Data":"19b7e892d5e45cd2af4559eb619dce50f171a90f2b6550dbd9ae3a56a42b3246"} Dec 06 03:59:30 crc kubenswrapper[4802]: I1206 03:59:30.869501 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7c67bffd47-mrlr7" podStartSLOduration=10.869486041 podStartE2EDuration="10.869486041s" podCreationTimestamp="2025-12-06 03:59:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:59:30.864040033 +0000 UTC m=+1163.735949185" watchObservedRunningTime="2025-12-06 03:59:30.869486041 +0000 UTC m=+1163.741395193" Dec 06 03:59:31 crc kubenswrapper[4802]: I1206 03:59:31.207782 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-qs8hv" Dec 06 03:59:31 crc kubenswrapper[4802]: I1206 03:59:31.336468 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c25c327-e3cd-4f28-855a-249e4bb0c5f6-combined-ca-bundle\") pod \"8c25c327-e3cd-4f28-855a-249e4bb0c5f6\" (UID: \"8c25c327-e3cd-4f28-855a-249e4bb0c5f6\") " Dec 06 03:59:31 crc kubenswrapper[4802]: I1206 03:59:31.336530 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8c25c327-e3cd-4f28-855a-249e4bb0c5f6-config\") pod \"8c25c327-e3cd-4f28-855a-249e4bb0c5f6\" (UID: \"8c25c327-e3cd-4f28-855a-249e4bb0c5f6\") " Dec 06 03:59:31 crc kubenswrapper[4802]: I1206 03:59:31.336784 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lt8jq\" (UniqueName: \"kubernetes.io/projected/8c25c327-e3cd-4f28-855a-249e4bb0c5f6-kube-api-access-lt8jq\") pod \"8c25c327-e3cd-4f28-855a-249e4bb0c5f6\" (UID: \"8c25c327-e3cd-4f28-855a-249e4bb0c5f6\") " Dec 06 03:59:31 crc kubenswrapper[4802]: I1206 03:59:31.343987 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c25c327-e3cd-4f28-855a-249e4bb0c5f6-kube-api-access-lt8jq" (OuterVolumeSpecName: "kube-api-access-lt8jq") pod "8c25c327-e3cd-4f28-855a-249e4bb0c5f6" (UID: "8c25c327-e3cd-4f28-855a-249e4bb0c5f6"). InnerVolumeSpecName "kube-api-access-lt8jq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:59:31 crc kubenswrapper[4802]: I1206 03:59:31.387206 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c25c327-e3cd-4f28-855a-249e4bb0c5f6-config" (OuterVolumeSpecName: "config") pod "8c25c327-e3cd-4f28-855a-249e4bb0c5f6" (UID: "8c25c327-e3cd-4f28-855a-249e4bb0c5f6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:59:31 crc kubenswrapper[4802]: I1206 03:59:31.425875 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c25c327-e3cd-4f28-855a-249e4bb0c5f6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8c25c327-e3cd-4f28-855a-249e4bb0c5f6" (UID: "8c25c327-e3cd-4f28-855a-249e4bb0c5f6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:59:31 crc kubenswrapper[4802]: I1206 03:59:31.439587 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lt8jq\" (UniqueName: \"kubernetes.io/projected/8c25c327-e3cd-4f28-855a-249e4bb0c5f6-kube-api-access-lt8jq\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:31 crc kubenswrapper[4802]: I1206 03:59:31.439634 4802 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c25c327-e3cd-4f28-855a-249e4bb0c5f6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:31 crc kubenswrapper[4802]: I1206 03:59:31.439644 4802 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/8c25c327-e3cd-4f28-855a-249e4bb0c5f6-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:31 crc kubenswrapper[4802]: I1206 03:59:31.847264 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7f79b98d58-l9qc5" event={"ID":"21e91f8f-b3aa-4a45-85e6-5192a3f85a04","Type":"ContainerStarted","Data":"b1916383cdd72584bba2be3194c5d44df09b16bec29b6d75071aa3cde9d93038"} Dec 06 03:59:31 crc kubenswrapper[4802]: I1206 03:59:31.847468 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7f79b98d58-l9qc5" Dec 06 03:59:31 crc kubenswrapper[4802]: I1206 03:59:31.849736 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-68998b57f5-6qgkl" event={"ID":"57a20a46-fdfb-410f-8a13-0b58fa5751da","Type":"ContainerStarted","Data":"4366a0248b6e74cd7cc08f528ecd12293c5bf9eb44560471f183b50ebdf0fee6"} Dec 06 03:59:31 crc kubenswrapper[4802]: I1206 03:59:31.851384 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-qs8hv" Dec 06 03:59:31 crc kubenswrapper[4802]: I1206 03:59:31.851550 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-qs8hv" event={"ID":"8c25c327-e3cd-4f28-855a-249e4bb0c5f6","Type":"ContainerDied","Data":"acbf648bc77efc79c07ddf28a963e7543f0acb21850d33865a63152a8d883989"} Dec 06 03:59:31 crc kubenswrapper[4802]: I1206 03:59:31.851587 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="acbf648bc77efc79c07ddf28a963e7543f0acb21850d33865a63152a8d883989" Dec 06 03:59:31 crc kubenswrapper[4802]: I1206 03:59:31.853628 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-5hj5p" event={"ID":"f1644907-b4b1-4e23-a273-19d535bd7f88","Type":"ContainerStarted","Data":"66f602546227aa86f52ceb621c4f145c78c1b12628e754ab524c6e32f88a0151"} Dec 06 03:59:31 crc kubenswrapper[4802]: I1206 03:59:31.856204 4802 generic.go:334] "Generic (PLEG): container finished" podID="4d4f4fff-a6ec-4fef-82b4-6ba392fd297c" containerID="fff16873781a4a16ca4ed5fa3a576de4e59fcca96bb97e804e8f0e742df14984" exitCode=2 Dec 06 03:59:31 crc kubenswrapper[4802]: I1206 03:59:31.856246 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4d4f4fff-a6ec-4fef-82b4-6ba392fd297c","Type":"ContainerDied","Data":"fff16873781a4a16ca4ed5fa3a576de4e59fcca96bb97e804e8f0e742df14984"} Dec 06 03:59:31 crc kubenswrapper[4802]: I1206 03:59:31.857804 4802 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 03:59:31 crc kubenswrapper[4802]: I1206 03:59:31.857831 4802 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 03:59:31 crc kubenswrapper[4802]: I1206 03:59:31.858865 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-554f6fb55d-hdgjt" event={"ID":"b443b1b9-ec2c-499d-ae4d-1937c6f156f7","Type":"ContainerStarted","Data":"66323b5d8c7be6149c2f3f876d321e869b82dddbe7ce68bb775819f08b6c21d1"} Dec 06 03:59:31 crc kubenswrapper[4802]: I1206 03:59:31.918165 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-7f79b98d58-l9qc5" podStartSLOduration=7.918144406 podStartE2EDuration="7.918144406s" podCreationTimestamp="2025-12-06 03:59:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:59:31.877715656 +0000 UTC m=+1164.749624818" watchObservedRunningTime="2025-12-06 03:59:31.918144406 +0000 UTC m=+1164.790053558" Dec 06 03:59:31 crc kubenswrapper[4802]: I1206 03:59:31.923197 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-5hj5p" podStartSLOduration=3.81206374 podStartE2EDuration="53.923174683s" podCreationTimestamp="2025-12-06 03:58:38 +0000 UTC" firstStartedPulling="2025-12-06 03:58:40.198301412 +0000 UTC m=+1113.070210564" lastFinishedPulling="2025-12-06 03:59:30.309412355 +0000 UTC m=+1163.181321507" observedRunningTime="2025-12-06 03:59:31.909066419 +0000 UTC m=+1164.780975571" watchObservedRunningTime="2025-12-06 03:59:31.923174683 +0000 UTC m=+1164.795083835" Dec 06 03:59:31 crc kubenswrapper[4802]: I1206 03:59:31.965558 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-554f6fb55d-hdgjt" podStartSLOduration=3.392225313 podStartE2EDuration="11.965532776s" podCreationTimestamp="2025-12-06 03:59:20 +0000 UTC" firstStartedPulling="2025-12-06 03:59:21.75733665 +0000 UTC m=+1154.629245802" lastFinishedPulling="2025-12-06 03:59:30.330644113 +0000 UTC m=+1163.202553265" observedRunningTime="2025-12-06 03:59:31.9326227 +0000 UTC m=+1164.804531852" watchObservedRunningTime="2025-12-06 03:59:31.965532776 +0000 UTC m=+1164.837441928" Dec 06 03:59:31 crc kubenswrapper[4802]: I1206 03:59:31.969387 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-68998b57f5-6qgkl" podStartSLOduration=3.32966455 podStartE2EDuration="11.9693644s" podCreationTimestamp="2025-12-06 03:59:20 +0000 UTC" firstStartedPulling="2025-12-06 03:59:21.6710049 +0000 UTC m=+1154.542914052" lastFinishedPulling="2025-12-06 03:59:30.31070474 +0000 UTC m=+1163.182613902" observedRunningTime="2025-12-06 03:59:31.960733795 +0000 UTC m=+1164.832642957" watchObservedRunningTime="2025-12-06 03:59:31.9693644 +0000 UTC m=+1164.841273562" Dec 06 03:59:32 crc kubenswrapper[4802]: I1206 03:59:32.157077 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c67bffd47-mrlr7"] Dec 06 03:59:32 crc kubenswrapper[4802]: I1206 03:59:32.259888 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-bq9w6"] Dec 06 03:59:32 crc kubenswrapper[4802]: E1206 03:59:32.260364 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c25c327-e3cd-4f28-855a-249e4bb0c5f6" containerName="neutron-db-sync" Dec 06 03:59:32 crc kubenswrapper[4802]: I1206 03:59:32.260384 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c25c327-e3cd-4f28-855a-249e4bb0c5f6" containerName="neutron-db-sync" Dec 06 03:59:32 crc kubenswrapper[4802]: I1206 03:59:32.260606 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c25c327-e3cd-4f28-855a-249e4bb0c5f6" containerName="neutron-db-sync" Dec 06 03:59:32 crc kubenswrapper[4802]: I1206 03:59:32.261712 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-bq9w6" Dec 06 03:59:32 crc kubenswrapper[4802]: I1206 03:59:32.262431 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 06 03:59:32 crc kubenswrapper[4802]: I1206 03:59:32.298191 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-bq9w6"] Dec 06 03:59:32 crc kubenswrapper[4802]: I1206 03:59:32.363422 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/11fd5eb5-07ce-4218-9f9e-50386ba1c4c4-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-bq9w6\" (UID: \"11fd5eb5-07ce-4218-9f9e-50386ba1c4c4\") " pod="openstack/dnsmasq-dns-848cf88cfc-bq9w6" Dec 06 03:59:32 crc kubenswrapper[4802]: I1206 03:59:32.363479 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11fd5eb5-07ce-4218-9f9e-50386ba1c4c4-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-bq9w6\" (UID: \"11fd5eb5-07ce-4218-9f9e-50386ba1c4c4\") " pod="openstack/dnsmasq-dns-848cf88cfc-bq9w6" Dec 06 03:59:32 crc kubenswrapper[4802]: I1206 03:59:32.363510 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/11fd5eb5-07ce-4218-9f9e-50386ba1c4c4-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-bq9w6\" (UID: \"11fd5eb5-07ce-4218-9f9e-50386ba1c4c4\") " pod="openstack/dnsmasq-dns-848cf88cfc-bq9w6" Dec 06 03:59:32 crc kubenswrapper[4802]: I1206 03:59:32.363570 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqcwv\" (UniqueName: \"kubernetes.io/projected/11fd5eb5-07ce-4218-9f9e-50386ba1c4c4-kube-api-access-xqcwv\") pod \"dnsmasq-dns-848cf88cfc-bq9w6\" (UID: \"11fd5eb5-07ce-4218-9f9e-50386ba1c4c4\") " pod="openstack/dnsmasq-dns-848cf88cfc-bq9w6" Dec 06 03:59:32 crc kubenswrapper[4802]: I1206 03:59:32.363652 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/11fd5eb5-07ce-4218-9f9e-50386ba1c4c4-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-bq9w6\" (UID: \"11fd5eb5-07ce-4218-9f9e-50386ba1c4c4\") " pod="openstack/dnsmasq-dns-848cf88cfc-bq9w6" Dec 06 03:59:32 crc kubenswrapper[4802]: I1206 03:59:32.363678 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11fd5eb5-07ce-4218-9f9e-50386ba1c4c4-config\") pod \"dnsmasq-dns-848cf88cfc-bq9w6\" (UID: \"11fd5eb5-07ce-4218-9f9e-50386ba1c4c4\") " pod="openstack/dnsmasq-dns-848cf88cfc-bq9w6" Dec 06 03:59:32 crc kubenswrapper[4802]: I1206 03:59:32.413495 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-844df74668-r698z"] Dec 06 03:59:32 crc kubenswrapper[4802]: I1206 03:59:32.415324 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-844df74668-r698z" Dec 06 03:59:32 crc kubenswrapper[4802]: I1206 03:59:32.422196 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 06 03:59:32 crc kubenswrapper[4802]: I1206 03:59:32.422552 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 06 03:59:32 crc kubenswrapper[4802]: I1206 03:59:32.422681 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 06 03:59:32 crc kubenswrapper[4802]: I1206 03:59:32.434832 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-844df74668-r698z"] Dec 06 03:59:32 crc kubenswrapper[4802]: I1206 03:59:32.443267 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-nr9zb" Dec 06 03:59:32 crc kubenswrapper[4802]: I1206 03:59:32.473566 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b364aef0-231f-472a-a8c9-51a5199c0997-httpd-config\") pod \"neutron-844df74668-r698z\" (UID: \"b364aef0-231f-472a-a8c9-51a5199c0997\") " pod="openstack/neutron-844df74668-r698z" Dec 06 03:59:32 crc kubenswrapper[4802]: I1206 03:59:32.473649 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/11fd5eb5-07ce-4218-9f9e-50386ba1c4c4-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-bq9w6\" (UID: \"11fd5eb5-07ce-4218-9f9e-50386ba1c4c4\") " pod="openstack/dnsmasq-dns-848cf88cfc-bq9w6" Dec 06 03:59:32 crc kubenswrapper[4802]: I1206 03:59:32.473682 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2599\" (UniqueName: \"kubernetes.io/projected/b364aef0-231f-472a-a8c9-51a5199c0997-kube-api-access-d2599\") pod \"neutron-844df74668-r698z\" (UID: \"b364aef0-231f-472a-a8c9-51a5199c0997\") " pod="openstack/neutron-844df74668-r698z" Dec 06 03:59:32 crc kubenswrapper[4802]: I1206 03:59:32.473710 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11fd5eb5-07ce-4218-9f9e-50386ba1c4c4-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-bq9w6\" (UID: \"11fd5eb5-07ce-4218-9f9e-50386ba1c4c4\") " pod="openstack/dnsmasq-dns-848cf88cfc-bq9w6" Dec 06 03:59:32 crc kubenswrapper[4802]: I1206 03:59:32.473743 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/11fd5eb5-07ce-4218-9f9e-50386ba1c4c4-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-bq9w6\" (UID: \"11fd5eb5-07ce-4218-9f9e-50386ba1c4c4\") " pod="openstack/dnsmasq-dns-848cf88cfc-bq9w6" Dec 06 03:59:32 crc kubenswrapper[4802]: I1206 03:59:32.473826 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b364aef0-231f-472a-a8c9-51a5199c0997-combined-ca-bundle\") pod \"neutron-844df74668-r698z\" (UID: \"b364aef0-231f-472a-a8c9-51a5199c0997\") " pod="openstack/neutron-844df74668-r698z" Dec 06 03:59:32 crc kubenswrapper[4802]: I1206 03:59:32.473852 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqcwv\" (UniqueName: \"kubernetes.io/projected/11fd5eb5-07ce-4218-9f9e-50386ba1c4c4-kube-api-access-xqcwv\") pod \"dnsmasq-dns-848cf88cfc-bq9w6\" (UID: \"11fd5eb5-07ce-4218-9f9e-50386ba1c4c4\") " pod="openstack/dnsmasq-dns-848cf88cfc-bq9w6" Dec 06 03:59:32 crc kubenswrapper[4802]: I1206 03:59:32.473896 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b364aef0-231f-472a-a8c9-51a5199c0997-config\") pod \"neutron-844df74668-r698z\" (UID: \"b364aef0-231f-472a-a8c9-51a5199c0997\") " pod="openstack/neutron-844df74668-r698z" Dec 06 03:59:32 crc kubenswrapper[4802]: I1206 03:59:32.473916 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b364aef0-231f-472a-a8c9-51a5199c0997-ovndb-tls-certs\") pod \"neutron-844df74668-r698z\" (UID: \"b364aef0-231f-472a-a8c9-51a5199c0997\") " pod="openstack/neutron-844df74668-r698z" Dec 06 03:59:32 crc kubenswrapper[4802]: I1206 03:59:32.473980 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/11fd5eb5-07ce-4218-9f9e-50386ba1c4c4-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-bq9w6\" (UID: \"11fd5eb5-07ce-4218-9f9e-50386ba1c4c4\") " pod="openstack/dnsmasq-dns-848cf88cfc-bq9w6" Dec 06 03:59:32 crc kubenswrapper[4802]: I1206 03:59:32.474013 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11fd5eb5-07ce-4218-9f9e-50386ba1c4c4-config\") pod \"dnsmasq-dns-848cf88cfc-bq9w6\" (UID: \"11fd5eb5-07ce-4218-9f9e-50386ba1c4c4\") " pod="openstack/dnsmasq-dns-848cf88cfc-bq9w6" Dec 06 03:59:32 crc kubenswrapper[4802]: I1206 03:59:32.475030 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11fd5eb5-07ce-4218-9f9e-50386ba1c4c4-config\") pod \"dnsmasq-dns-848cf88cfc-bq9w6\" (UID: \"11fd5eb5-07ce-4218-9f9e-50386ba1c4c4\") " pod="openstack/dnsmasq-dns-848cf88cfc-bq9w6" Dec 06 03:59:32 crc kubenswrapper[4802]: I1206 03:59:32.475663 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/11fd5eb5-07ce-4218-9f9e-50386ba1c4c4-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-bq9w6\" (UID: \"11fd5eb5-07ce-4218-9f9e-50386ba1c4c4\") " pod="openstack/dnsmasq-dns-848cf88cfc-bq9w6" Dec 06 03:59:32 crc kubenswrapper[4802]: I1206 03:59:32.476458 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/11fd5eb5-07ce-4218-9f9e-50386ba1c4c4-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-bq9w6\" (UID: \"11fd5eb5-07ce-4218-9f9e-50386ba1c4c4\") " pod="openstack/dnsmasq-dns-848cf88cfc-bq9w6" Dec 06 03:59:32 crc kubenswrapper[4802]: I1206 03:59:32.481027 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11fd5eb5-07ce-4218-9f9e-50386ba1c4c4-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-bq9w6\" (UID: \"11fd5eb5-07ce-4218-9f9e-50386ba1c4c4\") " pod="openstack/dnsmasq-dns-848cf88cfc-bq9w6" Dec 06 03:59:32 crc kubenswrapper[4802]: I1206 03:59:32.481352 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/11fd5eb5-07ce-4218-9f9e-50386ba1c4c4-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-bq9w6\" (UID: \"11fd5eb5-07ce-4218-9f9e-50386ba1c4c4\") " pod="openstack/dnsmasq-dns-848cf88cfc-bq9w6" Dec 06 03:59:32 crc kubenswrapper[4802]: I1206 03:59:32.509079 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqcwv\" (UniqueName: \"kubernetes.io/projected/11fd5eb5-07ce-4218-9f9e-50386ba1c4c4-kube-api-access-xqcwv\") pod \"dnsmasq-dns-848cf88cfc-bq9w6\" (UID: \"11fd5eb5-07ce-4218-9f9e-50386ba1c4c4\") " pod="openstack/dnsmasq-dns-848cf88cfc-bq9w6" Dec 06 03:59:32 crc kubenswrapper[4802]: I1206 03:59:32.575249 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b364aef0-231f-472a-a8c9-51a5199c0997-combined-ca-bundle\") pod \"neutron-844df74668-r698z\" (UID: \"b364aef0-231f-472a-a8c9-51a5199c0997\") " pod="openstack/neutron-844df74668-r698z" Dec 06 03:59:32 crc kubenswrapper[4802]: I1206 03:59:32.575345 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b364aef0-231f-472a-a8c9-51a5199c0997-config\") pod \"neutron-844df74668-r698z\" (UID: \"b364aef0-231f-472a-a8c9-51a5199c0997\") " pod="openstack/neutron-844df74668-r698z" Dec 06 03:59:32 crc kubenswrapper[4802]: I1206 03:59:32.575371 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b364aef0-231f-472a-a8c9-51a5199c0997-ovndb-tls-certs\") pod \"neutron-844df74668-r698z\" (UID: \"b364aef0-231f-472a-a8c9-51a5199c0997\") " pod="openstack/neutron-844df74668-r698z" Dec 06 03:59:32 crc kubenswrapper[4802]: I1206 03:59:32.575475 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b364aef0-231f-472a-a8c9-51a5199c0997-httpd-config\") pod \"neutron-844df74668-r698z\" (UID: \"b364aef0-231f-472a-a8c9-51a5199c0997\") " pod="openstack/neutron-844df74668-r698z" Dec 06 03:59:32 crc kubenswrapper[4802]: I1206 03:59:32.575553 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2599\" (UniqueName: \"kubernetes.io/projected/b364aef0-231f-472a-a8c9-51a5199c0997-kube-api-access-d2599\") pod \"neutron-844df74668-r698z\" (UID: \"b364aef0-231f-472a-a8c9-51a5199c0997\") " pod="openstack/neutron-844df74668-r698z" Dec 06 03:59:32 crc kubenswrapper[4802]: I1206 03:59:32.581913 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/b364aef0-231f-472a-a8c9-51a5199c0997-config\") pod \"neutron-844df74668-r698z\" (UID: \"b364aef0-231f-472a-a8c9-51a5199c0997\") " pod="openstack/neutron-844df74668-r698z" Dec 06 03:59:32 crc kubenswrapper[4802]: I1206 03:59:32.582593 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b364aef0-231f-472a-a8c9-51a5199c0997-ovndb-tls-certs\") pod \"neutron-844df74668-r698z\" (UID: \"b364aef0-231f-472a-a8c9-51a5199c0997\") " pod="openstack/neutron-844df74668-r698z" Dec 06 03:59:32 crc kubenswrapper[4802]: I1206 03:59:32.593577 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b364aef0-231f-472a-a8c9-51a5199c0997-httpd-config\") pod \"neutron-844df74668-r698z\" (UID: \"b364aef0-231f-472a-a8c9-51a5199c0997\") " pod="openstack/neutron-844df74668-r698z" Dec 06 03:59:32 crc kubenswrapper[4802]: I1206 03:59:32.596337 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b364aef0-231f-472a-a8c9-51a5199c0997-combined-ca-bundle\") pod \"neutron-844df74668-r698z\" (UID: \"b364aef0-231f-472a-a8c9-51a5199c0997\") " pod="openstack/neutron-844df74668-r698z" Dec 06 03:59:32 crc kubenswrapper[4802]: I1206 03:59:32.597865 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2599\" (UniqueName: \"kubernetes.io/projected/b364aef0-231f-472a-a8c9-51a5199c0997-kube-api-access-d2599\") pod \"neutron-844df74668-r698z\" (UID: \"b364aef0-231f-472a-a8c9-51a5199c0997\") " pod="openstack/neutron-844df74668-r698z" Dec 06 03:59:32 crc kubenswrapper[4802]: I1206 03:59:32.642135 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-bq9w6" Dec 06 03:59:32 crc kubenswrapper[4802]: I1206 03:59:32.761643 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-844df74668-r698z" Dec 06 03:59:32 crc kubenswrapper[4802]: I1206 03:59:32.940525 4802 generic.go:334] "Generic (PLEG): container finished" podID="4d4f4fff-a6ec-4fef-82b4-6ba392fd297c" containerID="fa037e55af9d2c4256d307c1f7e958d5e921062fe281b0182ef8ec24b83b1e30" exitCode=0 Dec 06 03:59:32 crc kubenswrapper[4802]: I1206 03:59:32.941543 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4d4f4fff-a6ec-4fef-82b4-6ba392fd297c","Type":"ContainerDied","Data":"fa037e55af9d2c4256d307c1f7e958d5e921062fe281b0182ef8ec24b83b1e30"} Dec 06 03:59:32 crc kubenswrapper[4802]: I1206 03:59:32.941604 4802 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 03:59:32 crc kubenswrapper[4802]: I1206 03:59:32.942489 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7c67bffd47-mrlr7" podUID="a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c" containerName="dnsmasq-dns" containerID="cri-o://5504caf2de012952cdbf63c5da40cb08607e111ca2ac64a80b98bbe613d99954" gracePeriod=10 Dec 06 03:59:32 crc kubenswrapper[4802]: I1206 03:59:32.943096 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7f79b98d58-l9qc5" Dec 06 03:59:33 crc kubenswrapper[4802]: I1206 03:59:33.068905 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 06 03:59:33 crc kubenswrapper[4802]: I1206 03:59:33.227335 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-bq9w6"] Dec 06 03:59:33 crc kubenswrapper[4802]: I1206 03:59:33.463169 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c67bffd47-mrlr7" Dec 06 03:59:33 crc kubenswrapper[4802]: I1206 03:59:33.502933 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85s6x\" (UniqueName: \"kubernetes.io/projected/a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c-kube-api-access-85s6x\") pod \"a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c\" (UID: \"a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c\") " Dec 06 03:59:33 crc kubenswrapper[4802]: I1206 03:59:33.503006 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c-dns-swift-storage-0\") pod \"a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c\" (UID: \"a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c\") " Dec 06 03:59:33 crc kubenswrapper[4802]: I1206 03:59:33.503038 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c-config\") pod \"a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c\" (UID: \"a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c\") " Dec 06 03:59:33 crc kubenswrapper[4802]: I1206 03:59:33.503166 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c-dns-svc\") pod \"a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c\" (UID: \"a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c\") " Dec 06 03:59:33 crc kubenswrapper[4802]: I1206 03:59:33.503223 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c-ovsdbserver-nb\") pod \"a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c\" (UID: \"a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c\") " Dec 06 03:59:33 crc kubenswrapper[4802]: I1206 03:59:33.503279 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c-ovsdbserver-sb\") pod \"a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c\" (UID: \"a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c\") " Dec 06 03:59:33 crc kubenswrapper[4802]: I1206 03:59:33.528985 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c-kube-api-access-85s6x" (OuterVolumeSpecName: "kube-api-access-85s6x") pod "a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c" (UID: "a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c"). InnerVolumeSpecName "kube-api-access-85s6x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:59:33 crc kubenswrapper[4802]: I1206 03:59:33.609114 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85s6x\" (UniqueName: \"kubernetes.io/projected/a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c-kube-api-access-85s6x\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:33 crc kubenswrapper[4802]: I1206 03:59:33.635135 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c" (UID: "a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:59:33 crc kubenswrapper[4802]: I1206 03:59:33.640322 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c-config" (OuterVolumeSpecName: "config") pod "a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c" (UID: "a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:59:33 crc kubenswrapper[4802]: I1206 03:59:33.659023 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c" (UID: "a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:59:33 crc kubenswrapper[4802]: I1206 03:59:33.662393 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c" (UID: "a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:59:33 crc kubenswrapper[4802]: I1206 03:59:33.677814 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c" (UID: "a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:59:33 crc kubenswrapper[4802]: I1206 03:59:33.711066 4802 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:33 crc kubenswrapper[4802]: I1206 03:59:33.711103 4802 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:33 crc kubenswrapper[4802]: I1206 03:59:33.711116 4802 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:33 crc kubenswrapper[4802]: I1206 03:59:33.711133 4802 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:33 crc kubenswrapper[4802]: I1206 03:59:33.711146 4802 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:33 crc kubenswrapper[4802]: I1206 03:59:33.822847 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-844df74668-r698z"] Dec 06 03:59:33 crc kubenswrapper[4802]: I1206 03:59:33.951098 4802 generic.go:334] "Generic (PLEG): container finished" podID="a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c" containerID="5504caf2de012952cdbf63c5da40cb08607e111ca2ac64a80b98bbe613d99954" exitCode=0 Dec 06 03:59:33 crc kubenswrapper[4802]: I1206 03:59:33.951290 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c67bffd47-mrlr7" event={"ID":"a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c","Type":"ContainerDied","Data":"5504caf2de012952cdbf63c5da40cb08607e111ca2ac64a80b98bbe613d99954"} Dec 06 03:59:33 crc kubenswrapper[4802]: I1206 03:59:33.952479 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c67bffd47-mrlr7" event={"ID":"a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c","Type":"ContainerDied","Data":"f09944a2b14a1af41f9b3ee21e231341ffe633826fe185e0eb96f81995808058"} Dec 06 03:59:33 crc kubenswrapper[4802]: I1206 03:59:33.952585 4802 scope.go:117] "RemoveContainer" containerID="5504caf2de012952cdbf63c5da40cb08607e111ca2ac64a80b98bbe613d99954" Dec 06 03:59:33 crc kubenswrapper[4802]: I1206 03:59:33.951364 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c67bffd47-mrlr7" Dec 06 03:59:33 crc kubenswrapper[4802]: I1206 03:59:33.957561 4802 generic.go:334] "Generic (PLEG): container finished" podID="11fd5eb5-07ce-4218-9f9e-50386ba1c4c4" containerID="7d4d29cec1b69ef3172164e6917be7c8d7b5288dfcfb522bbdb38d8ae950bfcb" exitCode=0 Dec 06 03:59:33 crc kubenswrapper[4802]: I1206 03:59:33.958361 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-bq9w6" event={"ID":"11fd5eb5-07ce-4218-9f9e-50386ba1c4c4","Type":"ContainerDied","Data":"7d4d29cec1b69ef3172164e6917be7c8d7b5288dfcfb522bbdb38d8ae950bfcb"} Dec 06 03:59:33 crc kubenswrapper[4802]: I1206 03:59:33.958402 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-bq9w6" event={"ID":"11fd5eb5-07ce-4218-9f9e-50386ba1c4c4","Type":"ContainerStarted","Data":"084a893e1525aeb0b40f311186ced5db14c99a84519c867217416fedab6dfe09"} Dec 06 03:59:33 crc kubenswrapper[4802]: I1206 03:59:33.963235 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-844df74668-r698z" event={"ID":"b364aef0-231f-472a-a8c9-51a5199c0997","Type":"ContainerStarted","Data":"894633cc495354c37aef490a08fa3810f4101afaa0484f5beaa33ccfc89b9bf9"} Dec 06 03:59:34 crc kubenswrapper[4802]: I1206 03:59:34.022890 4802 scope.go:117] "RemoveContainer" containerID="970fda877b59771db20a7752f0768084b73e8e6c9cd7b12537a9dc08c89f9e30" Dec 06 03:59:34 crc kubenswrapper[4802]: I1206 03:59:34.039044 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c67bffd47-mrlr7"] Dec 06 03:59:34 crc kubenswrapper[4802]: I1206 03:59:34.047115 4802 scope.go:117] "RemoveContainer" containerID="5504caf2de012952cdbf63c5da40cb08607e111ca2ac64a80b98bbe613d99954" Dec 06 03:59:34 crc kubenswrapper[4802]: I1206 03:59:34.047123 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7c67bffd47-mrlr7"] Dec 06 03:59:34 crc kubenswrapper[4802]: E1206 03:59:34.047654 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5504caf2de012952cdbf63c5da40cb08607e111ca2ac64a80b98bbe613d99954\": container with ID starting with 5504caf2de012952cdbf63c5da40cb08607e111ca2ac64a80b98bbe613d99954 not found: ID does not exist" containerID="5504caf2de012952cdbf63c5da40cb08607e111ca2ac64a80b98bbe613d99954" Dec 06 03:59:34 crc kubenswrapper[4802]: I1206 03:59:34.047706 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5504caf2de012952cdbf63c5da40cb08607e111ca2ac64a80b98bbe613d99954"} err="failed to get container status \"5504caf2de012952cdbf63c5da40cb08607e111ca2ac64a80b98bbe613d99954\": rpc error: code = NotFound desc = could not find container \"5504caf2de012952cdbf63c5da40cb08607e111ca2ac64a80b98bbe613d99954\": container with ID starting with 5504caf2de012952cdbf63c5da40cb08607e111ca2ac64a80b98bbe613d99954 not found: ID does not exist" Dec 06 03:59:34 crc kubenswrapper[4802]: I1206 03:59:34.047732 4802 scope.go:117] "RemoveContainer" containerID="970fda877b59771db20a7752f0768084b73e8e6c9cd7b12537a9dc08c89f9e30" Dec 06 03:59:34 crc kubenswrapper[4802]: E1206 03:59:34.048394 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"970fda877b59771db20a7752f0768084b73e8e6c9cd7b12537a9dc08c89f9e30\": container with ID starting with 970fda877b59771db20a7752f0768084b73e8e6c9cd7b12537a9dc08c89f9e30 not found: ID does not exist" containerID="970fda877b59771db20a7752f0768084b73e8e6c9cd7b12537a9dc08c89f9e30" Dec 06 03:59:34 crc kubenswrapper[4802]: I1206 03:59:34.048427 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"970fda877b59771db20a7752f0768084b73e8e6c9cd7b12537a9dc08c89f9e30"} err="failed to get container status \"970fda877b59771db20a7752f0768084b73e8e6c9cd7b12537a9dc08c89f9e30\": rpc error: code = NotFound desc = could not find container \"970fda877b59771db20a7752f0768084b73e8e6c9cd7b12537a9dc08c89f9e30\": container with ID starting with 970fda877b59771db20a7752f0768084b73e8e6c9cd7b12537a9dc08c89f9e30 not found: ID does not exist" Dec 06 03:59:34 crc kubenswrapper[4802]: I1206 03:59:34.990476 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-bq9w6" event={"ID":"11fd5eb5-07ce-4218-9f9e-50386ba1c4c4","Type":"ContainerStarted","Data":"034e8029d69444bbc17bc0698b97b4b46769903bd3005c879ee287e371adea4d"} Dec 06 03:59:34 crc kubenswrapper[4802]: I1206 03:59:34.991063 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-848cf88cfc-bq9w6" Dec 06 03:59:34 crc kubenswrapper[4802]: I1206 03:59:34.992533 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-844df74668-r698z" event={"ID":"b364aef0-231f-472a-a8c9-51a5199c0997","Type":"ContainerStarted","Data":"31dc097fa5cf9c76a286ae10dd83297f83da4c4ead70f0e1d9e17473d8ccab81"} Dec 06 03:59:34 crc kubenswrapper[4802]: I1206 03:59:34.992569 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-844df74668-r698z" event={"ID":"b364aef0-231f-472a-a8c9-51a5199c0997","Type":"ContainerStarted","Data":"1bd700182ce971dd2c603b23e1a7c8370de9caf2e1ce639678041c207ab36f16"} Dec 06 03:59:34 crc kubenswrapper[4802]: I1206 03:59:34.992740 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-844df74668-r698z" Dec 06 03:59:35 crc kubenswrapper[4802]: I1206 03:59:35.001935 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-68bd774b7c-9sjk9"] Dec 06 03:59:35 crc kubenswrapper[4802]: E1206 03:59:35.002266 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c" containerName="dnsmasq-dns" Dec 06 03:59:35 crc kubenswrapper[4802]: I1206 03:59:35.002285 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c" containerName="dnsmasq-dns" Dec 06 03:59:35 crc kubenswrapper[4802]: E1206 03:59:35.002303 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c" containerName="init" Dec 06 03:59:35 crc kubenswrapper[4802]: I1206 03:59:35.002309 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c" containerName="init" Dec 06 03:59:35 crc kubenswrapper[4802]: I1206 03:59:35.002522 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c" containerName="dnsmasq-dns" Dec 06 03:59:35 crc kubenswrapper[4802]: I1206 03:59:35.003408 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-68bd774b7c-9sjk9" Dec 06 03:59:35 crc kubenswrapper[4802]: I1206 03:59:35.008335 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 06 03:59:35 crc kubenswrapper[4802]: I1206 03:59:35.008672 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 06 03:59:35 crc kubenswrapper[4802]: I1206 03:59:35.026688 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-68bd774b7c-9sjk9"] Dec 06 03:59:35 crc kubenswrapper[4802]: I1206 03:59:35.030081 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-848cf88cfc-bq9w6" podStartSLOduration=3.030064165 podStartE2EDuration="3.030064165s" podCreationTimestamp="2025-12-06 03:59:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:59:35.024547685 +0000 UTC m=+1167.896456837" watchObservedRunningTime="2025-12-06 03:59:35.030064165 +0000 UTC m=+1167.901973317" Dec 06 03:59:35 crc kubenswrapper[4802]: I1206 03:59:35.032734 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8bf882a4-ac26-40da-8bcc-46052a9db457-public-tls-certs\") pod \"neutron-68bd774b7c-9sjk9\" (UID: \"8bf882a4-ac26-40da-8bcc-46052a9db457\") " pod="openstack/neutron-68bd774b7c-9sjk9" Dec 06 03:59:35 crc kubenswrapper[4802]: I1206 03:59:35.032795 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8bf882a4-ac26-40da-8bcc-46052a9db457-internal-tls-certs\") pod \"neutron-68bd774b7c-9sjk9\" (UID: \"8bf882a4-ac26-40da-8bcc-46052a9db457\") " pod="openstack/neutron-68bd774b7c-9sjk9" Dec 06 03:59:35 crc kubenswrapper[4802]: I1206 03:59:35.032835 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8bf882a4-ac26-40da-8bcc-46052a9db457-httpd-config\") pod \"neutron-68bd774b7c-9sjk9\" (UID: \"8bf882a4-ac26-40da-8bcc-46052a9db457\") " pod="openstack/neutron-68bd774b7c-9sjk9" Dec 06 03:59:35 crc kubenswrapper[4802]: I1206 03:59:35.032860 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bf882a4-ac26-40da-8bcc-46052a9db457-combined-ca-bundle\") pod \"neutron-68bd774b7c-9sjk9\" (UID: \"8bf882a4-ac26-40da-8bcc-46052a9db457\") " pod="openstack/neutron-68bd774b7c-9sjk9" Dec 06 03:59:35 crc kubenswrapper[4802]: I1206 03:59:35.032925 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8bf882a4-ac26-40da-8bcc-46052a9db457-config\") pod \"neutron-68bd774b7c-9sjk9\" (UID: \"8bf882a4-ac26-40da-8bcc-46052a9db457\") " pod="openstack/neutron-68bd774b7c-9sjk9" Dec 06 03:59:35 crc kubenswrapper[4802]: I1206 03:59:35.032945 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nt82x\" (UniqueName: \"kubernetes.io/projected/8bf882a4-ac26-40da-8bcc-46052a9db457-kube-api-access-nt82x\") pod \"neutron-68bd774b7c-9sjk9\" (UID: \"8bf882a4-ac26-40da-8bcc-46052a9db457\") " pod="openstack/neutron-68bd774b7c-9sjk9" Dec 06 03:59:35 crc kubenswrapper[4802]: I1206 03:59:35.032979 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8bf882a4-ac26-40da-8bcc-46052a9db457-ovndb-tls-certs\") pod \"neutron-68bd774b7c-9sjk9\" (UID: \"8bf882a4-ac26-40da-8bcc-46052a9db457\") " pod="openstack/neutron-68bd774b7c-9sjk9" Dec 06 03:59:35 crc kubenswrapper[4802]: I1206 03:59:35.071691 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-844df74668-r698z" podStartSLOduration=3.0716764579999998 podStartE2EDuration="3.071676458s" podCreationTimestamp="2025-12-06 03:59:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:59:35.069139129 +0000 UTC m=+1167.941048281" watchObservedRunningTime="2025-12-06 03:59:35.071676458 +0000 UTC m=+1167.943585610" Dec 06 03:59:35 crc kubenswrapper[4802]: I1206 03:59:35.135976 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8bf882a4-ac26-40da-8bcc-46052a9db457-config\") pod \"neutron-68bd774b7c-9sjk9\" (UID: \"8bf882a4-ac26-40da-8bcc-46052a9db457\") " pod="openstack/neutron-68bd774b7c-9sjk9" Dec 06 03:59:35 crc kubenswrapper[4802]: I1206 03:59:35.136255 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nt82x\" (UniqueName: \"kubernetes.io/projected/8bf882a4-ac26-40da-8bcc-46052a9db457-kube-api-access-nt82x\") pod \"neutron-68bd774b7c-9sjk9\" (UID: \"8bf882a4-ac26-40da-8bcc-46052a9db457\") " pod="openstack/neutron-68bd774b7c-9sjk9" Dec 06 03:59:35 crc kubenswrapper[4802]: I1206 03:59:35.136356 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8bf882a4-ac26-40da-8bcc-46052a9db457-ovndb-tls-certs\") pod \"neutron-68bd774b7c-9sjk9\" (UID: \"8bf882a4-ac26-40da-8bcc-46052a9db457\") " pod="openstack/neutron-68bd774b7c-9sjk9" Dec 06 03:59:35 crc kubenswrapper[4802]: I1206 03:59:35.137541 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8bf882a4-ac26-40da-8bcc-46052a9db457-public-tls-certs\") pod \"neutron-68bd774b7c-9sjk9\" (UID: \"8bf882a4-ac26-40da-8bcc-46052a9db457\") " pod="openstack/neutron-68bd774b7c-9sjk9" Dec 06 03:59:35 crc kubenswrapper[4802]: I1206 03:59:35.138283 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8bf882a4-ac26-40da-8bcc-46052a9db457-internal-tls-certs\") pod \"neutron-68bd774b7c-9sjk9\" (UID: \"8bf882a4-ac26-40da-8bcc-46052a9db457\") " pod="openstack/neutron-68bd774b7c-9sjk9" Dec 06 03:59:35 crc kubenswrapper[4802]: I1206 03:59:35.138433 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8bf882a4-ac26-40da-8bcc-46052a9db457-httpd-config\") pod \"neutron-68bd774b7c-9sjk9\" (UID: \"8bf882a4-ac26-40da-8bcc-46052a9db457\") " pod="openstack/neutron-68bd774b7c-9sjk9" Dec 06 03:59:35 crc kubenswrapper[4802]: I1206 03:59:35.138601 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bf882a4-ac26-40da-8bcc-46052a9db457-combined-ca-bundle\") pod \"neutron-68bd774b7c-9sjk9\" (UID: \"8bf882a4-ac26-40da-8bcc-46052a9db457\") " pod="openstack/neutron-68bd774b7c-9sjk9" Dec 06 03:59:35 crc kubenswrapper[4802]: I1206 03:59:35.142222 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/8bf882a4-ac26-40da-8bcc-46052a9db457-config\") pod \"neutron-68bd774b7c-9sjk9\" (UID: \"8bf882a4-ac26-40da-8bcc-46052a9db457\") " pod="openstack/neutron-68bd774b7c-9sjk9" Dec 06 03:59:35 crc kubenswrapper[4802]: I1206 03:59:35.142952 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8bf882a4-ac26-40da-8bcc-46052a9db457-public-tls-certs\") pod \"neutron-68bd774b7c-9sjk9\" (UID: \"8bf882a4-ac26-40da-8bcc-46052a9db457\") " pod="openstack/neutron-68bd774b7c-9sjk9" Dec 06 03:59:35 crc kubenswrapper[4802]: I1206 03:59:35.147540 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bf882a4-ac26-40da-8bcc-46052a9db457-combined-ca-bundle\") pod \"neutron-68bd774b7c-9sjk9\" (UID: \"8bf882a4-ac26-40da-8bcc-46052a9db457\") " pod="openstack/neutron-68bd774b7c-9sjk9" Dec 06 03:59:35 crc kubenswrapper[4802]: I1206 03:59:35.148177 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8bf882a4-ac26-40da-8bcc-46052a9db457-httpd-config\") pod \"neutron-68bd774b7c-9sjk9\" (UID: \"8bf882a4-ac26-40da-8bcc-46052a9db457\") " pod="openstack/neutron-68bd774b7c-9sjk9" Dec 06 03:59:35 crc kubenswrapper[4802]: I1206 03:59:35.155531 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8bf882a4-ac26-40da-8bcc-46052a9db457-ovndb-tls-certs\") pod \"neutron-68bd774b7c-9sjk9\" (UID: \"8bf882a4-ac26-40da-8bcc-46052a9db457\") " pod="openstack/neutron-68bd774b7c-9sjk9" Dec 06 03:59:35 crc kubenswrapper[4802]: I1206 03:59:35.157432 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nt82x\" (UniqueName: \"kubernetes.io/projected/8bf882a4-ac26-40da-8bcc-46052a9db457-kube-api-access-nt82x\") pod \"neutron-68bd774b7c-9sjk9\" (UID: \"8bf882a4-ac26-40da-8bcc-46052a9db457\") " pod="openstack/neutron-68bd774b7c-9sjk9" Dec 06 03:59:35 crc kubenswrapper[4802]: I1206 03:59:35.163106 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8bf882a4-ac26-40da-8bcc-46052a9db457-internal-tls-certs\") pod \"neutron-68bd774b7c-9sjk9\" (UID: \"8bf882a4-ac26-40da-8bcc-46052a9db457\") " pod="openstack/neutron-68bd774b7c-9sjk9" Dec 06 03:59:35 crc kubenswrapper[4802]: I1206 03:59:35.323063 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-68bd774b7c-9sjk9" Dec 06 03:59:35 crc kubenswrapper[4802]: I1206 03:59:35.464111 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c" path="/var/lib/kubelet/pods/a57fa7b2-4bcd-47b0-9ef5-5c005e8fb52c/volumes" Dec 06 03:59:35 crc kubenswrapper[4802]: I1206 03:59:35.898437 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-68bd774b7c-9sjk9"] Dec 06 03:59:35 crc kubenswrapper[4802]: W1206 03:59:35.901350 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8bf882a4_ac26_40da_8bcc_46052a9db457.slice/crio-aa9fb905810c8ad2782b3df165c10aa2c9091a52dbdbac4049769159cdae1599 WatchSource:0}: Error finding container aa9fb905810c8ad2782b3df165c10aa2c9091a52dbdbac4049769159cdae1599: Status 404 returned error can't find the container with id aa9fb905810c8ad2782b3df165c10aa2c9091a52dbdbac4049769159cdae1599 Dec 06 03:59:36 crc kubenswrapper[4802]: I1206 03:59:36.005307 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-68bd774b7c-9sjk9" event={"ID":"8bf882a4-ac26-40da-8bcc-46052a9db457","Type":"ContainerStarted","Data":"aa9fb905810c8ad2782b3df165c10aa2c9091a52dbdbac4049769159cdae1599"} Dec 06 03:59:37 crc kubenswrapper[4802]: I1206 03:59:37.017480 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-68bd774b7c-9sjk9" event={"ID":"8bf882a4-ac26-40da-8bcc-46052a9db457","Type":"ContainerStarted","Data":"512e15f941369580c9651065deb57df5f36a4e9a5ded0d72413f05e0501aeb80"} Dec 06 03:59:37 crc kubenswrapper[4802]: I1206 03:59:37.018094 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-68bd774b7c-9sjk9" event={"ID":"8bf882a4-ac26-40da-8bcc-46052a9db457","Type":"ContainerStarted","Data":"eab109844c716eb6d9464fcaa7909f8d6b2486e03f8475882c77909097c6d35e"} Dec 06 03:59:37 crc kubenswrapper[4802]: I1206 03:59:37.018179 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-68bd774b7c-9sjk9" Dec 06 03:59:38 crc kubenswrapper[4802]: I1206 03:59:38.013180 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6f6b9c98b6-8ndrw" Dec 06 03:59:38 crc kubenswrapper[4802]: I1206 03:59:38.039036 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-68bd774b7c-9sjk9" podStartSLOduration=4.039002571 podStartE2EDuration="4.039002571s" podCreationTimestamp="2025-12-06 03:59:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:59:37.045347492 +0000 UTC m=+1169.917256644" watchObservedRunningTime="2025-12-06 03:59:38.039002571 +0000 UTC m=+1170.910911723" Dec 06 03:59:41 crc kubenswrapper[4802]: I1206 03:59:41.017056 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7f79b98d58-l9qc5" Dec 06 03:59:41 crc kubenswrapper[4802]: I1206 03:59:41.057190 4802 generic.go:334] "Generic (PLEG): container finished" podID="f1644907-b4b1-4e23-a273-19d535bd7f88" containerID="66f602546227aa86f52ceb621c4f145c78c1b12628e754ab524c6e32f88a0151" exitCode=0 Dec 06 03:59:41 crc kubenswrapper[4802]: I1206 03:59:41.057312 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-5hj5p" event={"ID":"f1644907-b4b1-4e23-a273-19d535bd7f88","Type":"ContainerDied","Data":"66f602546227aa86f52ceb621c4f145c78c1b12628e754ab524c6e32f88a0151"} Dec 06 03:59:41 crc kubenswrapper[4802]: I1206 03:59:41.060098 4802 generic.go:334] "Generic (PLEG): container finished" podID="2a639667-be04-429f-a2ce-b20a011425f8" containerID="285d42f44e31edd7beb22eac4c447d12107075be937e5c0e69d394dc90c0ac17" exitCode=0 Dec 06 03:59:41 crc kubenswrapper[4802]: I1206 03:59:41.060149 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-2hzlt" event={"ID":"2a639667-be04-429f-a2ce-b20a011425f8","Type":"ContainerDied","Data":"285d42f44e31edd7beb22eac4c447d12107075be937e5c0e69d394dc90c0ac17"} Dec 06 03:59:41 crc kubenswrapper[4802]: I1206 03:59:41.090175 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7f79b98d58-l9qc5" Dec 06 03:59:41 crc kubenswrapper[4802]: I1206 03:59:41.176484 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6f6b9c98b6-8ndrw"] Dec 06 03:59:41 crc kubenswrapper[4802]: I1206 03:59:41.176706 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6f6b9c98b6-8ndrw" podUID="d098e8a2-f622-44f7-997f-51c6ff7eceaa" containerName="barbican-api-log" containerID="cri-o://dac72a8abf11cd63050015a15e0298021665dc56d9b2efceac05497ee84a7a29" gracePeriod=30 Dec 06 03:59:41 crc kubenswrapper[4802]: I1206 03:59:41.176880 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6f6b9c98b6-8ndrw" podUID="d098e8a2-f622-44f7-997f-51c6ff7eceaa" containerName="barbican-api" containerID="cri-o://748b82913c72c4ebf47e1088833f5e5da1132a05f680ef7cb0ff05f10fe522c5" gracePeriod=30 Dec 06 03:59:42 crc kubenswrapper[4802]: I1206 03:59:42.069622 4802 generic.go:334] "Generic (PLEG): container finished" podID="d098e8a2-f622-44f7-997f-51c6ff7eceaa" containerID="dac72a8abf11cd63050015a15e0298021665dc56d9b2efceac05497ee84a7a29" exitCode=143 Dec 06 03:59:42 crc kubenswrapper[4802]: I1206 03:59:42.069898 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6f6b9c98b6-8ndrw" event={"ID":"d098e8a2-f622-44f7-997f-51c6ff7eceaa","Type":"ContainerDied","Data":"dac72a8abf11cd63050015a15e0298021665dc56d9b2efceac05497ee84a7a29"} Dec 06 03:59:42 crc kubenswrapper[4802]: I1206 03:59:42.507705 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-5hj5p" Dec 06 03:59:42 crc kubenswrapper[4802]: I1206 03:59:42.527111 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-2hzlt" Dec 06 03:59:42 crc kubenswrapper[4802]: I1206 03:59:42.610693 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1644907-b4b1-4e23-a273-19d535bd7f88-combined-ca-bundle\") pod \"f1644907-b4b1-4e23-a273-19d535bd7f88\" (UID: \"f1644907-b4b1-4e23-a273-19d535bd7f88\") " Dec 06 03:59:42 crc kubenswrapper[4802]: I1206 03:59:42.610829 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f1644907-b4b1-4e23-a273-19d535bd7f88-scripts\") pod \"f1644907-b4b1-4e23-a273-19d535bd7f88\" (UID: \"f1644907-b4b1-4e23-a273-19d535bd7f88\") " Dec 06 03:59:42 crc kubenswrapper[4802]: I1206 03:59:42.610877 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f1644907-b4b1-4e23-a273-19d535bd7f88-etc-machine-id\") pod \"f1644907-b4b1-4e23-a273-19d535bd7f88\" (UID: \"f1644907-b4b1-4e23-a273-19d535bd7f88\") " Dec 06 03:59:42 crc kubenswrapper[4802]: I1206 03:59:42.610907 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pln97\" (UniqueName: \"kubernetes.io/projected/f1644907-b4b1-4e23-a273-19d535bd7f88-kube-api-access-pln97\") pod \"f1644907-b4b1-4e23-a273-19d535bd7f88\" (UID: \"f1644907-b4b1-4e23-a273-19d535bd7f88\") " Dec 06 03:59:42 crc kubenswrapper[4802]: I1206 03:59:42.610953 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f1644907-b4b1-4e23-a273-19d535bd7f88-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "f1644907-b4b1-4e23-a273-19d535bd7f88" (UID: "f1644907-b4b1-4e23-a273-19d535bd7f88"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:59:42 crc kubenswrapper[4802]: I1206 03:59:42.611635 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f1644907-b4b1-4e23-a273-19d535bd7f88-db-sync-config-data\") pod \"f1644907-b4b1-4e23-a273-19d535bd7f88\" (UID: \"f1644907-b4b1-4e23-a273-19d535bd7f88\") " Dec 06 03:59:42 crc kubenswrapper[4802]: I1206 03:59:42.611733 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1644907-b4b1-4e23-a273-19d535bd7f88-config-data\") pod \"f1644907-b4b1-4e23-a273-19d535bd7f88\" (UID: \"f1644907-b4b1-4e23-a273-19d535bd7f88\") " Dec 06 03:59:42 crc kubenswrapper[4802]: I1206 03:59:42.614416 4802 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f1644907-b4b1-4e23-a273-19d535bd7f88-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:42 crc kubenswrapper[4802]: I1206 03:59:42.616471 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1644907-b4b1-4e23-a273-19d535bd7f88-kube-api-access-pln97" (OuterVolumeSpecName: "kube-api-access-pln97") pod "f1644907-b4b1-4e23-a273-19d535bd7f88" (UID: "f1644907-b4b1-4e23-a273-19d535bd7f88"). InnerVolumeSpecName "kube-api-access-pln97". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:59:42 crc kubenswrapper[4802]: I1206 03:59:42.617248 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1644907-b4b1-4e23-a273-19d535bd7f88-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "f1644907-b4b1-4e23-a273-19d535bd7f88" (UID: "f1644907-b4b1-4e23-a273-19d535bd7f88"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:59:42 crc kubenswrapper[4802]: I1206 03:59:42.625898 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1644907-b4b1-4e23-a273-19d535bd7f88-scripts" (OuterVolumeSpecName: "scripts") pod "f1644907-b4b1-4e23-a273-19d535bd7f88" (UID: "f1644907-b4b1-4e23-a273-19d535bd7f88"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:59:42 crc kubenswrapper[4802]: I1206 03:59:42.639083 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1644907-b4b1-4e23-a273-19d535bd7f88-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f1644907-b4b1-4e23-a273-19d535bd7f88" (UID: "f1644907-b4b1-4e23-a273-19d535bd7f88"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:59:42 crc kubenswrapper[4802]: I1206 03:59:42.643946 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-848cf88cfc-bq9w6" Dec 06 03:59:42 crc kubenswrapper[4802]: I1206 03:59:42.658561 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1644907-b4b1-4e23-a273-19d535bd7f88-config-data" (OuterVolumeSpecName: "config-data") pod "f1644907-b4b1-4e23-a273-19d535bd7f88" (UID: "f1644907-b4b1-4e23-a273-19d535bd7f88"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:59:42 crc kubenswrapper[4802]: I1206 03:59:42.715012 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a639667-be04-429f-a2ce-b20a011425f8-config-data\") pod \"2a639667-be04-429f-a2ce-b20a011425f8\" (UID: \"2a639667-be04-429f-a2ce-b20a011425f8\") " Dec 06 03:59:42 crc kubenswrapper[4802]: I1206 03:59:42.715249 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ssjb\" (UniqueName: \"kubernetes.io/projected/2a639667-be04-429f-a2ce-b20a011425f8-kube-api-access-6ssjb\") pod \"2a639667-be04-429f-a2ce-b20a011425f8\" (UID: \"2a639667-be04-429f-a2ce-b20a011425f8\") " Dec 06 03:59:42 crc kubenswrapper[4802]: I1206 03:59:42.715285 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a639667-be04-429f-a2ce-b20a011425f8-combined-ca-bundle\") pod \"2a639667-be04-429f-a2ce-b20a011425f8\" (UID: \"2a639667-be04-429f-a2ce-b20a011425f8\") " Dec 06 03:59:42 crc kubenswrapper[4802]: I1206 03:59:42.716399 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1644907-b4b1-4e23-a273-19d535bd7f88-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:42 crc kubenswrapper[4802]: I1206 03:59:42.716417 4802 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1644907-b4b1-4e23-a273-19d535bd7f88-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:42 crc kubenswrapper[4802]: I1206 03:59:42.716426 4802 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f1644907-b4b1-4e23-a273-19d535bd7f88-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:42 crc kubenswrapper[4802]: I1206 03:59:42.716437 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pln97\" (UniqueName: \"kubernetes.io/projected/f1644907-b4b1-4e23-a273-19d535bd7f88-kube-api-access-pln97\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:42 crc kubenswrapper[4802]: I1206 03:59:42.716445 4802 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f1644907-b4b1-4e23-a273-19d535bd7f88-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:42 crc kubenswrapper[4802]: I1206 03:59:42.716474 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-z4x9c"] Dec 06 03:59:42 crc kubenswrapper[4802]: I1206 03:59:42.716693 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-56df8fb6b7-z4x9c" podUID="f91ac3c0-8611-458e-bb4a-cfacb60d8b5b" containerName="dnsmasq-dns" containerID="cri-o://dbc39ed94f5b7975108e7a698f80ef79b3a5290662fc25802033727d945074ef" gracePeriod=10 Dec 06 03:59:42 crc kubenswrapper[4802]: I1206 03:59:42.727243 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a639667-be04-429f-a2ce-b20a011425f8-kube-api-access-6ssjb" (OuterVolumeSpecName: "kube-api-access-6ssjb") pod "2a639667-be04-429f-a2ce-b20a011425f8" (UID: "2a639667-be04-429f-a2ce-b20a011425f8"). InnerVolumeSpecName "kube-api-access-6ssjb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:59:42 crc kubenswrapper[4802]: I1206 03:59:42.757985 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a639667-be04-429f-a2ce-b20a011425f8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2a639667-be04-429f-a2ce-b20a011425f8" (UID: "2a639667-be04-429f-a2ce-b20a011425f8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:59:42 crc kubenswrapper[4802]: I1206 03:59:42.821071 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ssjb\" (UniqueName: \"kubernetes.io/projected/2a639667-be04-429f-a2ce-b20a011425f8-kube-api-access-6ssjb\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:42 crc kubenswrapper[4802]: I1206 03:59:42.821112 4802 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a639667-be04-429f-a2ce-b20a011425f8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:42 crc kubenswrapper[4802]: I1206 03:59:42.878949 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a639667-be04-429f-a2ce-b20a011425f8-config-data" (OuterVolumeSpecName: "config-data") pod "2a639667-be04-429f-a2ce-b20a011425f8" (UID: "2a639667-be04-429f-a2ce-b20a011425f8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:59:42 crc kubenswrapper[4802]: I1206 03:59:42.923066 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a639667-be04-429f-a2ce-b20a011425f8-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.084645 4802 generic.go:334] "Generic (PLEG): container finished" podID="f91ac3c0-8611-458e-bb4a-cfacb60d8b5b" containerID="dbc39ed94f5b7975108e7a698f80ef79b3a5290662fc25802033727d945074ef" exitCode=0 Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.084868 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-z4x9c" event={"ID":"f91ac3c0-8611-458e-bb4a-cfacb60d8b5b","Type":"ContainerDied","Data":"dbc39ed94f5b7975108e7a698f80ef79b3a5290662fc25802033727d945074ef"} Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.102302 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-2hzlt" event={"ID":"2a639667-be04-429f-a2ce-b20a011425f8","Type":"ContainerDied","Data":"3173bdb2f47b8aaeecd93e04af3df770499eef5d3268913395d23e8481c67cc6"} Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.102536 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3173bdb2f47b8aaeecd93e04af3df770499eef5d3268913395d23e8481c67cc6" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.102732 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-2hzlt" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.124064 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-5hj5p" event={"ID":"f1644907-b4b1-4e23-a273-19d535bd7f88","Type":"ContainerDied","Data":"e0760f7b4403432817d1e174766269537a2b9a6108c03ff28e47ddde093058e0"} Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.124107 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e0760f7b4403432817d1e174766269537a2b9a6108c03ff28e47ddde093058e0" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.124222 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-5hj5p" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.216795 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-z4x9c" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.285106 4802 patch_prober.go:28] interesting pod/machine-config-daemon-zpxxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.285346 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.285450 4802 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.286175 4802 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"04d0056d567a84a69f17a8c41df89172171e22968697aae49e1b7d82534da79c"} pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.286317 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" containerID="cri-o://04d0056d567a84a69f17a8c41df89172171e22968697aae49e1b7d82534da79c" gracePeriod=600 Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.338432 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f91ac3c0-8611-458e-bb4a-cfacb60d8b5b-config\") pod \"f91ac3c0-8611-458e-bb4a-cfacb60d8b5b\" (UID: \"f91ac3c0-8611-458e-bb4a-cfacb60d8b5b\") " Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.338524 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f91ac3c0-8611-458e-bb4a-cfacb60d8b5b-dns-svc\") pod \"f91ac3c0-8611-458e-bb4a-cfacb60d8b5b\" (UID: \"f91ac3c0-8611-458e-bb4a-cfacb60d8b5b\") " Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.338577 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f91ac3c0-8611-458e-bb4a-cfacb60d8b5b-dns-swift-storage-0\") pod \"f91ac3c0-8611-458e-bb4a-cfacb60d8b5b\" (UID: \"f91ac3c0-8611-458e-bb4a-cfacb60d8b5b\") " Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.338639 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hsfz5\" (UniqueName: \"kubernetes.io/projected/f91ac3c0-8611-458e-bb4a-cfacb60d8b5b-kube-api-access-hsfz5\") pod \"f91ac3c0-8611-458e-bb4a-cfacb60d8b5b\" (UID: \"f91ac3c0-8611-458e-bb4a-cfacb60d8b5b\") " Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.338658 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f91ac3c0-8611-458e-bb4a-cfacb60d8b5b-ovsdbserver-nb\") pod \"f91ac3c0-8611-458e-bb4a-cfacb60d8b5b\" (UID: \"f91ac3c0-8611-458e-bb4a-cfacb60d8b5b\") " Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.338735 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f91ac3c0-8611-458e-bb4a-cfacb60d8b5b-ovsdbserver-sb\") pod \"f91ac3c0-8611-458e-bb4a-cfacb60d8b5b\" (UID: \"f91ac3c0-8611-458e-bb4a-cfacb60d8b5b\") " Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.360685 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 03:59:43 crc kubenswrapper[4802]: E1206 03:59:43.361040 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a639667-be04-429f-a2ce-b20a011425f8" containerName="heat-db-sync" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.361056 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a639667-be04-429f-a2ce-b20a011425f8" containerName="heat-db-sync" Dec 06 03:59:43 crc kubenswrapper[4802]: E1206 03:59:43.361069 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1644907-b4b1-4e23-a273-19d535bd7f88" containerName="cinder-db-sync" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.361075 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1644907-b4b1-4e23-a273-19d535bd7f88" containerName="cinder-db-sync" Dec 06 03:59:43 crc kubenswrapper[4802]: E1206 03:59:43.361087 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f91ac3c0-8611-458e-bb4a-cfacb60d8b5b" containerName="dnsmasq-dns" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.361093 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="f91ac3c0-8611-458e-bb4a-cfacb60d8b5b" containerName="dnsmasq-dns" Dec 06 03:59:43 crc kubenswrapper[4802]: E1206 03:59:43.361105 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f91ac3c0-8611-458e-bb4a-cfacb60d8b5b" containerName="init" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.361111 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="f91ac3c0-8611-458e-bb4a-cfacb60d8b5b" containerName="init" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.361285 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="f91ac3c0-8611-458e-bb4a-cfacb60d8b5b" containerName="dnsmasq-dns" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.361299 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a639667-be04-429f-a2ce-b20a011425f8" containerName="heat-db-sync" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.361317 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1644907-b4b1-4e23-a273-19d535bd7f88" containerName="cinder-db-sync" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.362183 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.366688 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.366919 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.367109 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-5k5s6" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.367259 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.380695 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.388955 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f91ac3c0-8611-458e-bb4a-cfacb60d8b5b-kube-api-access-hsfz5" (OuterVolumeSpecName: "kube-api-access-hsfz5") pod "f91ac3c0-8611-458e-bb4a-cfacb60d8b5b" (UID: "f91ac3c0-8611-458e-bb4a-cfacb60d8b5b"). InnerVolumeSpecName "kube-api-access-hsfz5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.444619 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hsfz5\" (UniqueName: \"kubernetes.io/projected/f91ac3c0-8611-458e-bb4a-cfacb60d8b5b-kube-api-access-hsfz5\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.494900 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f91ac3c0-8611-458e-bb4a-cfacb60d8b5b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f91ac3c0-8611-458e-bb4a-cfacb60d8b5b" (UID: "f91ac3c0-8611-458e-bb4a-cfacb60d8b5b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.525268 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f91ac3c0-8611-458e-bb4a-cfacb60d8b5b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f91ac3c0-8611-458e-bb4a-cfacb60d8b5b" (UID: "f91ac3c0-8611-458e-bb4a-cfacb60d8b5b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.546343 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfefa56f-3099-4f29-b43b-47e1fcb7ee3d-config-data\") pod \"cinder-scheduler-0\" (UID: \"bfefa56f-3099-4f29-b43b-47e1fcb7ee3d\") " pod="openstack/cinder-scheduler-0" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.546769 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bfefa56f-3099-4f29-b43b-47e1fcb7ee3d-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"bfefa56f-3099-4f29-b43b-47e1fcb7ee3d\") " pod="openstack/cinder-scheduler-0" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.546831 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfefa56f-3099-4f29-b43b-47e1fcb7ee3d-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"bfefa56f-3099-4f29-b43b-47e1fcb7ee3d\") " pod="openstack/cinder-scheduler-0" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.546873 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bfefa56f-3099-4f29-b43b-47e1fcb7ee3d-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"bfefa56f-3099-4f29-b43b-47e1fcb7ee3d\") " pod="openstack/cinder-scheduler-0" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.546905 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfefa56f-3099-4f29-b43b-47e1fcb7ee3d-scripts\") pod \"cinder-scheduler-0\" (UID: \"bfefa56f-3099-4f29-b43b-47e1fcb7ee3d\") " pod="openstack/cinder-scheduler-0" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.546955 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79vhb\" (UniqueName: \"kubernetes.io/projected/bfefa56f-3099-4f29-b43b-47e1fcb7ee3d-kube-api-access-79vhb\") pod \"cinder-scheduler-0\" (UID: \"bfefa56f-3099-4f29-b43b-47e1fcb7ee3d\") " pod="openstack/cinder-scheduler-0" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.555697 4802 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f91ac3c0-8611-458e-bb4a-cfacb60d8b5b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.555730 4802 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f91ac3c0-8611-458e-bb4a-cfacb60d8b5b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.556263 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f91ac3c0-8611-458e-bb4a-cfacb60d8b5b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f91ac3c0-8611-458e-bb4a-cfacb60d8b5b" (UID: "f91ac3c0-8611-458e-bb4a-cfacb60d8b5b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.567383 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f91ac3c0-8611-458e-bb4a-cfacb60d8b5b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "f91ac3c0-8611-458e-bb4a-cfacb60d8b5b" (UID: "f91ac3c0-8611-458e-bb4a-cfacb60d8b5b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.577290 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f91ac3c0-8611-458e-bb4a-cfacb60d8b5b-config" (OuterVolumeSpecName: "config") pod "f91ac3c0-8611-458e-bb4a-cfacb60d8b5b" (UID: "f91ac3c0-8611-458e-bb4a-cfacb60d8b5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.642735 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-h9zv8"] Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.644586 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-h9zv8"] Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.644641 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.644709 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-h9zv8" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.646478 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.648382 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.653586 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.659414 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bfefa56f-3099-4f29-b43b-47e1fcb7ee3d-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"bfefa56f-3099-4f29-b43b-47e1fcb7ee3d\") " pod="openstack/cinder-scheduler-0" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.659488 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfefa56f-3099-4f29-b43b-47e1fcb7ee3d-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"bfefa56f-3099-4f29-b43b-47e1fcb7ee3d\") " pod="openstack/cinder-scheduler-0" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.659525 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bfefa56f-3099-4f29-b43b-47e1fcb7ee3d-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"bfefa56f-3099-4f29-b43b-47e1fcb7ee3d\") " pod="openstack/cinder-scheduler-0" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.659548 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfefa56f-3099-4f29-b43b-47e1fcb7ee3d-scripts\") pod \"cinder-scheduler-0\" (UID: \"bfefa56f-3099-4f29-b43b-47e1fcb7ee3d\") " pod="openstack/cinder-scheduler-0" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.659584 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79vhb\" (UniqueName: \"kubernetes.io/projected/bfefa56f-3099-4f29-b43b-47e1fcb7ee3d-kube-api-access-79vhb\") pod \"cinder-scheduler-0\" (UID: \"bfefa56f-3099-4f29-b43b-47e1fcb7ee3d\") " pod="openstack/cinder-scheduler-0" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.659632 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfefa56f-3099-4f29-b43b-47e1fcb7ee3d-config-data\") pod \"cinder-scheduler-0\" (UID: \"bfefa56f-3099-4f29-b43b-47e1fcb7ee3d\") " pod="openstack/cinder-scheduler-0" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.659711 4802 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f91ac3c0-8611-458e-bb4a-cfacb60d8b5b-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.659724 4802 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f91ac3c0-8611-458e-bb4a-cfacb60d8b5b-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.659733 4802 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f91ac3c0-8611-458e-bb4a-cfacb60d8b5b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.660763 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bfefa56f-3099-4f29-b43b-47e1fcb7ee3d-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"bfefa56f-3099-4f29-b43b-47e1fcb7ee3d\") " pod="openstack/cinder-scheduler-0" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.664388 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfefa56f-3099-4f29-b43b-47e1fcb7ee3d-scripts\") pod \"cinder-scheduler-0\" (UID: \"bfefa56f-3099-4f29-b43b-47e1fcb7ee3d\") " pod="openstack/cinder-scheduler-0" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.667739 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bfefa56f-3099-4f29-b43b-47e1fcb7ee3d-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"bfefa56f-3099-4f29-b43b-47e1fcb7ee3d\") " pod="openstack/cinder-scheduler-0" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.671379 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfefa56f-3099-4f29-b43b-47e1fcb7ee3d-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"bfefa56f-3099-4f29-b43b-47e1fcb7ee3d\") " pod="openstack/cinder-scheduler-0" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.672294 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfefa56f-3099-4f29-b43b-47e1fcb7ee3d-config-data\") pod \"cinder-scheduler-0\" (UID: \"bfefa56f-3099-4f29-b43b-47e1fcb7ee3d\") " pod="openstack/cinder-scheduler-0" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.680188 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79vhb\" (UniqueName: \"kubernetes.io/projected/bfefa56f-3099-4f29-b43b-47e1fcb7ee3d-kube-api-access-79vhb\") pod \"cinder-scheduler-0\" (UID: \"bfefa56f-3099-4f29-b43b-47e1fcb7ee3d\") " pod="openstack/cinder-scheduler-0" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.760768 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/605903a1-6fe5-44e2-b130-349ada640d4b-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"605903a1-6fe5-44e2-b130-349ada640d4b\") " pod="openstack/cinder-api-0" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.760815 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5a5637e5-3258-4eab-9438-479d0aac8099-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-h9zv8\" (UID: \"5a5637e5-3258-4eab-9438-479d0aac8099\") " pod="openstack/dnsmasq-dns-6578955fd5-h9zv8" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.760848 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5a5637e5-3258-4eab-9438-479d0aac8099-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-h9zv8\" (UID: \"5a5637e5-3258-4eab-9438-479d0aac8099\") " pod="openstack/dnsmasq-dns-6578955fd5-h9zv8" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.760868 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5a5637e5-3258-4eab-9438-479d0aac8099-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-h9zv8\" (UID: \"5a5637e5-3258-4eab-9438-479d0aac8099\") " pod="openstack/dnsmasq-dns-6578955fd5-h9zv8" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.760897 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbzws\" (UniqueName: \"kubernetes.io/projected/5a5637e5-3258-4eab-9438-479d0aac8099-kube-api-access-hbzws\") pod \"dnsmasq-dns-6578955fd5-h9zv8\" (UID: \"5a5637e5-3258-4eab-9438-479d0aac8099\") " pod="openstack/dnsmasq-dns-6578955fd5-h9zv8" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.760926 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5a5637e5-3258-4eab-9438-479d0aac8099-dns-svc\") pod \"dnsmasq-dns-6578955fd5-h9zv8\" (UID: \"5a5637e5-3258-4eab-9438-479d0aac8099\") " pod="openstack/dnsmasq-dns-6578955fd5-h9zv8" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.760952 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9hqn\" (UniqueName: \"kubernetes.io/projected/605903a1-6fe5-44e2-b130-349ada640d4b-kube-api-access-t9hqn\") pod \"cinder-api-0\" (UID: \"605903a1-6fe5-44e2-b130-349ada640d4b\") " pod="openstack/cinder-api-0" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.761085 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a5637e5-3258-4eab-9438-479d0aac8099-config\") pod \"dnsmasq-dns-6578955fd5-h9zv8\" (UID: \"5a5637e5-3258-4eab-9438-479d0aac8099\") " pod="openstack/dnsmasq-dns-6578955fd5-h9zv8" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.761135 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/605903a1-6fe5-44e2-b130-349ada640d4b-config-data-custom\") pod \"cinder-api-0\" (UID: \"605903a1-6fe5-44e2-b130-349ada640d4b\") " pod="openstack/cinder-api-0" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.761161 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/605903a1-6fe5-44e2-b130-349ada640d4b-scripts\") pod \"cinder-api-0\" (UID: \"605903a1-6fe5-44e2-b130-349ada640d4b\") " pod="openstack/cinder-api-0" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.761249 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/605903a1-6fe5-44e2-b130-349ada640d4b-logs\") pod \"cinder-api-0\" (UID: \"605903a1-6fe5-44e2-b130-349ada640d4b\") " pod="openstack/cinder-api-0" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.761283 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/605903a1-6fe5-44e2-b130-349ada640d4b-config-data\") pod \"cinder-api-0\" (UID: \"605903a1-6fe5-44e2-b130-349ada640d4b\") " pod="openstack/cinder-api-0" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.761309 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/605903a1-6fe5-44e2-b130-349ada640d4b-etc-machine-id\") pod \"cinder-api-0\" (UID: \"605903a1-6fe5-44e2-b130-349ada640d4b\") " pod="openstack/cinder-api-0" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.863422 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/605903a1-6fe5-44e2-b130-349ada640d4b-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"605903a1-6fe5-44e2-b130-349ada640d4b\") " pod="openstack/cinder-api-0" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.863476 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5a5637e5-3258-4eab-9438-479d0aac8099-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-h9zv8\" (UID: \"5a5637e5-3258-4eab-9438-479d0aac8099\") " pod="openstack/dnsmasq-dns-6578955fd5-h9zv8" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.863514 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5a5637e5-3258-4eab-9438-479d0aac8099-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-h9zv8\" (UID: \"5a5637e5-3258-4eab-9438-479d0aac8099\") " pod="openstack/dnsmasq-dns-6578955fd5-h9zv8" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.863538 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5a5637e5-3258-4eab-9438-479d0aac8099-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-h9zv8\" (UID: \"5a5637e5-3258-4eab-9438-479d0aac8099\") " pod="openstack/dnsmasq-dns-6578955fd5-h9zv8" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.863570 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbzws\" (UniqueName: \"kubernetes.io/projected/5a5637e5-3258-4eab-9438-479d0aac8099-kube-api-access-hbzws\") pod \"dnsmasq-dns-6578955fd5-h9zv8\" (UID: \"5a5637e5-3258-4eab-9438-479d0aac8099\") " pod="openstack/dnsmasq-dns-6578955fd5-h9zv8" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.863604 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5a5637e5-3258-4eab-9438-479d0aac8099-dns-svc\") pod \"dnsmasq-dns-6578955fd5-h9zv8\" (UID: \"5a5637e5-3258-4eab-9438-479d0aac8099\") " pod="openstack/dnsmasq-dns-6578955fd5-h9zv8" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.863630 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9hqn\" (UniqueName: \"kubernetes.io/projected/605903a1-6fe5-44e2-b130-349ada640d4b-kube-api-access-t9hqn\") pod \"cinder-api-0\" (UID: \"605903a1-6fe5-44e2-b130-349ada640d4b\") " pod="openstack/cinder-api-0" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.863662 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a5637e5-3258-4eab-9438-479d0aac8099-config\") pod \"dnsmasq-dns-6578955fd5-h9zv8\" (UID: \"5a5637e5-3258-4eab-9438-479d0aac8099\") " pod="openstack/dnsmasq-dns-6578955fd5-h9zv8" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.863680 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/605903a1-6fe5-44e2-b130-349ada640d4b-config-data-custom\") pod \"cinder-api-0\" (UID: \"605903a1-6fe5-44e2-b130-349ada640d4b\") " pod="openstack/cinder-api-0" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.863696 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/605903a1-6fe5-44e2-b130-349ada640d4b-scripts\") pod \"cinder-api-0\" (UID: \"605903a1-6fe5-44e2-b130-349ada640d4b\") " pod="openstack/cinder-api-0" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.863725 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/605903a1-6fe5-44e2-b130-349ada640d4b-logs\") pod \"cinder-api-0\" (UID: \"605903a1-6fe5-44e2-b130-349ada640d4b\") " pod="openstack/cinder-api-0" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.863743 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/605903a1-6fe5-44e2-b130-349ada640d4b-config-data\") pod \"cinder-api-0\" (UID: \"605903a1-6fe5-44e2-b130-349ada640d4b\") " pod="openstack/cinder-api-0" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.863775 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/605903a1-6fe5-44e2-b130-349ada640d4b-etc-machine-id\") pod \"cinder-api-0\" (UID: \"605903a1-6fe5-44e2-b130-349ada640d4b\") " pod="openstack/cinder-api-0" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.863845 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/605903a1-6fe5-44e2-b130-349ada640d4b-etc-machine-id\") pod \"cinder-api-0\" (UID: \"605903a1-6fe5-44e2-b130-349ada640d4b\") " pod="openstack/cinder-api-0" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.865080 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5a5637e5-3258-4eab-9438-479d0aac8099-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-h9zv8\" (UID: \"5a5637e5-3258-4eab-9438-479d0aac8099\") " pod="openstack/dnsmasq-dns-6578955fd5-h9zv8" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.865712 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5a5637e5-3258-4eab-9438-479d0aac8099-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-h9zv8\" (UID: \"5a5637e5-3258-4eab-9438-479d0aac8099\") " pod="openstack/dnsmasq-dns-6578955fd5-h9zv8" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.866377 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5a5637e5-3258-4eab-9438-479d0aac8099-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-h9zv8\" (UID: \"5a5637e5-3258-4eab-9438-479d0aac8099\") " pod="openstack/dnsmasq-dns-6578955fd5-h9zv8" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.866584 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/605903a1-6fe5-44e2-b130-349ada640d4b-logs\") pod \"cinder-api-0\" (UID: \"605903a1-6fe5-44e2-b130-349ada640d4b\") " pod="openstack/cinder-api-0" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.866909 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5a5637e5-3258-4eab-9438-479d0aac8099-dns-svc\") pod \"dnsmasq-dns-6578955fd5-h9zv8\" (UID: \"5a5637e5-3258-4eab-9438-479d0aac8099\") " pod="openstack/dnsmasq-dns-6578955fd5-h9zv8" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.866909 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a5637e5-3258-4eab-9438-479d0aac8099-config\") pod \"dnsmasq-dns-6578955fd5-h9zv8\" (UID: \"5a5637e5-3258-4eab-9438-479d0aac8099\") " pod="openstack/dnsmasq-dns-6578955fd5-h9zv8" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.872426 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/605903a1-6fe5-44e2-b130-349ada640d4b-config-data-custom\") pod \"cinder-api-0\" (UID: \"605903a1-6fe5-44e2-b130-349ada640d4b\") " pod="openstack/cinder-api-0" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.872530 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/605903a1-6fe5-44e2-b130-349ada640d4b-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"605903a1-6fe5-44e2-b130-349ada640d4b\") " pod="openstack/cinder-api-0" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.872799 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/605903a1-6fe5-44e2-b130-349ada640d4b-config-data\") pod \"cinder-api-0\" (UID: \"605903a1-6fe5-44e2-b130-349ada640d4b\") " pod="openstack/cinder-api-0" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.874244 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/605903a1-6fe5-44e2-b130-349ada640d4b-scripts\") pod \"cinder-api-0\" (UID: \"605903a1-6fe5-44e2-b130-349ada640d4b\") " pod="openstack/cinder-api-0" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.882769 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9hqn\" (UniqueName: \"kubernetes.io/projected/605903a1-6fe5-44e2-b130-349ada640d4b-kube-api-access-t9hqn\") pod \"cinder-api-0\" (UID: \"605903a1-6fe5-44e2-b130-349ada640d4b\") " pod="openstack/cinder-api-0" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.892215 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbzws\" (UniqueName: \"kubernetes.io/projected/5a5637e5-3258-4eab-9438-479d0aac8099-kube-api-access-hbzws\") pod \"dnsmasq-dns-6578955fd5-h9zv8\" (UID: \"5a5637e5-3258-4eab-9438-479d0aac8099\") " pod="openstack/dnsmasq-dns-6578955fd5-h9zv8" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.968636 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 06 03:59:43 crc kubenswrapper[4802]: I1206 03:59:43.988100 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-h9zv8" Dec 06 03:59:44 crc kubenswrapper[4802]: I1206 03:59:44.006177 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 06 03:59:44 crc kubenswrapper[4802]: I1206 03:59:44.165133 4802 generic.go:334] "Generic (PLEG): container finished" podID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerID="04d0056d567a84a69f17a8c41df89172171e22968697aae49e1b7d82534da79c" exitCode=0 Dec 06 03:59:44 crc kubenswrapper[4802]: I1206 03:59:44.165427 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" event={"ID":"6df38316-e0d3-4018-8d27-3620eba3a68d","Type":"ContainerDied","Data":"04d0056d567a84a69f17a8c41df89172171e22968697aae49e1b7d82534da79c"} Dec 06 03:59:44 crc kubenswrapper[4802]: I1206 03:59:44.165455 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" event={"ID":"6df38316-e0d3-4018-8d27-3620eba3a68d","Type":"ContainerStarted","Data":"61a067de7e4bf972bb44f2b14a9f2f9bdc755648ec9a47e6260216803462a3b1"} Dec 06 03:59:44 crc kubenswrapper[4802]: I1206 03:59:44.165470 4802 scope.go:117] "RemoveContainer" containerID="e13987c26ea7eba2b08231bba48c1e97a8d8b6df1eb4749d8e02e423ceb644c9" Dec 06 03:59:44 crc kubenswrapper[4802]: I1206 03:59:44.174535 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-z4x9c" event={"ID":"f91ac3c0-8611-458e-bb4a-cfacb60d8b5b","Type":"ContainerDied","Data":"1ff3e5fe6bf0f9d9b42183731e7c24bb6aaa5a6c9bec17525aef35cd8c01b62d"} Dec 06 03:59:44 crc kubenswrapper[4802]: I1206 03:59:44.174625 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-z4x9c" Dec 06 03:59:44 crc kubenswrapper[4802]: I1206 03:59:44.210922 4802 scope.go:117] "RemoveContainer" containerID="dbc39ed94f5b7975108e7a698f80ef79b3a5290662fc25802033727d945074ef" Dec 06 03:59:44 crc kubenswrapper[4802]: I1206 03:59:44.236032 4802 scope.go:117] "RemoveContainer" containerID="0167fa890261a5700e8a3554578562ee7c6e7304e40de2db2330b755086ae8e9" Dec 06 03:59:44 crc kubenswrapper[4802]: I1206 03:59:44.278926 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-z4x9c"] Dec 06 03:59:44 crc kubenswrapper[4802]: I1206 03:59:44.304655 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-z4x9c"] Dec 06 03:59:44 crc kubenswrapper[4802]: I1206 03:59:44.362393 4802 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6f6b9c98b6-8ndrw" podUID="d098e8a2-f622-44f7-997f-51c6ff7eceaa" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.155:9311/healthcheck\": read tcp 10.217.0.2:43028->10.217.0.155:9311: read: connection reset by peer" Dec 06 03:59:44 crc kubenswrapper[4802]: I1206 03:59:44.362399 4802 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6f6b9c98b6-8ndrw" podUID="d098e8a2-f622-44f7-997f-51c6ff7eceaa" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.155:9311/healthcheck\": read tcp 10.217.0.2:43018->10.217.0.155:9311: read: connection reset by peer" Dec 06 03:59:44 crc kubenswrapper[4802]: I1206 03:59:44.609932 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 03:59:44 crc kubenswrapper[4802]: I1206 03:59:44.983207 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 06 03:59:44 crc kubenswrapper[4802]: W1206 03:59:44.986718 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod605903a1_6fe5_44e2_b130_349ada640d4b.slice/crio-60835912b5ead3b33a8bf12130078fe15fa7ebf37b65533896dd590ff9146321 WatchSource:0}: Error finding container 60835912b5ead3b33a8bf12130078fe15fa7ebf37b65533896dd590ff9146321: Status 404 returned error can't find the container with id 60835912b5ead3b33a8bf12130078fe15fa7ebf37b65533896dd590ff9146321 Dec 06 03:59:45 crc kubenswrapper[4802]: I1206 03:59:45.001639 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-h9zv8"] Dec 06 03:59:45 crc kubenswrapper[4802]: W1206 03:59:45.012839 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5a5637e5_3258_4eab_9438_479d0aac8099.slice/crio-d0dafd87f3f702437aa75029b5742ef894c8f8b0d040c8a67e065c303ed988cf WatchSource:0}: Error finding container d0dafd87f3f702437aa75029b5742ef894c8f8b0d040c8a67e065c303ed988cf: Status 404 returned error can't find the container with id d0dafd87f3f702437aa75029b5742ef894c8f8b0d040c8a67e065c303ed988cf Dec 06 03:59:45 crc kubenswrapper[4802]: I1206 03:59:45.027557 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6f6b9c98b6-8ndrw" Dec 06 03:59:45 crc kubenswrapper[4802]: I1206 03:59:45.191576 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d098e8a2-f622-44f7-997f-51c6ff7eceaa-config-data-custom\") pod \"d098e8a2-f622-44f7-997f-51c6ff7eceaa\" (UID: \"d098e8a2-f622-44f7-997f-51c6ff7eceaa\") " Dec 06 03:59:45 crc kubenswrapper[4802]: I1206 03:59:45.191957 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4fr2d\" (UniqueName: \"kubernetes.io/projected/d098e8a2-f622-44f7-997f-51c6ff7eceaa-kube-api-access-4fr2d\") pod \"d098e8a2-f622-44f7-997f-51c6ff7eceaa\" (UID: \"d098e8a2-f622-44f7-997f-51c6ff7eceaa\") " Dec 06 03:59:45 crc kubenswrapper[4802]: I1206 03:59:45.192022 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d098e8a2-f622-44f7-997f-51c6ff7eceaa-combined-ca-bundle\") pod \"d098e8a2-f622-44f7-997f-51c6ff7eceaa\" (UID: \"d098e8a2-f622-44f7-997f-51c6ff7eceaa\") " Dec 06 03:59:45 crc kubenswrapper[4802]: I1206 03:59:45.192093 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d098e8a2-f622-44f7-997f-51c6ff7eceaa-logs\") pod \"d098e8a2-f622-44f7-997f-51c6ff7eceaa\" (UID: \"d098e8a2-f622-44f7-997f-51c6ff7eceaa\") " Dec 06 03:59:45 crc kubenswrapper[4802]: I1206 03:59:45.193305 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d098e8a2-f622-44f7-997f-51c6ff7eceaa-config-data\") pod \"d098e8a2-f622-44f7-997f-51c6ff7eceaa\" (UID: \"d098e8a2-f622-44f7-997f-51c6ff7eceaa\") " Dec 06 03:59:45 crc kubenswrapper[4802]: I1206 03:59:45.201236 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d098e8a2-f622-44f7-997f-51c6ff7eceaa-logs" (OuterVolumeSpecName: "logs") pod "d098e8a2-f622-44f7-997f-51c6ff7eceaa" (UID: "d098e8a2-f622-44f7-997f-51c6ff7eceaa"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:59:45 crc kubenswrapper[4802]: I1206 03:59:45.214942 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d098e8a2-f622-44f7-997f-51c6ff7eceaa-kube-api-access-4fr2d" (OuterVolumeSpecName: "kube-api-access-4fr2d") pod "d098e8a2-f622-44f7-997f-51c6ff7eceaa" (UID: "d098e8a2-f622-44f7-997f-51c6ff7eceaa"). InnerVolumeSpecName "kube-api-access-4fr2d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:59:45 crc kubenswrapper[4802]: I1206 03:59:45.215676 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d098e8a2-f622-44f7-997f-51c6ff7eceaa-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "d098e8a2-f622-44f7-997f-51c6ff7eceaa" (UID: "d098e8a2-f622-44f7-997f-51c6ff7eceaa"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:59:45 crc kubenswrapper[4802]: I1206 03:59:45.226422 4802 generic.go:334] "Generic (PLEG): container finished" podID="d098e8a2-f622-44f7-997f-51c6ff7eceaa" containerID="748b82913c72c4ebf47e1088833f5e5da1132a05f680ef7cb0ff05f10fe522c5" exitCode=0 Dec 06 03:59:45 crc kubenswrapper[4802]: I1206 03:59:45.226488 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6f6b9c98b6-8ndrw" event={"ID":"d098e8a2-f622-44f7-997f-51c6ff7eceaa","Type":"ContainerDied","Data":"748b82913c72c4ebf47e1088833f5e5da1132a05f680ef7cb0ff05f10fe522c5"} Dec 06 03:59:45 crc kubenswrapper[4802]: I1206 03:59:45.226516 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6f6b9c98b6-8ndrw" event={"ID":"d098e8a2-f622-44f7-997f-51c6ff7eceaa","Type":"ContainerDied","Data":"eccf0d8cb38ce9260b1f4b144eb77a94a99e70245a85f95655db5812158f3d30"} Dec 06 03:59:45 crc kubenswrapper[4802]: I1206 03:59:45.226532 4802 scope.go:117] "RemoveContainer" containerID="748b82913c72c4ebf47e1088833f5e5da1132a05f680ef7cb0ff05f10fe522c5" Dec 06 03:59:45 crc kubenswrapper[4802]: I1206 03:59:45.226591 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6f6b9c98b6-8ndrw" Dec 06 03:59:45 crc kubenswrapper[4802]: I1206 03:59:45.229055 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bfefa56f-3099-4f29-b43b-47e1fcb7ee3d","Type":"ContainerStarted","Data":"6d5d5901c49a822b29593438266b9a9d1998e4d00ab361e84e0cc72ddddfcd0d"} Dec 06 03:59:45 crc kubenswrapper[4802]: I1206 03:59:45.233938 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"605903a1-6fe5-44e2-b130-349ada640d4b","Type":"ContainerStarted","Data":"60835912b5ead3b33a8bf12130078fe15fa7ebf37b65533896dd590ff9146321"} Dec 06 03:59:45 crc kubenswrapper[4802]: I1206 03:59:45.241337 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-h9zv8" event={"ID":"5a5637e5-3258-4eab-9438-479d0aac8099","Type":"ContainerStarted","Data":"d0dafd87f3f702437aa75029b5742ef894c8f8b0d040c8a67e065c303ed988cf"} Dec 06 03:59:45 crc kubenswrapper[4802]: I1206 03:59:45.247087 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d098e8a2-f622-44f7-997f-51c6ff7eceaa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d098e8a2-f622-44f7-997f-51c6ff7eceaa" (UID: "d098e8a2-f622-44f7-997f-51c6ff7eceaa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:59:45 crc kubenswrapper[4802]: I1206 03:59:45.276389 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d098e8a2-f622-44f7-997f-51c6ff7eceaa-config-data" (OuterVolumeSpecName: "config-data") pod "d098e8a2-f622-44f7-997f-51c6ff7eceaa" (UID: "d098e8a2-f622-44f7-997f-51c6ff7eceaa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:59:45 crc kubenswrapper[4802]: I1206 03:59:45.297005 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d098e8a2-f622-44f7-997f-51c6ff7eceaa-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:45 crc kubenswrapper[4802]: I1206 03:59:45.297043 4802 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d098e8a2-f622-44f7-997f-51c6ff7eceaa-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:45 crc kubenswrapper[4802]: I1206 03:59:45.297060 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4fr2d\" (UniqueName: \"kubernetes.io/projected/d098e8a2-f622-44f7-997f-51c6ff7eceaa-kube-api-access-4fr2d\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:45 crc kubenswrapper[4802]: I1206 03:59:45.297071 4802 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d098e8a2-f622-44f7-997f-51c6ff7eceaa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:45 crc kubenswrapper[4802]: I1206 03:59:45.297082 4802 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d098e8a2-f622-44f7-997f-51c6ff7eceaa-logs\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:45 crc kubenswrapper[4802]: I1206 03:59:45.304538 4802 scope.go:117] "RemoveContainer" containerID="dac72a8abf11cd63050015a15e0298021665dc56d9b2efceac05497ee84a7a29" Dec 06 03:59:45 crc kubenswrapper[4802]: I1206 03:59:45.386535 4802 scope.go:117] "RemoveContainer" containerID="748b82913c72c4ebf47e1088833f5e5da1132a05f680ef7cb0ff05f10fe522c5" Dec 06 03:59:45 crc kubenswrapper[4802]: E1206 03:59:45.386991 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"748b82913c72c4ebf47e1088833f5e5da1132a05f680ef7cb0ff05f10fe522c5\": container with ID starting with 748b82913c72c4ebf47e1088833f5e5da1132a05f680ef7cb0ff05f10fe522c5 not found: ID does not exist" containerID="748b82913c72c4ebf47e1088833f5e5da1132a05f680ef7cb0ff05f10fe522c5" Dec 06 03:59:45 crc kubenswrapper[4802]: I1206 03:59:45.387024 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"748b82913c72c4ebf47e1088833f5e5da1132a05f680ef7cb0ff05f10fe522c5"} err="failed to get container status \"748b82913c72c4ebf47e1088833f5e5da1132a05f680ef7cb0ff05f10fe522c5\": rpc error: code = NotFound desc = could not find container \"748b82913c72c4ebf47e1088833f5e5da1132a05f680ef7cb0ff05f10fe522c5\": container with ID starting with 748b82913c72c4ebf47e1088833f5e5da1132a05f680ef7cb0ff05f10fe522c5 not found: ID does not exist" Dec 06 03:59:45 crc kubenswrapper[4802]: I1206 03:59:45.387089 4802 scope.go:117] "RemoveContainer" containerID="dac72a8abf11cd63050015a15e0298021665dc56d9b2efceac05497ee84a7a29" Dec 06 03:59:45 crc kubenswrapper[4802]: E1206 03:59:45.387892 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dac72a8abf11cd63050015a15e0298021665dc56d9b2efceac05497ee84a7a29\": container with ID starting with dac72a8abf11cd63050015a15e0298021665dc56d9b2efceac05497ee84a7a29 not found: ID does not exist" containerID="dac72a8abf11cd63050015a15e0298021665dc56d9b2efceac05497ee84a7a29" Dec 06 03:59:45 crc kubenswrapper[4802]: I1206 03:59:45.387917 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dac72a8abf11cd63050015a15e0298021665dc56d9b2efceac05497ee84a7a29"} err="failed to get container status \"dac72a8abf11cd63050015a15e0298021665dc56d9b2efceac05497ee84a7a29\": rpc error: code = NotFound desc = could not find container \"dac72a8abf11cd63050015a15e0298021665dc56d9b2efceac05497ee84a7a29\": container with ID starting with dac72a8abf11cd63050015a15e0298021665dc56d9b2efceac05497ee84a7a29 not found: ID does not exist" Dec 06 03:59:45 crc kubenswrapper[4802]: I1206 03:59:45.464975 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f91ac3c0-8611-458e-bb4a-cfacb60d8b5b" path="/var/lib/kubelet/pods/f91ac3c0-8611-458e-bb4a-cfacb60d8b5b/volumes" Dec 06 03:59:45 crc kubenswrapper[4802]: I1206 03:59:45.551972 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6f6b9c98b6-8ndrw"] Dec 06 03:59:45 crc kubenswrapper[4802]: I1206 03:59:45.558512 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-6f6b9c98b6-8ndrw"] Dec 06 03:59:46 crc kubenswrapper[4802]: I1206 03:59:46.108735 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 06 03:59:46 crc kubenswrapper[4802]: I1206 03:59:46.255038 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bfefa56f-3099-4f29-b43b-47e1fcb7ee3d","Type":"ContainerStarted","Data":"dcca102ecce1b33a2926573c61c19c6fe74c70e2b98318750c10253d54f15f5c"} Dec 06 03:59:46 crc kubenswrapper[4802]: I1206 03:59:46.256578 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"605903a1-6fe5-44e2-b130-349ada640d4b","Type":"ContainerStarted","Data":"6df92749abe5f9e52931fc9ab886048efd141ab35280059910832d9ca121ccb7"} Dec 06 03:59:46 crc kubenswrapper[4802]: I1206 03:59:46.258379 4802 generic.go:334] "Generic (PLEG): container finished" podID="5a5637e5-3258-4eab-9438-479d0aac8099" containerID="2f8f82847a070c0100fa0db88d5b3b94d65dd663421a778e3015430b4291f86e" exitCode=0 Dec 06 03:59:46 crc kubenswrapper[4802]: I1206 03:59:46.258438 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-h9zv8" event={"ID":"5a5637e5-3258-4eab-9438-479d0aac8099","Type":"ContainerDied","Data":"2f8f82847a070c0100fa0db88d5b3b94d65dd663421a778e3015430b4291f86e"} Dec 06 03:59:47 crc kubenswrapper[4802]: I1206 03:59:47.269279 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"605903a1-6fe5-44e2-b130-349ada640d4b","Type":"ContainerStarted","Data":"34f2790e81e3aaf0fdd499afe55740738d60cb1559efe0946ee78df890368b61"} Dec 06 03:59:47 crc kubenswrapper[4802]: I1206 03:59:47.269916 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 06 03:59:47 crc kubenswrapper[4802]: I1206 03:59:47.269395 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="605903a1-6fe5-44e2-b130-349ada640d4b" containerName="cinder-api" containerID="cri-o://34f2790e81e3aaf0fdd499afe55740738d60cb1559efe0946ee78df890368b61" gracePeriod=30 Dec 06 03:59:47 crc kubenswrapper[4802]: I1206 03:59:47.269368 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="605903a1-6fe5-44e2-b130-349ada640d4b" containerName="cinder-api-log" containerID="cri-o://6df92749abe5f9e52931fc9ab886048efd141ab35280059910832d9ca121ccb7" gracePeriod=30 Dec 06 03:59:47 crc kubenswrapper[4802]: I1206 03:59:47.275411 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-h9zv8" event={"ID":"5a5637e5-3258-4eab-9438-479d0aac8099","Type":"ContainerStarted","Data":"26c2f7af0077ebae0eb9505c78dae937555d12b3ce110a02311e89cbbfae5345"} Dec 06 03:59:47 crc kubenswrapper[4802]: I1206 03:59:47.275557 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6578955fd5-h9zv8" Dec 06 03:59:47 crc kubenswrapper[4802]: I1206 03:59:47.281078 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bfefa56f-3099-4f29-b43b-47e1fcb7ee3d","Type":"ContainerStarted","Data":"44682c5fb47ae56c3da4fdc038c58337805efffb60c98e2e3c8fdba5f7f4afdd"} Dec 06 03:59:47 crc kubenswrapper[4802]: I1206 03:59:47.312926 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.312909133 podStartE2EDuration="4.312909133s" podCreationTimestamp="2025-12-06 03:59:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:59:47.292087027 +0000 UTC m=+1180.163996179" watchObservedRunningTime="2025-12-06 03:59:47.312909133 +0000 UTC m=+1180.184818285" Dec 06 03:59:47 crc kubenswrapper[4802]: I1206 03:59:47.313553 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.595047542 podStartE2EDuration="4.313548031s" podCreationTimestamp="2025-12-06 03:59:43 +0000 UTC" firstStartedPulling="2025-12-06 03:59:44.619945129 +0000 UTC m=+1177.491854281" lastFinishedPulling="2025-12-06 03:59:45.338445618 +0000 UTC m=+1178.210354770" observedRunningTime="2025-12-06 03:59:47.311622769 +0000 UTC m=+1180.183531921" watchObservedRunningTime="2025-12-06 03:59:47.313548031 +0000 UTC m=+1180.185457183" Dec 06 03:59:47 crc kubenswrapper[4802]: I1206 03:59:47.338034 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6578955fd5-h9zv8" podStartSLOduration=4.338013726 podStartE2EDuration="4.338013726s" podCreationTimestamp="2025-12-06 03:59:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:59:47.332927109 +0000 UTC m=+1180.204836281" watchObservedRunningTime="2025-12-06 03:59:47.338013726 +0000 UTC m=+1180.209922888" Dec 06 03:59:47 crc kubenswrapper[4802]: I1206 03:59:47.470666 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d098e8a2-f622-44f7-997f-51c6ff7eceaa" path="/var/lib/kubelet/pods/d098e8a2-f622-44f7-997f-51c6ff7eceaa/volumes" Dec 06 03:59:47 crc kubenswrapper[4802]: I1206 03:59:47.922419 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.046563 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/605903a1-6fe5-44e2-b130-349ada640d4b-combined-ca-bundle\") pod \"605903a1-6fe5-44e2-b130-349ada640d4b\" (UID: \"605903a1-6fe5-44e2-b130-349ada640d4b\") " Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.046722 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/605903a1-6fe5-44e2-b130-349ada640d4b-etc-machine-id\") pod \"605903a1-6fe5-44e2-b130-349ada640d4b\" (UID: \"605903a1-6fe5-44e2-b130-349ada640d4b\") " Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.046793 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/605903a1-6fe5-44e2-b130-349ada640d4b-scripts\") pod \"605903a1-6fe5-44e2-b130-349ada640d4b\" (UID: \"605903a1-6fe5-44e2-b130-349ada640d4b\") " Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.046850 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t9hqn\" (UniqueName: \"kubernetes.io/projected/605903a1-6fe5-44e2-b130-349ada640d4b-kube-api-access-t9hqn\") pod \"605903a1-6fe5-44e2-b130-349ada640d4b\" (UID: \"605903a1-6fe5-44e2-b130-349ada640d4b\") " Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.046908 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/605903a1-6fe5-44e2-b130-349ada640d4b-logs\") pod \"605903a1-6fe5-44e2-b130-349ada640d4b\" (UID: \"605903a1-6fe5-44e2-b130-349ada640d4b\") " Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.046941 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/605903a1-6fe5-44e2-b130-349ada640d4b-config-data\") pod \"605903a1-6fe5-44e2-b130-349ada640d4b\" (UID: \"605903a1-6fe5-44e2-b130-349ada640d4b\") " Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.046969 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/605903a1-6fe5-44e2-b130-349ada640d4b-config-data-custom\") pod \"605903a1-6fe5-44e2-b130-349ada640d4b\" (UID: \"605903a1-6fe5-44e2-b130-349ada640d4b\") " Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.048916 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/605903a1-6fe5-44e2-b130-349ada640d4b-logs" (OuterVolumeSpecName: "logs") pod "605903a1-6fe5-44e2-b130-349ada640d4b" (UID: "605903a1-6fe5-44e2-b130-349ada640d4b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.049169 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/605903a1-6fe5-44e2-b130-349ada640d4b-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "605903a1-6fe5-44e2-b130-349ada640d4b" (UID: "605903a1-6fe5-44e2-b130-349ada640d4b"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.052543 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/605903a1-6fe5-44e2-b130-349ada640d4b-kube-api-access-t9hqn" (OuterVolumeSpecName: "kube-api-access-t9hqn") pod "605903a1-6fe5-44e2-b130-349ada640d4b" (UID: "605903a1-6fe5-44e2-b130-349ada640d4b"). InnerVolumeSpecName "kube-api-access-t9hqn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.054634 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/605903a1-6fe5-44e2-b130-349ada640d4b-scripts" (OuterVolumeSpecName: "scripts") pod "605903a1-6fe5-44e2-b130-349ada640d4b" (UID: "605903a1-6fe5-44e2-b130-349ada640d4b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.056922 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/605903a1-6fe5-44e2-b130-349ada640d4b-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "605903a1-6fe5-44e2-b130-349ada640d4b" (UID: "605903a1-6fe5-44e2-b130-349ada640d4b"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.089575 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/605903a1-6fe5-44e2-b130-349ada640d4b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "605903a1-6fe5-44e2-b130-349ada640d4b" (UID: "605903a1-6fe5-44e2-b130-349ada640d4b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.110067 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/605903a1-6fe5-44e2-b130-349ada640d4b-config-data" (OuterVolumeSpecName: "config-data") pod "605903a1-6fe5-44e2-b130-349ada640d4b" (UID: "605903a1-6fe5-44e2-b130-349ada640d4b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.148437 4802 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/605903a1-6fe5-44e2-b130-349ada640d4b-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.148464 4802 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/605903a1-6fe5-44e2-b130-349ada640d4b-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.148475 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t9hqn\" (UniqueName: \"kubernetes.io/projected/605903a1-6fe5-44e2-b130-349ada640d4b-kube-api-access-t9hqn\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.148487 4802 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/605903a1-6fe5-44e2-b130-349ada640d4b-logs\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.148497 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/605903a1-6fe5-44e2-b130-349ada640d4b-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.148505 4802 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/605903a1-6fe5-44e2-b130-349ada640d4b-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.148513 4802 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/605903a1-6fe5-44e2-b130-349ada640d4b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.290920 4802 generic.go:334] "Generic (PLEG): container finished" podID="605903a1-6fe5-44e2-b130-349ada640d4b" containerID="34f2790e81e3aaf0fdd499afe55740738d60cb1559efe0946ee78df890368b61" exitCode=0 Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.291841 4802 generic.go:334] "Generic (PLEG): container finished" podID="605903a1-6fe5-44e2-b130-349ada640d4b" containerID="6df92749abe5f9e52931fc9ab886048efd141ab35280059910832d9ca121ccb7" exitCode=143 Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.291179 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.291098 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"605903a1-6fe5-44e2-b130-349ada640d4b","Type":"ContainerDied","Data":"34f2790e81e3aaf0fdd499afe55740738d60cb1559efe0946ee78df890368b61"} Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.292035 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"605903a1-6fe5-44e2-b130-349ada640d4b","Type":"ContainerDied","Data":"6df92749abe5f9e52931fc9ab886048efd141ab35280059910832d9ca121ccb7"} Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.292057 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"605903a1-6fe5-44e2-b130-349ada640d4b","Type":"ContainerDied","Data":"60835912b5ead3b33a8bf12130078fe15fa7ebf37b65533896dd590ff9146321"} Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.292077 4802 scope.go:117] "RemoveContainer" containerID="34f2790e81e3aaf0fdd499afe55740738d60cb1559efe0946ee78df890368b61" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.325972 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.332595 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.333977 4802 scope.go:117] "RemoveContainer" containerID="6df92749abe5f9e52931fc9ab886048efd141ab35280059910832d9ca121ccb7" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.352687 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 06 03:59:48 crc kubenswrapper[4802]: E1206 03:59:48.353213 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d098e8a2-f622-44f7-997f-51c6ff7eceaa" containerName="barbican-api-log" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.353287 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="d098e8a2-f622-44f7-997f-51c6ff7eceaa" containerName="barbican-api-log" Dec 06 03:59:48 crc kubenswrapper[4802]: E1206 03:59:48.353382 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="605903a1-6fe5-44e2-b130-349ada640d4b" containerName="cinder-api-log" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.353433 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="605903a1-6fe5-44e2-b130-349ada640d4b" containerName="cinder-api-log" Dec 06 03:59:48 crc kubenswrapper[4802]: E1206 03:59:48.353495 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d098e8a2-f622-44f7-997f-51c6ff7eceaa" containerName="barbican-api" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.353887 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="d098e8a2-f622-44f7-997f-51c6ff7eceaa" containerName="barbican-api" Dec 06 03:59:48 crc kubenswrapper[4802]: E1206 03:59:48.353982 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="605903a1-6fe5-44e2-b130-349ada640d4b" containerName="cinder-api" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.353990 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="605903a1-6fe5-44e2-b130-349ada640d4b" containerName="cinder-api" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.354406 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="d098e8a2-f622-44f7-997f-51c6ff7eceaa" containerName="barbican-api-log" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.354423 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="605903a1-6fe5-44e2-b130-349ada640d4b" containerName="cinder-api-log" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.354432 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="d098e8a2-f622-44f7-997f-51c6ff7eceaa" containerName="barbican-api" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.354451 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="605903a1-6fe5-44e2-b130-349ada640d4b" containerName="cinder-api" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.355586 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.359601 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.359946 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.360146 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.375020 4802 scope.go:117] "RemoveContainer" containerID="34f2790e81e3aaf0fdd499afe55740738d60cb1559efe0946ee78df890368b61" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.375864 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 06 03:59:48 crc kubenswrapper[4802]: E1206 03:59:48.376130 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34f2790e81e3aaf0fdd499afe55740738d60cb1559efe0946ee78df890368b61\": container with ID starting with 34f2790e81e3aaf0fdd499afe55740738d60cb1559efe0946ee78df890368b61 not found: ID does not exist" containerID="34f2790e81e3aaf0fdd499afe55740738d60cb1559efe0946ee78df890368b61" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.376206 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34f2790e81e3aaf0fdd499afe55740738d60cb1559efe0946ee78df890368b61"} err="failed to get container status \"34f2790e81e3aaf0fdd499afe55740738d60cb1559efe0946ee78df890368b61\": rpc error: code = NotFound desc = could not find container \"34f2790e81e3aaf0fdd499afe55740738d60cb1559efe0946ee78df890368b61\": container with ID starting with 34f2790e81e3aaf0fdd499afe55740738d60cb1559efe0946ee78df890368b61 not found: ID does not exist" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.376311 4802 scope.go:117] "RemoveContainer" containerID="6df92749abe5f9e52931fc9ab886048efd141ab35280059910832d9ca121ccb7" Dec 06 03:59:48 crc kubenswrapper[4802]: E1206 03:59:48.376615 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6df92749abe5f9e52931fc9ab886048efd141ab35280059910832d9ca121ccb7\": container with ID starting with 6df92749abe5f9e52931fc9ab886048efd141ab35280059910832d9ca121ccb7 not found: ID does not exist" containerID="6df92749abe5f9e52931fc9ab886048efd141ab35280059910832d9ca121ccb7" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.376688 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6df92749abe5f9e52931fc9ab886048efd141ab35280059910832d9ca121ccb7"} err="failed to get container status \"6df92749abe5f9e52931fc9ab886048efd141ab35280059910832d9ca121ccb7\": rpc error: code = NotFound desc = could not find container \"6df92749abe5f9e52931fc9ab886048efd141ab35280059910832d9ca121ccb7\": container with ID starting with 6df92749abe5f9e52931fc9ab886048efd141ab35280059910832d9ca121ccb7 not found: ID does not exist" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.376769 4802 scope.go:117] "RemoveContainer" containerID="34f2790e81e3aaf0fdd499afe55740738d60cb1559efe0946ee78df890368b61" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.379378 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34f2790e81e3aaf0fdd499afe55740738d60cb1559efe0946ee78df890368b61"} err="failed to get container status \"34f2790e81e3aaf0fdd499afe55740738d60cb1559efe0946ee78df890368b61\": rpc error: code = NotFound desc = could not find container \"34f2790e81e3aaf0fdd499afe55740738d60cb1559efe0946ee78df890368b61\": container with ID starting with 34f2790e81e3aaf0fdd499afe55740738d60cb1559efe0946ee78df890368b61 not found: ID does not exist" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.379443 4802 scope.go:117] "RemoveContainer" containerID="6df92749abe5f9e52931fc9ab886048efd141ab35280059910832d9ca121ccb7" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.379805 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6df92749abe5f9e52931fc9ab886048efd141ab35280059910832d9ca121ccb7"} err="failed to get container status \"6df92749abe5f9e52931fc9ab886048efd141ab35280059910832d9ca121ccb7\": rpc error: code = NotFound desc = could not find container \"6df92749abe5f9e52931fc9ab886048efd141ab35280059910832d9ca121ccb7\": container with ID starting with 6df92749abe5f9e52931fc9ab886048efd141ab35280059910832d9ca121ccb7 not found: ID does not exist" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.453132 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45876a53-aa92-4ca3-bbc2-170e9855b034-config-data\") pod \"cinder-api-0\" (UID: \"45876a53-aa92-4ca3-bbc2-170e9855b034\") " pod="openstack/cinder-api-0" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.453262 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/45876a53-aa92-4ca3-bbc2-170e9855b034-config-data-custom\") pod \"cinder-api-0\" (UID: \"45876a53-aa92-4ca3-bbc2-170e9855b034\") " pod="openstack/cinder-api-0" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.453290 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/45876a53-aa92-4ca3-bbc2-170e9855b034-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"45876a53-aa92-4ca3-bbc2-170e9855b034\") " pod="openstack/cinder-api-0" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.453307 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kttbv\" (UniqueName: \"kubernetes.io/projected/45876a53-aa92-4ca3-bbc2-170e9855b034-kube-api-access-kttbv\") pod \"cinder-api-0\" (UID: \"45876a53-aa92-4ca3-bbc2-170e9855b034\") " pod="openstack/cinder-api-0" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.453328 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45876a53-aa92-4ca3-bbc2-170e9855b034-scripts\") pod \"cinder-api-0\" (UID: \"45876a53-aa92-4ca3-bbc2-170e9855b034\") " pod="openstack/cinder-api-0" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.453381 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/45876a53-aa92-4ca3-bbc2-170e9855b034-etc-machine-id\") pod \"cinder-api-0\" (UID: \"45876a53-aa92-4ca3-bbc2-170e9855b034\") " pod="openstack/cinder-api-0" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.453422 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/45876a53-aa92-4ca3-bbc2-170e9855b034-logs\") pod \"cinder-api-0\" (UID: \"45876a53-aa92-4ca3-bbc2-170e9855b034\") " pod="openstack/cinder-api-0" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.453450 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45876a53-aa92-4ca3-bbc2-170e9855b034-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"45876a53-aa92-4ca3-bbc2-170e9855b034\") " pod="openstack/cinder-api-0" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.453467 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/45876a53-aa92-4ca3-bbc2-170e9855b034-public-tls-certs\") pod \"cinder-api-0\" (UID: \"45876a53-aa92-4ca3-bbc2-170e9855b034\") " pod="openstack/cinder-api-0" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.555364 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/45876a53-aa92-4ca3-bbc2-170e9855b034-logs\") pod \"cinder-api-0\" (UID: \"45876a53-aa92-4ca3-bbc2-170e9855b034\") " pod="openstack/cinder-api-0" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.555710 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45876a53-aa92-4ca3-bbc2-170e9855b034-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"45876a53-aa92-4ca3-bbc2-170e9855b034\") " pod="openstack/cinder-api-0" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.555846 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/45876a53-aa92-4ca3-bbc2-170e9855b034-public-tls-certs\") pod \"cinder-api-0\" (UID: \"45876a53-aa92-4ca3-bbc2-170e9855b034\") " pod="openstack/cinder-api-0" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.555984 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45876a53-aa92-4ca3-bbc2-170e9855b034-config-data\") pod \"cinder-api-0\" (UID: \"45876a53-aa92-4ca3-bbc2-170e9855b034\") " pod="openstack/cinder-api-0" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.556141 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/45876a53-aa92-4ca3-bbc2-170e9855b034-logs\") pod \"cinder-api-0\" (UID: \"45876a53-aa92-4ca3-bbc2-170e9855b034\") " pod="openstack/cinder-api-0" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.556154 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/45876a53-aa92-4ca3-bbc2-170e9855b034-config-data-custom\") pod \"cinder-api-0\" (UID: \"45876a53-aa92-4ca3-bbc2-170e9855b034\") " pod="openstack/cinder-api-0" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.556264 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/45876a53-aa92-4ca3-bbc2-170e9855b034-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"45876a53-aa92-4ca3-bbc2-170e9855b034\") " pod="openstack/cinder-api-0" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.556290 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kttbv\" (UniqueName: \"kubernetes.io/projected/45876a53-aa92-4ca3-bbc2-170e9855b034-kube-api-access-kttbv\") pod \"cinder-api-0\" (UID: \"45876a53-aa92-4ca3-bbc2-170e9855b034\") " pod="openstack/cinder-api-0" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.556331 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45876a53-aa92-4ca3-bbc2-170e9855b034-scripts\") pod \"cinder-api-0\" (UID: \"45876a53-aa92-4ca3-bbc2-170e9855b034\") " pod="openstack/cinder-api-0" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.556428 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/45876a53-aa92-4ca3-bbc2-170e9855b034-etc-machine-id\") pod \"cinder-api-0\" (UID: \"45876a53-aa92-4ca3-bbc2-170e9855b034\") " pod="openstack/cinder-api-0" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.556544 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/45876a53-aa92-4ca3-bbc2-170e9855b034-etc-machine-id\") pod \"cinder-api-0\" (UID: \"45876a53-aa92-4ca3-bbc2-170e9855b034\") " pod="openstack/cinder-api-0" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.559954 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/45876a53-aa92-4ca3-bbc2-170e9855b034-config-data-custom\") pod \"cinder-api-0\" (UID: \"45876a53-aa92-4ca3-bbc2-170e9855b034\") " pod="openstack/cinder-api-0" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.560572 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/45876a53-aa92-4ca3-bbc2-170e9855b034-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"45876a53-aa92-4ca3-bbc2-170e9855b034\") " pod="openstack/cinder-api-0" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.560965 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45876a53-aa92-4ca3-bbc2-170e9855b034-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"45876a53-aa92-4ca3-bbc2-170e9855b034\") " pod="openstack/cinder-api-0" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.561357 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45876a53-aa92-4ca3-bbc2-170e9855b034-scripts\") pod \"cinder-api-0\" (UID: \"45876a53-aa92-4ca3-bbc2-170e9855b034\") " pod="openstack/cinder-api-0" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.561410 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45876a53-aa92-4ca3-bbc2-170e9855b034-config-data\") pod \"cinder-api-0\" (UID: \"45876a53-aa92-4ca3-bbc2-170e9855b034\") " pod="openstack/cinder-api-0" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.561834 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/45876a53-aa92-4ca3-bbc2-170e9855b034-public-tls-certs\") pod \"cinder-api-0\" (UID: \"45876a53-aa92-4ca3-bbc2-170e9855b034\") " pod="openstack/cinder-api-0" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.573784 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kttbv\" (UniqueName: \"kubernetes.io/projected/45876a53-aa92-4ca3-bbc2-170e9855b034-kube-api-access-kttbv\") pod \"cinder-api-0\" (UID: \"45876a53-aa92-4ca3-bbc2-170e9855b034\") " pod="openstack/cinder-api-0" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.687148 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 06 03:59:48 crc kubenswrapper[4802]: I1206 03:59:48.969556 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 06 03:59:49 crc kubenswrapper[4802]: W1206 03:59:49.159585 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod45876a53_aa92_4ca3_bbc2_170e9855b034.slice/crio-e51ce73fec771a89d9a26fe0e0d30c64635eed420141b08e837bfd8a204344f6 WatchSource:0}: Error finding container e51ce73fec771a89d9a26fe0e0d30c64635eed420141b08e837bfd8a204344f6: Status 404 returned error can't find the container with id e51ce73fec771a89d9a26fe0e0d30c64635eed420141b08e837bfd8a204344f6 Dec 06 03:59:49 crc kubenswrapper[4802]: I1206 03:59:49.159746 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 06 03:59:49 crc kubenswrapper[4802]: I1206 03:59:49.306453 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"45876a53-aa92-4ca3-bbc2-170e9855b034","Type":"ContainerStarted","Data":"e51ce73fec771a89d9a26fe0e0d30c64635eed420141b08e837bfd8a204344f6"} Dec 06 03:59:49 crc kubenswrapper[4802]: I1206 03:59:49.464263 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="605903a1-6fe5-44e2-b130-349ada640d4b" path="/var/lib/kubelet/pods/605903a1-6fe5-44e2-b130-349ada640d4b/volumes" Dec 06 03:59:50 crc kubenswrapper[4802]: I1206 03:59:50.321798 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"45876a53-aa92-4ca3-bbc2-170e9855b034","Type":"ContainerStarted","Data":"6238ca80fdbead656b719d515b688c48ecbd6a1ccbd38c0c878a5db0da9cc951"} Dec 06 03:59:50 crc kubenswrapper[4802]: I1206 03:59:50.799193 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-75dc574896-4vrbs" Dec 06 03:59:51 crc kubenswrapper[4802]: I1206 03:59:51.332694 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"45876a53-aa92-4ca3-bbc2-170e9855b034","Type":"ContainerStarted","Data":"f7ab8ccdb8677f0b650cd7574f5c3dfa539610b7732a444ff5d9bcc703adf600"} Dec 06 03:59:51 crc kubenswrapper[4802]: I1206 03:59:51.333052 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 06 03:59:51 crc kubenswrapper[4802]: I1206 03:59:51.349018 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.3490010789999998 podStartE2EDuration="3.349001079s" podCreationTimestamp="2025-12-06 03:59:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:59:51.347870338 +0000 UTC m=+1184.219779500" watchObservedRunningTime="2025-12-06 03:59:51.349001079 +0000 UTC m=+1184.220910241" Dec 06 03:59:51 crc kubenswrapper[4802]: I1206 03:59:51.663714 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 06 03:59:51 crc kubenswrapper[4802]: I1206 03:59:51.664889 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 06 03:59:51 crc kubenswrapper[4802]: I1206 03:59:51.668062 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 06 03:59:51 crc kubenswrapper[4802]: I1206 03:59:51.669804 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-b2rmc" Dec 06 03:59:51 crc kubenswrapper[4802]: I1206 03:59:51.674403 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 06 03:59:51 crc kubenswrapper[4802]: I1206 03:59:51.679709 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 06 03:59:51 crc kubenswrapper[4802]: I1206 03:59:51.822647 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20c275f9-f37c-427a-b8f0-77c35fc880b5-combined-ca-bundle\") pod \"openstackclient\" (UID: \"20c275f9-f37c-427a-b8f0-77c35fc880b5\") " pod="openstack/openstackclient" Dec 06 03:59:51 crc kubenswrapper[4802]: I1206 03:59:51.823026 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zszql\" (UniqueName: \"kubernetes.io/projected/20c275f9-f37c-427a-b8f0-77c35fc880b5-kube-api-access-zszql\") pod \"openstackclient\" (UID: \"20c275f9-f37c-427a-b8f0-77c35fc880b5\") " pod="openstack/openstackclient" Dec 06 03:59:51 crc kubenswrapper[4802]: I1206 03:59:51.823072 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/20c275f9-f37c-427a-b8f0-77c35fc880b5-openstack-config\") pod \"openstackclient\" (UID: \"20c275f9-f37c-427a-b8f0-77c35fc880b5\") " pod="openstack/openstackclient" Dec 06 03:59:51 crc kubenswrapper[4802]: I1206 03:59:51.823178 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/20c275f9-f37c-427a-b8f0-77c35fc880b5-openstack-config-secret\") pod \"openstackclient\" (UID: \"20c275f9-f37c-427a-b8f0-77c35fc880b5\") " pod="openstack/openstackclient" Dec 06 03:59:51 crc kubenswrapper[4802]: I1206 03:59:51.924452 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zszql\" (UniqueName: \"kubernetes.io/projected/20c275f9-f37c-427a-b8f0-77c35fc880b5-kube-api-access-zszql\") pod \"openstackclient\" (UID: \"20c275f9-f37c-427a-b8f0-77c35fc880b5\") " pod="openstack/openstackclient" Dec 06 03:59:51 crc kubenswrapper[4802]: I1206 03:59:51.924524 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/20c275f9-f37c-427a-b8f0-77c35fc880b5-openstack-config\") pod \"openstackclient\" (UID: \"20c275f9-f37c-427a-b8f0-77c35fc880b5\") " pod="openstack/openstackclient" Dec 06 03:59:51 crc kubenswrapper[4802]: I1206 03:59:51.924639 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/20c275f9-f37c-427a-b8f0-77c35fc880b5-openstack-config-secret\") pod \"openstackclient\" (UID: \"20c275f9-f37c-427a-b8f0-77c35fc880b5\") " pod="openstack/openstackclient" Dec 06 03:59:51 crc kubenswrapper[4802]: I1206 03:59:51.924701 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20c275f9-f37c-427a-b8f0-77c35fc880b5-combined-ca-bundle\") pod \"openstackclient\" (UID: \"20c275f9-f37c-427a-b8f0-77c35fc880b5\") " pod="openstack/openstackclient" Dec 06 03:59:51 crc kubenswrapper[4802]: I1206 03:59:51.925893 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/20c275f9-f37c-427a-b8f0-77c35fc880b5-openstack-config\") pod \"openstackclient\" (UID: \"20c275f9-f37c-427a-b8f0-77c35fc880b5\") " pod="openstack/openstackclient" Dec 06 03:59:51 crc kubenswrapper[4802]: I1206 03:59:51.942276 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/20c275f9-f37c-427a-b8f0-77c35fc880b5-openstack-config-secret\") pod \"openstackclient\" (UID: \"20c275f9-f37c-427a-b8f0-77c35fc880b5\") " pod="openstack/openstackclient" Dec 06 03:59:51 crc kubenswrapper[4802]: I1206 03:59:51.942530 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20c275f9-f37c-427a-b8f0-77c35fc880b5-combined-ca-bundle\") pod \"openstackclient\" (UID: \"20c275f9-f37c-427a-b8f0-77c35fc880b5\") " pod="openstack/openstackclient" Dec 06 03:59:51 crc kubenswrapper[4802]: I1206 03:59:51.971402 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zszql\" (UniqueName: \"kubernetes.io/projected/20c275f9-f37c-427a-b8f0-77c35fc880b5-kube-api-access-zszql\") pod \"openstackclient\" (UID: \"20c275f9-f37c-427a-b8f0-77c35fc880b5\") " pod="openstack/openstackclient" Dec 06 03:59:52 crc kubenswrapper[4802]: I1206 03:59:52.002298 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 06 03:59:52 crc kubenswrapper[4802]: I1206 03:59:52.256178 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-788d784b-hjc7w" Dec 06 03:59:52 crc kubenswrapper[4802]: I1206 03:59:52.265911 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-788d784b-hjc7w" Dec 06 03:59:52 crc kubenswrapper[4802]: I1206 03:59:52.481502 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 06 03:59:52 crc kubenswrapper[4802]: W1206 03:59:52.488089 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod20c275f9_f37c_427a_b8f0_77c35fc880b5.slice/crio-5c4fa56efa0671d252f19663086c75ee72dabd77e428f43668505d4f90ae2c32 WatchSource:0}: Error finding container 5c4fa56efa0671d252f19663086c75ee72dabd77e428f43668505d4f90ae2c32: Status 404 returned error can't find the container with id 5c4fa56efa0671d252f19663086c75ee72dabd77e428f43668505d4f90ae2c32 Dec 06 03:59:53 crc kubenswrapper[4802]: I1206 03:59:53.358890 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"20c275f9-f37c-427a-b8f0-77c35fc880b5","Type":"ContainerStarted","Data":"5c4fa56efa0671d252f19663086c75ee72dabd77e428f43668505d4f90ae2c32"} Dec 06 03:59:53 crc kubenswrapper[4802]: I1206 03:59:53.995609 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6578955fd5-h9zv8" Dec 06 03:59:54 crc kubenswrapper[4802]: I1206 03:59:54.069450 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-bq9w6"] Dec 06 03:59:54 crc kubenswrapper[4802]: I1206 03:59:54.070457 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-848cf88cfc-bq9w6" podUID="11fd5eb5-07ce-4218-9f9e-50386ba1c4c4" containerName="dnsmasq-dns" containerID="cri-o://034e8029d69444bbc17bc0698b97b4b46769903bd3005c879ee287e371adea4d" gracePeriod=10 Dec 06 03:59:54 crc kubenswrapper[4802]: I1206 03:59:54.207064 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 06 03:59:54 crc kubenswrapper[4802]: I1206 03:59:54.273663 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 03:59:54 crc kubenswrapper[4802]: I1206 03:59:54.376543 4802 generic.go:334] "Generic (PLEG): container finished" podID="11fd5eb5-07ce-4218-9f9e-50386ba1c4c4" containerID="034e8029d69444bbc17bc0698b97b4b46769903bd3005c879ee287e371adea4d" exitCode=0 Dec 06 03:59:54 crc kubenswrapper[4802]: I1206 03:59:54.376744 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="bfefa56f-3099-4f29-b43b-47e1fcb7ee3d" containerName="cinder-scheduler" containerID="cri-o://dcca102ecce1b33a2926573c61c19c6fe74c70e2b98318750c10253d54f15f5c" gracePeriod=30 Dec 06 03:59:54 crc kubenswrapper[4802]: I1206 03:59:54.377076 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-bq9w6" event={"ID":"11fd5eb5-07ce-4218-9f9e-50386ba1c4c4","Type":"ContainerDied","Data":"034e8029d69444bbc17bc0698b97b4b46769903bd3005c879ee287e371adea4d"} Dec 06 03:59:54 crc kubenswrapper[4802]: I1206 03:59:54.377336 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="bfefa56f-3099-4f29-b43b-47e1fcb7ee3d" containerName="probe" containerID="cri-o://44682c5fb47ae56c3da4fdc038c58337805efffb60c98e2e3c8fdba5f7f4afdd" gracePeriod=30 Dec 06 03:59:54 crc kubenswrapper[4802]: I1206 03:59:54.686049 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-bq9w6" Dec 06 03:59:54 crc kubenswrapper[4802]: I1206 03:59:54.786976 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xqcwv\" (UniqueName: \"kubernetes.io/projected/11fd5eb5-07ce-4218-9f9e-50386ba1c4c4-kube-api-access-xqcwv\") pod \"11fd5eb5-07ce-4218-9f9e-50386ba1c4c4\" (UID: \"11fd5eb5-07ce-4218-9f9e-50386ba1c4c4\") " Dec 06 03:59:54 crc kubenswrapper[4802]: I1206 03:59:54.787070 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/11fd5eb5-07ce-4218-9f9e-50386ba1c4c4-ovsdbserver-nb\") pod \"11fd5eb5-07ce-4218-9f9e-50386ba1c4c4\" (UID: \"11fd5eb5-07ce-4218-9f9e-50386ba1c4c4\") " Dec 06 03:59:54 crc kubenswrapper[4802]: I1206 03:59:54.787101 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11fd5eb5-07ce-4218-9f9e-50386ba1c4c4-config\") pod \"11fd5eb5-07ce-4218-9f9e-50386ba1c4c4\" (UID: \"11fd5eb5-07ce-4218-9f9e-50386ba1c4c4\") " Dec 06 03:59:54 crc kubenswrapper[4802]: I1206 03:59:54.787174 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/11fd5eb5-07ce-4218-9f9e-50386ba1c4c4-ovsdbserver-sb\") pod \"11fd5eb5-07ce-4218-9f9e-50386ba1c4c4\" (UID: \"11fd5eb5-07ce-4218-9f9e-50386ba1c4c4\") " Dec 06 03:59:54 crc kubenswrapper[4802]: I1206 03:59:54.787208 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11fd5eb5-07ce-4218-9f9e-50386ba1c4c4-dns-svc\") pod \"11fd5eb5-07ce-4218-9f9e-50386ba1c4c4\" (UID: \"11fd5eb5-07ce-4218-9f9e-50386ba1c4c4\") " Dec 06 03:59:54 crc kubenswrapper[4802]: I1206 03:59:54.787232 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/11fd5eb5-07ce-4218-9f9e-50386ba1c4c4-dns-swift-storage-0\") pod \"11fd5eb5-07ce-4218-9f9e-50386ba1c4c4\" (UID: \"11fd5eb5-07ce-4218-9f9e-50386ba1c4c4\") " Dec 06 03:59:54 crc kubenswrapper[4802]: I1206 03:59:54.803458 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11fd5eb5-07ce-4218-9f9e-50386ba1c4c4-kube-api-access-xqcwv" (OuterVolumeSpecName: "kube-api-access-xqcwv") pod "11fd5eb5-07ce-4218-9f9e-50386ba1c4c4" (UID: "11fd5eb5-07ce-4218-9f9e-50386ba1c4c4"). InnerVolumeSpecName "kube-api-access-xqcwv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:59:54 crc kubenswrapper[4802]: I1206 03:59:54.852134 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11fd5eb5-07ce-4218-9f9e-50386ba1c4c4-config" (OuterVolumeSpecName: "config") pod "11fd5eb5-07ce-4218-9f9e-50386ba1c4c4" (UID: "11fd5eb5-07ce-4218-9f9e-50386ba1c4c4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:59:54 crc kubenswrapper[4802]: I1206 03:59:54.852572 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11fd5eb5-07ce-4218-9f9e-50386ba1c4c4-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "11fd5eb5-07ce-4218-9f9e-50386ba1c4c4" (UID: "11fd5eb5-07ce-4218-9f9e-50386ba1c4c4"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:59:54 crc kubenswrapper[4802]: I1206 03:59:54.889608 4802 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/11fd5eb5-07ce-4218-9f9e-50386ba1c4c4-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:54 crc kubenswrapper[4802]: I1206 03:59:54.889650 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xqcwv\" (UniqueName: \"kubernetes.io/projected/11fd5eb5-07ce-4218-9f9e-50386ba1c4c4-kube-api-access-xqcwv\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:54 crc kubenswrapper[4802]: I1206 03:59:54.889665 4802 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11fd5eb5-07ce-4218-9f9e-50386ba1c4c4-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:54 crc kubenswrapper[4802]: I1206 03:59:54.893376 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11fd5eb5-07ce-4218-9f9e-50386ba1c4c4-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "11fd5eb5-07ce-4218-9f9e-50386ba1c4c4" (UID: "11fd5eb5-07ce-4218-9f9e-50386ba1c4c4"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:59:54 crc kubenswrapper[4802]: I1206 03:59:54.904598 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11fd5eb5-07ce-4218-9f9e-50386ba1c4c4-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "11fd5eb5-07ce-4218-9f9e-50386ba1c4c4" (UID: "11fd5eb5-07ce-4218-9f9e-50386ba1c4c4"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:59:54 crc kubenswrapper[4802]: I1206 03:59:54.912692 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11fd5eb5-07ce-4218-9f9e-50386ba1c4c4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "11fd5eb5-07ce-4218-9f9e-50386ba1c4c4" (UID: "11fd5eb5-07ce-4218-9f9e-50386ba1c4c4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:59:54 crc kubenswrapper[4802]: I1206 03:59:54.991278 4802 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/11fd5eb5-07ce-4218-9f9e-50386ba1c4c4-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:54 crc kubenswrapper[4802]: I1206 03:59:54.991340 4802 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/11fd5eb5-07ce-4218-9f9e-50386ba1c4c4-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:54 crc kubenswrapper[4802]: I1206 03:59:54.991353 4802 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11fd5eb5-07ce-4218-9f9e-50386ba1c4c4-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 03:59:55 crc kubenswrapper[4802]: I1206 03:59:55.387311 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-bq9w6" Dec 06 03:59:55 crc kubenswrapper[4802]: I1206 03:59:55.387329 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-bq9w6" event={"ID":"11fd5eb5-07ce-4218-9f9e-50386ba1c4c4","Type":"ContainerDied","Data":"084a893e1525aeb0b40f311186ced5db14c99a84519c867217416fedab6dfe09"} Dec 06 03:59:55 crc kubenswrapper[4802]: I1206 03:59:55.390865 4802 scope.go:117] "RemoveContainer" containerID="034e8029d69444bbc17bc0698b97b4b46769903bd3005c879ee287e371adea4d" Dec 06 03:59:55 crc kubenswrapper[4802]: I1206 03:59:55.402564 4802 generic.go:334] "Generic (PLEG): container finished" podID="bfefa56f-3099-4f29-b43b-47e1fcb7ee3d" containerID="44682c5fb47ae56c3da4fdc038c58337805efffb60c98e2e3c8fdba5f7f4afdd" exitCode=0 Dec 06 03:59:55 crc kubenswrapper[4802]: I1206 03:59:55.402610 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bfefa56f-3099-4f29-b43b-47e1fcb7ee3d","Type":"ContainerDied","Data":"44682c5fb47ae56c3da4fdc038c58337805efffb60c98e2e3c8fdba5f7f4afdd"} Dec 06 03:59:55 crc kubenswrapper[4802]: I1206 03:59:55.433522 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-bq9w6"] Dec 06 03:59:55 crc kubenswrapper[4802]: I1206 03:59:55.436985 4802 scope.go:117] "RemoveContainer" containerID="7d4d29cec1b69ef3172164e6917be7c8d7b5288dfcfb522bbdb38d8ae950bfcb" Dec 06 03:59:55 crc kubenswrapper[4802]: I1206 03:59:55.449055 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-bq9w6"] Dec 06 03:59:55 crc kubenswrapper[4802]: I1206 03:59:55.463180 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11fd5eb5-07ce-4218-9f9e-50386ba1c4c4" path="/var/lib/kubelet/pods/11fd5eb5-07ce-4218-9f9e-50386ba1c4c4/volumes" Dec 06 03:59:56 crc kubenswrapper[4802]: I1206 03:59:56.731495 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-686fbdfd7f-sfrn5"] Dec 06 03:59:56 crc kubenswrapper[4802]: E1206 03:59:56.731866 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11fd5eb5-07ce-4218-9f9e-50386ba1c4c4" containerName="init" Dec 06 03:59:56 crc kubenswrapper[4802]: I1206 03:59:56.731878 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="11fd5eb5-07ce-4218-9f9e-50386ba1c4c4" containerName="init" Dec 06 03:59:56 crc kubenswrapper[4802]: E1206 03:59:56.731896 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11fd5eb5-07ce-4218-9f9e-50386ba1c4c4" containerName="dnsmasq-dns" Dec 06 03:59:56 crc kubenswrapper[4802]: I1206 03:59:56.731902 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="11fd5eb5-07ce-4218-9f9e-50386ba1c4c4" containerName="dnsmasq-dns" Dec 06 03:59:56 crc kubenswrapper[4802]: I1206 03:59:56.732088 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="11fd5eb5-07ce-4218-9f9e-50386ba1c4c4" containerName="dnsmasq-dns" Dec 06 03:59:56 crc kubenswrapper[4802]: I1206 03:59:56.733008 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-686fbdfd7f-sfrn5" Dec 06 03:59:56 crc kubenswrapper[4802]: I1206 03:59:56.734786 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Dec 06 03:59:56 crc kubenswrapper[4802]: I1206 03:59:56.734963 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 06 03:59:56 crc kubenswrapper[4802]: I1206 03:59:56.735147 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Dec 06 03:59:56 crc kubenswrapper[4802]: I1206 03:59:56.767792 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-686fbdfd7f-sfrn5"] Dec 06 03:59:56 crc kubenswrapper[4802]: I1206 03:59:56.777514 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-5d96658447-rg8zs"] Dec 06 03:59:56 crc kubenswrapper[4802]: I1206 03:59:56.778894 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-5d96658447-rg8zs" Dec 06 03:59:56 crc kubenswrapper[4802]: I1206 03:59:56.786287 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-engine-config-data" Dec 06 03:59:56 crc kubenswrapper[4802]: I1206 03:59:56.786516 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-kxppf" Dec 06 03:59:56 crc kubenswrapper[4802]: I1206 03:59:56.793683 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-5d96658447-rg8zs"] Dec 06 03:59:56 crc kubenswrapper[4802]: I1206 03:59:56.794954 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Dec 06 03:59:56 crc kubenswrapper[4802]: I1206 03:59:56.832979 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4985762a-1d53-40ad-85da-c6728bdbaa3d-log-httpd\") pod \"swift-proxy-686fbdfd7f-sfrn5\" (UID: \"4985762a-1d53-40ad-85da-c6728bdbaa3d\") " pod="openstack/swift-proxy-686fbdfd7f-sfrn5" Dec 06 03:59:56 crc kubenswrapper[4802]: I1206 03:59:56.833036 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pg9h\" (UniqueName: \"kubernetes.io/projected/4985762a-1d53-40ad-85da-c6728bdbaa3d-kube-api-access-8pg9h\") pod \"swift-proxy-686fbdfd7f-sfrn5\" (UID: \"4985762a-1d53-40ad-85da-c6728bdbaa3d\") " pod="openstack/swift-proxy-686fbdfd7f-sfrn5" Dec 06 03:59:56 crc kubenswrapper[4802]: I1206 03:59:56.833064 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4985762a-1d53-40ad-85da-c6728bdbaa3d-config-data\") pod \"swift-proxy-686fbdfd7f-sfrn5\" (UID: \"4985762a-1d53-40ad-85da-c6728bdbaa3d\") " pod="openstack/swift-proxy-686fbdfd7f-sfrn5" Dec 06 03:59:56 crc kubenswrapper[4802]: I1206 03:59:56.833094 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4985762a-1d53-40ad-85da-c6728bdbaa3d-internal-tls-certs\") pod \"swift-proxy-686fbdfd7f-sfrn5\" (UID: \"4985762a-1d53-40ad-85da-c6728bdbaa3d\") " pod="openstack/swift-proxy-686fbdfd7f-sfrn5" Dec 06 03:59:56 crc kubenswrapper[4802]: I1206 03:59:56.833127 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4985762a-1d53-40ad-85da-c6728bdbaa3d-run-httpd\") pod \"swift-proxy-686fbdfd7f-sfrn5\" (UID: \"4985762a-1d53-40ad-85da-c6728bdbaa3d\") " pod="openstack/swift-proxy-686fbdfd7f-sfrn5" Dec 06 03:59:56 crc kubenswrapper[4802]: I1206 03:59:56.833223 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4985762a-1d53-40ad-85da-c6728bdbaa3d-combined-ca-bundle\") pod \"swift-proxy-686fbdfd7f-sfrn5\" (UID: \"4985762a-1d53-40ad-85da-c6728bdbaa3d\") " pod="openstack/swift-proxy-686fbdfd7f-sfrn5" Dec 06 03:59:56 crc kubenswrapper[4802]: I1206 03:59:56.833270 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4985762a-1d53-40ad-85da-c6728bdbaa3d-etc-swift\") pod \"swift-proxy-686fbdfd7f-sfrn5\" (UID: \"4985762a-1d53-40ad-85da-c6728bdbaa3d\") " pod="openstack/swift-proxy-686fbdfd7f-sfrn5" Dec 06 03:59:56 crc kubenswrapper[4802]: I1206 03:59:56.833308 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4985762a-1d53-40ad-85da-c6728bdbaa3d-public-tls-certs\") pod \"swift-proxy-686fbdfd7f-sfrn5\" (UID: \"4985762a-1d53-40ad-85da-c6728bdbaa3d\") " pod="openstack/swift-proxy-686fbdfd7f-sfrn5" Dec 06 03:59:56 crc kubenswrapper[4802]: I1206 03:59:56.937135 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-688b9f5b49-vxlxf"] Dec 06 03:59:56 crc kubenswrapper[4802]: I1206 03:59:56.937967 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sv9rl\" (UniqueName: \"kubernetes.io/projected/d3da4180-e8fe-4785-8d4f-6cd172fb6399-kube-api-access-sv9rl\") pod \"heat-engine-5d96658447-rg8zs\" (UID: \"d3da4180-e8fe-4785-8d4f-6cd172fb6399\") " pod="openstack/heat-engine-5d96658447-rg8zs" Dec 06 03:59:56 crc kubenswrapper[4802]: I1206 03:59:56.938153 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4985762a-1d53-40ad-85da-c6728bdbaa3d-combined-ca-bundle\") pod \"swift-proxy-686fbdfd7f-sfrn5\" (UID: \"4985762a-1d53-40ad-85da-c6728bdbaa3d\") " pod="openstack/swift-proxy-686fbdfd7f-sfrn5" Dec 06 03:59:56 crc kubenswrapper[4802]: I1206 03:59:56.938294 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4985762a-1d53-40ad-85da-c6728bdbaa3d-etc-swift\") pod \"swift-proxy-686fbdfd7f-sfrn5\" (UID: \"4985762a-1d53-40ad-85da-c6728bdbaa3d\") " pod="openstack/swift-proxy-686fbdfd7f-sfrn5" Dec 06 03:59:56 crc kubenswrapper[4802]: I1206 03:59:56.938407 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4985762a-1d53-40ad-85da-c6728bdbaa3d-public-tls-certs\") pod \"swift-proxy-686fbdfd7f-sfrn5\" (UID: \"4985762a-1d53-40ad-85da-c6728bdbaa3d\") " pod="openstack/swift-proxy-686fbdfd7f-sfrn5" Dec 06 03:59:56 crc kubenswrapper[4802]: I1206 03:59:56.938517 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3da4180-e8fe-4785-8d4f-6cd172fb6399-config-data\") pod \"heat-engine-5d96658447-rg8zs\" (UID: \"d3da4180-e8fe-4785-8d4f-6cd172fb6399\") " pod="openstack/heat-engine-5d96658447-rg8zs" Dec 06 03:59:56 crc kubenswrapper[4802]: I1206 03:59:56.938591 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4985762a-1d53-40ad-85da-c6728bdbaa3d-log-httpd\") pod \"swift-proxy-686fbdfd7f-sfrn5\" (UID: \"4985762a-1d53-40ad-85da-c6728bdbaa3d\") " pod="openstack/swift-proxy-686fbdfd7f-sfrn5" Dec 06 03:59:56 crc kubenswrapper[4802]: I1206 03:59:56.938661 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d3da4180-e8fe-4785-8d4f-6cd172fb6399-config-data-custom\") pod \"heat-engine-5d96658447-rg8zs\" (UID: \"d3da4180-e8fe-4785-8d4f-6cd172fb6399\") " pod="openstack/heat-engine-5d96658447-rg8zs" Dec 06 03:59:56 crc kubenswrapper[4802]: I1206 03:59:56.938732 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pg9h\" (UniqueName: \"kubernetes.io/projected/4985762a-1d53-40ad-85da-c6728bdbaa3d-kube-api-access-8pg9h\") pod \"swift-proxy-686fbdfd7f-sfrn5\" (UID: \"4985762a-1d53-40ad-85da-c6728bdbaa3d\") " pod="openstack/swift-proxy-686fbdfd7f-sfrn5" Dec 06 03:59:56 crc kubenswrapper[4802]: I1206 03:59:56.938829 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4985762a-1d53-40ad-85da-c6728bdbaa3d-config-data\") pod \"swift-proxy-686fbdfd7f-sfrn5\" (UID: \"4985762a-1d53-40ad-85da-c6728bdbaa3d\") " pod="openstack/swift-proxy-686fbdfd7f-sfrn5" Dec 06 03:59:56 crc kubenswrapper[4802]: I1206 03:59:56.938900 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4985762a-1d53-40ad-85da-c6728bdbaa3d-internal-tls-certs\") pod \"swift-proxy-686fbdfd7f-sfrn5\" (UID: \"4985762a-1d53-40ad-85da-c6728bdbaa3d\") " pod="openstack/swift-proxy-686fbdfd7f-sfrn5" Dec 06 03:59:56 crc kubenswrapper[4802]: I1206 03:59:56.938977 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3da4180-e8fe-4785-8d4f-6cd172fb6399-combined-ca-bundle\") pod \"heat-engine-5d96658447-rg8zs\" (UID: \"d3da4180-e8fe-4785-8d4f-6cd172fb6399\") " pod="openstack/heat-engine-5d96658447-rg8zs" Dec 06 03:59:56 crc kubenswrapper[4802]: I1206 03:59:56.939047 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4985762a-1d53-40ad-85da-c6728bdbaa3d-run-httpd\") pod \"swift-proxy-686fbdfd7f-sfrn5\" (UID: \"4985762a-1d53-40ad-85da-c6728bdbaa3d\") " pod="openstack/swift-proxy-686fbdfd7f-sfrn5" Dec 06 03:59:56 crc kubenswrapper[4802]: I1206 03:59:56.939531 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4985762a-1d53-40ad-85da-c6728bdbaa3d-run-httpd\") pod \"swift-proxy-686fbdfd7f-sfrn5\" (UID: \"4985762a-1d53-40ad-85da-c6728bdbaa3d\") " pod="openstack/swift-proxy-686fbdfd7f-sfrn5" Dec 06 03:59:56 crc kubenswrapper[4802]: I1206 03:59:56.943855 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688b9f5b49-vxlxf" Dec 06 03:59:56 crc kubenswrapper[4802]: I1206 03:59:56.950946 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4985762a-1d53-40ad-85da-c6728bdbaa3d-public-tls-certs\") pod \"swift-proxy-686fbdfd7f-sfrn5\" (UID: \"4985762a-1d53-40ad-85da-c6728bdbaa3d\") " pod="openstack/swift-proxy-686fbdfd7f-sfrn5" Dec 06 03:59:56 crc kubenswrapper[4802]: I1206 03:59:56.951984 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4985762a-1d53-40ad-85da-c6728bdbaa3d-config-data\") pod \"swift-proxy-686fbdfd7f-sfrn5\" (UID: \"4985762a-1d53-40ad-85da-c6728bdbaa3d\") " pod="openstack/swift-proxy-686fbdfd7f-sfrn5" Dec 06 03:59:56 crc kubenswrapper[4802]: I1206 03:59:56.952564 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4985762a-1d53-40ad-85da-c6728bdbaa3d-log-httpd\") pod \"swift-proxy-686fbdfd7f-sfrn5\" (UID: \"4985762a-1d53-40ad-85da-c6728bdbaa3d\") " pod="openstack/swift-proxy-686fbdfd7f-sfrn5" Dec 06 03:59:56 crc kubenswrapper[4802]: I1206 03:59:56.959414 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4985762a-1d53-40ad-85da-c6728bdbaa3d-internal-tls-certs\") pod \"swift-proxy-686fbdfd7f-sfrn5\" (UID: \"4985762a-1d53-40ad-85da-c6728bdbaa3d\") " pod="openstack/swift-proxy-686fbdfd7f-sfrn5" Dec 06 03:59:56 crc kubenswrapper[4802]: I1206 03:59:56.959807 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4985762a-1d53-40ad-85da-c6728bdbaa3d-combined-ca-bundle\") pod \"swift-proxy-686fbdfd7f-sfrn5\" (UID: \"4985762a-1d53-40ad-85da-c6728bdbaa3d\") " pod="openstack/swift-proxy-686fbdfd7f-sfrn5" Dec 06 03:59:56 crc kubenswrapper[4802]: I1206 03:59:56.968139 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4985762a-1d53-40ad-85da-c6728bdbaa3d-etc-swift\") pod \"swift-proxy-686fbdfd7f-sfrn5\" (UID: \"4985762a-1d53-40ad-85da-c6728bdbaa3d\") " pod="openstack/swift-proxy-686fbdfd7f-sfrn5" Dec 06 03:59:56 crc kubenswrapper[4802]: I1206 03:59:56.968931 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-688b9f5b49-vxlxf"] Dec 06 03:59:56 crc kubenswrapper[4802]: I1206 03:59:56.985393 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pg9h\" (UniqueName: \"kubernetes.io/projected/4985762a-1d53-40ad-85da-c6728bdbaa3d-kube-api-access-8pg9h\") pod \"swift-proxy-686fbdfd7f-sfrn5\" (UID: \"4985762a-1d53-40ad-85da-c6728bdbaa3d\") " pod="openstack/swift-proxy-686fbdfd7f-sfrn5" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.026005 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-6fc9b98cc8-7swm4"] Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.028099 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6fc9b98cc8-7swm4" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.032554 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-67b6f8b59b-ghkb7"] Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.033530 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-67b6f8b59b-ghkb7" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.044966 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa11f460-8357-4771-9941-0d1682ed0b74-ovsdbserver-sb\") pod \"dnsmasq-dns-688b9f5b49-vxlxf\" (UID: \"aa11f460-8357-4771-9941-0d1682ed0b74\") " pod="openstack/dnsmasq-dns-688b9f5b49-vxlxf" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.045070 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sv9rl\" (UniqueName: \"kubernetes.io/projected/d3da4180-e8fe-4785-8d4f-6cd172fb6399-kube-api-access-sv9rl\") pod \"heat-engine-5d96658447-rg8zs\" (UID: \"d3da4180-e8fe-4785-8d4f-6cd172fb6399\") " pod="openstack/heat-engine-5d96658447-rg8zs" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.045105 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67sbh\" (UniqueName: \"kubernetes.io/projected/aa11f460-8357-4771-9941-0d1682ed0b74-kube-api-access-67sbh\") pod \"dnsmasq-dns-688b9f5b49-vxlxf\" (UID: \"aa11f460-8357-4771-9941-0d1682ed0b74\") " pod="openstack/dnsmasq-dns-688b9f5b49-vxlxf" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.045147 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa11f460-8357-4771-9941-0d1682ed0b74-config\") pod \"dnsmasq-dns-688b9f5b49-vxlxf\" (UID: \"aa11f460-8357-4771-9941-0d1682ed0b74\") " pod="openstack/dnsmasq-dns-688b9f5b49-vxlxf" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.045186 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa11f460-8357-4771-9941-0d1682ed0b74-dns-svc\") pod \"dnsmasq-dns-688b9f5b49-vxlxf\" (UID: \"aa11f460-8357-4771-9941-0d1682ed0b74\") " pod="openstack/dnsmasq-dns-688b9f5b49-vxlxf" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.045264 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3da4180-e8fe-4785-8d4f-6cd172fb6399-config-data\") pod \"heat-engine-5d96658447-rg8zs\" (UID: \"d3da4180-e8fe-4785-8d4f-6cd172fb6399\") " pod="openstack/heat-engine-5d96658447-rg8zs" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.045292 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa11f460-8357-4771-9941-0d1682ed0b74-ovsdbserver-nb\") pod \"dnsmasq-dns-688b9f5b49-vxlxf\" (UID: \"aa11f460-8357-4771-9941-0d1682ed0b74\") " pod="openstack/dnsmasq-dns-688b9f5b49-vxlxf" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.045324 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d3da4180-e8fe-4785-8d4f-6cd172fb6399-config-data-custom\") pod \"heat-engine-5d96658447-rg8zs\" (UID: \"d3da4180-e8fe-4785-8d4f-6cd172fb6399\") " pod="openstack/heat-engine-5d96658447-rg8zs" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.045357 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aa11f460-8357-4771-9941-0d1682ed0b74-dns-swift-storage-0\") pod \"dnsmasq-dns-688b9f5b49-vxlxf\" (UID: \"aa11f460-8357-4771-9941-0d1682ed0b74\") " pod="openstack/dnsmasq-dns-688b9f5b49-vxlxf" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.045383 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3da4180-e8fe-4785-8d4f-6cd172fb6399-combined-ca-bundle\") pod \"heat-engine-5d96658447-rg8zs\" (UID: \"d3da4180-e8fe-4785-8d4f-6cd172fb6399\") " pod="openstack/heat-engine-5d96658447-rg8zs" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.047085 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-api-config-data" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.047649 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-cfnapi-config-data" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.050368 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-686fbdfd7f-sfrn5" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.056628 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d3da4180-e8fe-4785-8d4f-6cd172fb6399-config-data-custom\") pod \"heat-engine-5d96658447-rg8zs\" (UID: \"d3da4180-e8fe-4785-8d4f-6cd172fb6399\") " pod="openstack/heat-engine-5d96658447-rg8zs" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.080078 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3da4180-e8fe-4785-8d4f-6cd172fb6399-combined-ca-bundle\") pod \"heat-engine-5d96658447-rg8zs\" (UID: \"d3da4180-e8fe-4785-8d4f-6cd172fb6399\") " pod="openstack/heat-engine-5d96658447-rg8zs" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.086530 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3da4180-e8fe-4785-8d4f-6cd172fb6399-config-data\") pod \"heat-engine-5d96658447-rg8zs\" (UID: \"d3da4180-e8fe-4785-8d4f-6cd172fb6399\") " pod="openstack/heat-engine-5d96658447-rg8zs" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.086663 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-6fc9b98cc8-7swm4"] Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.138804 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-67b6f8b59b-ghkb7"] Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.155615 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa11f460-8357-4771-9941-0d1682ed0b74-dns-svc\") pod \"dnsmasq-dns-688b9f5b49-vxlxf\" (UID: \"aa11f460-8357-4771-9941-0d1682ed0b74\") " pod="openstack/dnsmasq-dns-688b9f5b49-vxlxf" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.156686 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa11f460-8357-4771-9941-0d1682ed0b74-dns-svc\") pod \"dnsmasq-dns-688b9f5b49-vxlxf\" (UID: \"aa11f460-8357-4771-9941-0d1682ed0b74\") " pod="openstack/dnsmasq-dns-688b9f5b49-vxlxf" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.160554 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sv9rl\" (UniqueName: \"kubernetes.io/projected/d3da4180-e8fe-4785-8d4f-6cd172fb6399-kube-api-access-sv9rl\") pod \"heat-engine-5d96658447-rg8zs\" (UID: \"d3da4180-e8fe-4785-8d4f-6cd172fb6399\") " pod="openstack/heat-engine-5d96658447-rg8zs" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.162657 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa11f460-8357-4771-9941-0d1682ed0b74-ovsdbserver-nb\") pod \"dnsmasq-dns-688b9f5b49-vxlxf\" (UID: \"aa11f460-8357-4771-9941-0d1682ed0b74\") " pod="openstack/dnsmasq-dns-688b9f5b49-vxlxf" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.162965 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thv7m\" (UniqueName: \"kubernetes.io/projected/16c7bcbb-1a4f-4c68-ade9-3cafed6d6876-kube-api-access-thv7m\") pod \"heat-api-6fc9b98cc8-7swm4\" (UID: \"16c7bcbb-1a4f-4c68-ade9-3cafed6d6876\") " pod="openstack/heat-api-6fc9b98cc8-7swm4" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.163050 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a355c0a5-b839-4928-b7ce-0b2d0389190d-config-data-custom\") pod \"heat-cfnapi-67b6f8b59b-ghkb7\" (UID: \"a355c0a5-b839-4928-b7ce-0b2d0389190d\") " pod="openstack/heat-cfnapi-67b6f8b59b-ghkb7" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.163146 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16c7bcbb-1a4f-4c68-ade9-3cafed6d6876-config-data\") pod \"heat-api-6fc9b98cc8-7swm4\" (UID: \"16c7bcbb-1a4f-4c68-ade9-3cafed6d6876\") " pod="openstack/heat-api-6fc9b98cc8-7swm4" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.163234 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aa11f460-8357-4771-9941-0d1682ed0b74-dns-swift-storage-0\") pod \"dnsmasq-dns-688b9f5b49-vxlxf\" (UID: \"aa11f460-8357-4771-9941-0d1682ed0b74\") " pod="openstack/dnsmasq-dns-688b9f5b49-vxlxf" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.163343 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa11f460-8357-4771-9941-0d1682ed0b74-ovsdbserver-nb\") pod \"dnsmasq-dns-688b9f5b49-vxlxf\" (UID: \"aa11f460-8357-4771-9941-0d1682ed0b74\") " pod="openstack/dnsmasq-dns-688b9f5b49-vxlxf" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.163431 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a355c0a5-b839-4928-b7ce-0b2d0389190d-combined-ca-bundle\") pod \"heat-cfnapi-67b6f8b59b-ghkb7\" (UID: \"a355c0a5-b839-4928-b7ce-0b2d0389190d\") " pod="openstack/heat-cfnapi-67b6f8b59b-ghkb7" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.163525 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa11f460-8357-4771-9941-0d1682ed0b74-ovsdbserver-sb\") pod \"dnsmasq-dns-688b9f5b49-vxlxf\" (UID: \"aa11f460-8357-4771-9941-0d1682ed0b74\") " pod="openstack/dnsmasq-dns-688b9f5b49-vxlxf" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.163596 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/16c7bcbb-1a4f-4c68-ade9-3cafed6d6876-config-data-custom\") pod \"heat-api-6fc9b98cc8-7swm4\" (UID: \"16c7bcbb-1a4f-4c68-ade9-3cafed6d6876\") " pod="openstack/heat-api-6fc9b98cc8-7swm4" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.163714 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wl2gt\" (UniqueName: \"kubernetes.io/projected/a355c0a5-b839-4928-b7ce-0b2d0389190d-kube-api-access-wl2gt\") pod \"heat-cfnapi-67b6f8b59b-ghkb7\" (UID: \"a355c0a5-b839-4928-b7ce-0b2d0389190d\") " pod="openstack/heat-cfnapi-67b6f8b59b-ghkb7" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.163831 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67sbh\" (UniqueName: \"kubernetes.io/projected/aa11f460-8357-4771-9941-0d1682ed0b74-kube-api-access-67sbh\") pod \"dnsmasq-dns-688b9f5b49-vxlxf\" (UID: \"aa11f460-8357-4771-9941-0d1682ed0b74\") " pod="openstack/dnsmasq-dns-688b9f5b49-vxlxf" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.163941 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a355c0a5-b839-4928-b7ce-0b2d0389190d-config-data\") pod \"heat-cfnapi-67b6f8b59b-ghkb7\" (UID: \"a355c0a5-b839-4928-b7ce-0b2d0389190d\") " pod="openstack/heat-cfnapi-67b6f8b59b-ghkb7" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.164017 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa11f460-8357-4771-9941-0d1682ed0b74-config\") pod \"dnsmasq-dns-688b9f5b49-vxlxf\" (UID: \"aa11f460-8357-4771-9941-0d1682ed0b74\") " pod="openstack/dnsmasq-dns-688b9f5b49-vxlxf" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.164111 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16c7bcbb-1a4f-4c68-ade9-3cafed6d6876-combined-ca-bundle\") pod \"heat-api-6fc9b98cc8-7swm4\" (UID: \"16c7bcbb-1a4f-4c68-ade9-3cafed6d6876\") " pod="openstack/heat-api-6fc9b98cc8-7swm4" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.164271 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aa11f460-8357-4771-9941-0d1682ed0b74-dns-swift-storage-0\") pod \"dnsmasq-dns-688b9f5b49-vxlxf\" (UID: \"aa11f460-8357-4771-9941-0d1682ed0b74\") " pod="openstack/dnsmasq-dns-688b9f5b49-vxlxf" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.166288 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa11f460-8357-4771-9941-0d1682ed0b74-config\") pod \"dnsmasq-dns-688b9f5b49-vxlxf\" (UID: \"aa11f460-8357-4771-9941-0d1682ed0b74\") " pod="openstack/dnsmasq-dns-688b9f5b49-vxlxf" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.168507 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa11f460-8357-4771-9941-0d1682ed0b74-ovsdbserver-sb\") pod \"dnsmasq-dns-688b9f5b49-vxlxf\" (UID: \"aa11f460-8357-4771-9941-0d1682ed0b74\") " pod="openstack/dnsmasq-dns-688b9f5b49-vxlxf" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.211364 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67sbh\" (UniqueName: \"kubernetes.io/projected/aa11f460-8357-4771-9941-0d1682ed0b74-kube-api-access-67sbh\") pod \"dnsmasq-dns-688b9f5b49-vxlxf\" (UID: \"aa11f460-8357-4771-9941-0d1682ed0b74\") " pod="openstack/dnsmasq-dns-688b9f5b49-vxlxf" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.266303 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thv7m\" (UniqueName: \"kubernetes.io/projected/16c7bcbb-1a4f-4c68-ade9-3cafed6d6876-kube-api-access-thv7m\") pod \"heat-api-6fc9b98cc8-7swm4\" (UID: \"16c7bcbb-1a4f-4c68-ade9-3cafed6d6876\") " pod="openstack/heat-api-6fc9b98cc8-7swm4" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.268047 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a355c0a5-b839-4928-b7ce-0b2d0389190d-config-data-custom\") pod \"heat-cfnapi-67b6f8b59b-ghkb7\" (UID: \"a355c0a5-b839-4928-b7ce-0b2d0389190d\") " pod="openstack/heat-cfnapi-67b6f8b59b-ghkb7" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.268175 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16c7bcbb-1a4f-4c68-ade9-3cafed6d6876-config-data\") pod \"heat-api-6fc9b98cc8-7swm4\" (UID: \"16c7bcbb-1a4f-4c68-ade9-3cafed6d6876\") " pod="openstack/heat-api-6fc9b98cc8-7swm4" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.268326 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a355c0a5-b839-4928-b7ce-0b2d0389190d-combined-ca-bundle\") pod \"heat-cfnapi-67b6f8b59b-ghkb7\" (UID: \"a355c0a5-b839-4928-b7ce-0b2d0389190d\") " pod="openstack/heat-cfnapi-67b6f8b59b-ghkb7" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.268448 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/16c7bcbb-1a4f-4c68-ade9-3cafed6d6876-config-data-custom\") pod \"heat-api-6fc9b98cc8-7swm4\" (UID: \"16c7bcbb-1a4f-4c68-ade9-3cafed6d6876\") " pod="openstack/heat-api-6fc9b98cc8-7swm4" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.268570 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wl2gt\" (UniqueName: \"kubernetes.io/projected/a355c0a5-b839-4928-b7ce-0b2d0389190d-kube-api-access-wl2gt\") pod \"heat-cfnapi-67b6f8b59b-ghkb7\" (UID: \"a355c0a5-b839-4928-b7ce-0b2d0389190d\") " pod="openstack/heat-cfnapi-67b6f8b59b-ghkb7" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.268708 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a355c0a5-b839-4928-b7ce-0b2d0389190d-config-data\") pod \"heat-cfnapi-67b6f8b59b-ghkb7\" (UID: \"a355c0a5-b839-4928-b7ce-0b2d0389190d\") " pod="openstack/heat-cfnapi-67b6f8b59b-ghkb7" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.268923 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16c7bcbb-1a4f-4c68-ade9-3cafed6d6876-combined-ca-bundle\") pod \"heat-api-6fc9b98cc8-7swm4\" (UID: \"16c7bcbb-1a4f-4c68-ade9-3cafed6d6876\") " pod="openstack/heat-api-6fc9b98cc8-7swm4" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.281096 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a355c0a5-b839-4928-b7ce-0b2d0389190d-config-data-custom\") pod \"heat-cfnapi-67b6f8b59b-ghkb7\" (UID: \"a355c0a5-b839-4928-b7ce-0b2d0389190d\") " pod="openstack/heat-cfnapi-67b6f8b59b-ghkb7" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.281860 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a355c0a5-b839-4928-b7ce-0b2d0389190d-combined-ca-bundle\") pod \"heat-cfnapi-67b6f8b59b-ghkb7\" (UID: \"a355c0a5-b839-4928-b7ce-0b2d0389190d\") " pod="openstack/heat-cfnapi-67b6f8b59b-ghkb7" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.284091 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a355c0a5-b839-4928-b7ce-0b2d0389190d-config-data\") pod \"heat-cfnapi-67b6f8b59b-ghkb7\" (UID: \"a355c0a5-b839-4928-b7ce-0b2d0389190d\") " pod="openstack/heat-cfnapi-67b6f8b59b-ghkb7" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.291397 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16c7bcbb-1a4f-4c68-ade9-3cafed6d6876-combined-ca-bundle\") pod \"heat-api-6fc9b98cc8-7swm4\" (UID: \"16c7bcbb-1a4f-4c68-ade9-3cafed6d6876\") " pod="openstack/heat-api-6fc9b98cc8-7swm4" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.293415 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/16c7bcbb-1a4f-4c68-ade9-3cafed6d6876-config-data-custom\") pod \"heat-api-6fc9b98cc8-7swm4\" (UID: \"16c7bcbb-1a4f-4c68-ade9-3cafed6d6876\") " pod="openstack/heat-api-6fc9b98cc8-7swm4" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.296211 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16c7bcbb-1a4f-4c68-ade9-3cafed6d6876-config-data\") pod \"heat-api-6fc9b98cc8-7swm4\" (UID: \"16c7bcbb-1a4f-4c68-ade9-3cafed6d6876\") " pod="openstack/heat-api-6fc9b98cc8-7swm4" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.298276 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wl2gt\" (UniqueName: \"kubernetes.io/projected/a355c0a5-b839-4928-b7ce-0b2d0389190d-kube-api-access-wl2gt\") pod \"heat-cfnapi-67b6f8b59b-ghkb7\" (UID: \"a355c0a5-b839-4928-b7ce-0b2d0389190d\") " pod="openstack/heat-cfnapi-67b6f8b59b-ghkb7" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.301418 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thv7m\" (UniqueName: \"kubernetes.io/projected/16c7bcbb-1a4f-4c68-ade9-3cafed6d6876-kube-api-access-thv7m\") pod \"heat-api-6fc9b98cc8-7swm4\" (UID: \"16c7bcbb-1a4f-4c68-ade9-3cafed6d6876\") " pod="openstack/heat-api-6fc9b98cc8-7swm4" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.438796 4802 generic.go:334] "Generic (PLEG): container finished" podID="bfefa56f-3099-4f29-b43b-47e1fcb7ee3d" containerID="dcca102ecce1b33a2926573c61c19c6fe74c70e2b98318750c10253d54f15f5c" exitCode=0 Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.438842 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bfefa56f-3099-4f29-b43b-47e1fcb7ee3d","Type":"ContainerDied","Data":"dcca102ecce1b33a2926573c61c19c6fe74c70e2b98318750c10253d54f15f5c"} Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.452327 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688b9f5b49-vxlxf" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.453003 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-5d96658447-rg8zs" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.522628 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6fc9b98cc8-7swm4" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.562393 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-67b6f8b59b-ghkb7" Dec 06 03:59:57 crc kubenswrapper[4802]: I1206 03:59:57.869378 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-686fbdfd7f-sfrn5"] Dec 06 03:59:59 crc kubenswrapper[4802]: I1206 03:59:59.019999 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 03:59:59 crc kubenswrapper[4802]: I1206 03:59:59.020289 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="f8c7901e-3653-4d82-a15f-dd12cf4ab8c5" containerName="glance-log" containerID="cri-o://bd7d4fa9d3441f11c0fea392f92f249b999c25be443db3a969902839362960b2" gracePeriod=30 Dec 06 03:59:59 crc kubenswrapper[4802]: I1206 03:59:59.020442 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="f8c7901e-3653-4d82-a15f-dd12cf4ab8c5" containerName="glance-httpd" containerID="cri-o://43393bb092a35409606f84f7092e94a2e0eec77e324a88aed2126f8d32b543b1" gracePeriod=30 Dec 06 03:59:59 crc kubenswrapper[4802]: I1206 03:59:59.482694 4802 generic.go:334] "Generic (PLEG): container finished" podID="f8c7901e-3653-4d82-a15f-dd12cf4ab8c5" containerID="bd7d4fa9d3441f11c0fea392f92f249b999c25be443db3a969902839362960b2" exitCode=143 Dec 06 03:59:59 crc kubenswrapper[4802]: I1206 03:59:59.482737 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f8c7901e-3653-4d82-a15f-dd12cf4ab8c5","Type":"ContainerDied","Data":"bd7d4fa9d3441f11c0fea392f92f249b999c25be443db3a969902839362960b2"} Dec 06 04:00:00 crc kubenswrapper[4802]: I1206 04:00:00.135449 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416560-zzktw"] Dec 06 04:00:00 crc kubenswrapper[4802]: I1206 04:00:00.136529 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416560-zzktw" Dec 06 04:00:00 crc kubenswrapper[4802]: I1206 04:00:00.139274 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 04:00:00 crc kubenswrapper[4802]: I1206 04:00:00.140382 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 04:00:00 crc kubenswrapper[4802]: I1206 04:00:00.148226 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416560-zzktw"] Dec 06 04:00:00 crc kubenswrapper[4802]: I1206 04:00:00.225167 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6qgq\" (UniqueName: \"kubernetes.io/projected/5898b86d-3906-4c87-8509-8c37fe50f544-kube-api-access-h6qgq\") pod \"collect-profiles-29416560-zzktw\" (UID: \"5898b86d-3906-4c87-8509-8c37fe50f544\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416560-zzktw" Dec 06 04:00:00 crc kubenswrapper[4802]: I1206 04:00:00.225302 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5898b86d-3906-4c87-8509-8c37fe50f544-config-volume\") pod \"collect-profiles-29416560-zzktw\" (UID: \"5898b86d-3906-4c87-8509-8c37fe50f544\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416560-zzktw" Dec 06 04:00:00 crc kubenswrapper[4802]: I1206 04:00:00.225693 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5898b86d-3906-4c87-8509-8c37fe50f544-secret-volume\") pod \"collect-profiles-29416560-zzktw\" (UID: \"5898b86d-3906-4c87-8509-8c37fe50f544\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416560-zzktw" Dec 06 04:00:00 crc kubenswrapper[4802]: I1206 04:00:00.329356 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5898b86d-3906-4c87-8509-8c37fe50f544-secret-volume\") pod \"collect-profiles-29416560-zzktw\" (UID: \"5898b86d-3906-4c87-8509-8c37fe50f544\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416560-zzktw" Dec 06 04:00:00 crc kubenswrapper[4802]: I1206 04:00:00.329455 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6qgq\" (UniqueName: \"kubernetes.io/projected/5898b86d-3906-4c87-8509-8c37fe50f544-kube-api-access-h6qgq\") pod \"collect-profiles-29416560-zzktw\" (UID: \"5898b86d-3906-4c87-8509-8c37fe50f544\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416560-zzktw" Dec 06 04:00:00 crc kubenswrapper[4802]: I1206 04:00:00.329504 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5898b86d-3906-4c87-8509-8c37fe50f544-config-volume\") pod \"collect-profiles-29416560-zzktw\" (UID: \"5898b86d-3906-4c87-8509-8c37fe50f544\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416560-zzktw" Dec 06 04:00:00 crc kubenswrapper[4802]: I1206 04:00:00.330870 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5898b86d-3906-4c87-8509-8c37fe50f544-config-volume\") pod \"collect-profiles-29416560-zzktw\" (UID: \"5898b86d-3906-4c87-8509-8c37fe50f544\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416560-zzktw" Dec 06 04:00:00 crc kubenswrapper[4802]: I1206 04:00:00.336219 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5898b86d-3906-4c87-8509-8c37fe50f544-secret-volume\") pod \"collect-profiles-29416560-zzktw\" (UID: \"5898b86d-3906-4c87-8509-8c37fe50f544\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416560-zzktw" Dec 06 04:00:00 crc kubenswrapper[4802]: I1206 04:00:00.347980 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6qgq\" (UniqueName: \"kubernetes.io/projected/5898b86d-3906-4c87-8509-8c37fe50f544-kube-api-access-h6qgq\") pod \"collect-profiles-29416560-zzktw\" (UID: \"5898b86d-3906-4c87-8509-8c37fe50f544\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416560-zzktw" Dec 06 04:00:00 crc kubenswrapper[4802]: I1206 04:00:00.461036 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416560-zzktw" Dec 06 04:00:00 crc kubenswrapper[4802]: I1206 04:00:00.708977 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 04:00:00 crc kubenswrapper[4802]: I1206 04:00:00.709224 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="22294441-50c6-4dc0-80f1-4b14c1fc62a9" containerName="glance-log" containerID="cri-o://c6a619cd1b3bf4fef4f5bd45c896b6e8b20d75cfc21ef84bac1fb69e9b572e9a" gracePeriod=30 Dec 06 04:00:00 crc kubenswrapper[4802]: I1206 04:00:00.709327 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="22294441-50c6-4dc0-80f1-4b14c1fc62a9" containerName="glance-httpd" containerID="cri-o://90662e183b0e95670b64555522541d9155b259f309bcf2d1939c8051e4b9fb83" gracePeriod=30 Dec 06 04:00:00 crc kubenswrapper[4802]: I1206 04:00:00.843711 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 06 04:00:01 crc kubenswrapper[4802]: I1206 04:00:01.512357 4802 generic.go:334] "Generic (PLEG): container finished" podID="4d4f4fff-a6ec-4fef-82b4-6ba392fd297c" containerID="ec333f217f947fe8e066593d0ae87d8fd22fa9027bc4f3bc00f5f9ae7a30b116" exitCode=137 Dec 06 04:00:01 crc kubenswrapper[4802]: I1206 04:00:01.512675 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4d4f4fff-a6ec-4fef-82b4-6ba392fd297c","Type":"ContainerDied","Data":"ec333f217f947fe8e066593d0ae87d8fd22fa9027bc4f3bc00f5f9ae7a30b116"} Dec 06 04:00:01 crc kubenswrapper[4802]: I1206 04:00:01.523131 4802 generic.go:334] "Generic (PLEG): container finished" podID="22294441-50c6-4dc0-80f1-4b14c1fc62a9" containerID="c6a619cd1b3bf4fef4f5bd45c896b6e8b20d75cfc21ef84bac1fb69e9b572e9a" exitCode=143 Dec 06 04:00:01 crc kubenswrapper[4802]: I1206 04:00:01.523179 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"22294441-50c6-4dc0-80f1-4b14c1fc62a9","Type":"ContainerDied","Data":"c6a619cd1b3bf4fef4f5bd45c896b6e8b20d75cfc21ef84bac1fb69e9b572e9a"} Dec 06 04:00:02 crc kubenswrapper[4802]: I1206 04:00:02.537989 4802 generic.go:334] "Generic (PLEG): container finished" podID="f8c7901e-3653-4d82-a15f-dd12cf4ab8c5" containerID="43393bb092a35409606f84f7092e94a2e0eec77e324a88aed2126f8d32b543b1" exitCode=0 Dec 06 04:00:02 crc kubenswrapper[4802]: I1206 04:00:02.538286 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f8c7901e-3653-4d82-a15f-dd12cf4ab8c5","Type":"ContainerDied","Data":"43393bb092a35409606f84f7092e94a2e0eec77e324a88aed2126f8d32b543b1"} Dec 06 04:00:02 crc kubenswrapper[4802]: I1206 04:00:02.838285 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-844df74668-r698z" Dec 06 04:00:03 crc kubenswrapper[4802]: I1206 04:00:03.868926 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-55fc64bfdd-779rm"] Dec 06 04:00:03 crc kubenswrapper[4802]: I1206 04:00:03.871203 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-55fc64bfdd-779rm" Dec 06 04:00:03 crc kubenswrapper[4802]: I1206 04:00:03.919464 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-55fc64bfdd-779rm"] Dec 06 04:00:03 crc kubenswrapper[4802]: I1206 04:00:03.955973 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-798947df95-gjr2t"] Dec 06 04:00:03 crc kubenswrapper[4802]: I1206 04:00:03.957598 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-798947df95-gjr2t" Dec 06 04:00:03 crc kubenswrapper[4802]: I1206 04:00:03.985965 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-798947df95-gjr2t"] Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.000256 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a86ea592-5d80-4415-8c1f-5471274999d4-combined-ca-bundle\") pod \"heat-engine-55fc64bfdd-779rm\" (UID: \"a86ea592-5d80-4415-8c1f-5471274999d4\") " pod="openstack/heat-engine-55fc64bfdd-779rm" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.000399 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eda29ee7-1690-43ed-be0f-c8aac34d6daa-config-data\") pod \"heat-api-798947df95-gjr2t\" (UID: \"eda29ee7-1690-43ed-be0f-c8aac34d6daa\") " pod="openstack/heat-api-798947df95-gjr2t" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.000423 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eda29ee7-1690-43ed-be0f-c8aac34d6daa-config-data-custom\") pod \"heat-api-798947df95-gjr2t\" (UID: \"eda29ee7-1690-43ed-be0f-c8aac34d6daa\") " pod="openstack/heat-api-798947df95-gjr2t" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.000458 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eda29ee7-1690-43ed-be0f-c8aac34d6daa-combined-ca-bundle\") pod \"heat-api-798947df95-gjr2t\" (UID: \"eda29ee7-1690-43ed-be0f-c8aac34d6daa\") " pod="openstack/heat-api-798947df95-gjr2t" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.000475 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a86ea592-5d80-4415-8c1f-5471274999d4-config-data\") pod \"heat-engine-55fc64bfdd-779rm\" (UID: \"a86ea592-5d80-4415-8c1f-5471274999d4\") " pod="openstack/heat-engine-55fc64bfdd-779rm" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.000523 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jkfv\" (UniqueName: \"kubernetes.io/projected/eda29ee7-1690-43ed-be0f-c8aac34d6daa-kube-api-access-6jkfv\") pod \"heat-api-798947df95-gjr2t\" (UID: \"eda29ee7-1690-43ed-be0f-c8aac34d6daa\") " pod="openstack/heat-api-798947df95-gjr2t" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.000557 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vhbg\" (UniqueName: \"kubernetes.io/projected/a86ea592-5d80-4415-8c1f-5471274999d4-kube-api-access-4vhbg\") pod \"heat-engine-55fc64bfdd-779rm\" (UID: \"a86ea592-5d80-4415-8c1f-5471274999d4\") " pod="openstack/heat-engine-55fc64bfdd-779rm" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.000605 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a86ea592-5d80-4415-8c1f-5471274999d4-config-data-custom\") pod \"heat-engine-55fc64bfdd-779rm\" (UID: \"a86ea592-5d80-4415-8c1f-5471274999d4\") " pod="openstack/heat-engine-55fc64bfdd-779rm" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.021648 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-7c696dc874-7894k"] Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.023059 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7c696dc874-7894k" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.028516 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-7c696dc874-7894k"] Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.102785 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5bzt\" (UniqueName: \"kubernetes.io/projected/246893e5-50e3-402e-8b4f-f21b31af732e-kube-api-access-s5bzt\") pod \"heat-cfnapi-7c696dc874-7894k\" (UID: \"246893e5-50e3-402e-8b4f-f21b31af732e\") " pod="openstack/heat-cfnapi-7c696dc874-7894k" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.102845 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eda29ee7-1690-43ed-be0f-c8aac34d6daa-config-data\") pod \"heat-api-798947df95-gjr2t\" (UID: \"eda29ee7-1690-43ed-be0f-c8aac34d6daa\") " pod="openstack/heat-api-798947df95-gjr2t" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.102874 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eda29ee7-1690-43ed-be0f-c8aac34d6daa-config-data-custom\") pod \"heat-api-798947df95-gjr2t\" (UID: \"eda29ee7-1690-43ed-be0f-c8aac34d6daa\") " pod="openstack/heat-api-798947df95-gjr2t" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.102916 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eda29ee7-1690-43ed-be0f-c8aac34d6daa-combined-ca-bundle\") pod \"heat-api-798947df95-gjr2t\" (UID: \"eda29ee7-1690-43ed-be0f-c8aac34d6daa\") " pod="openstack/heat-api-798947df95-gjr2t" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.102939 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a86ea592-5d80-4415-8c1f-5471274999d4-config-data\") pod \"heat-engine-55fc64bfdd-779rm\" (UID: \"a86ea592-5d80-4415-8c1f-5471274999d4\") " pod="openstack/heat-engine-55fc64bfdd-779rm" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.102981 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/246893e5-50e3-402e-8b4f-f21b31af732e-combined-ca-bundle\") pod \"heat-cfnapi-7c696dc874-7894k\" (UID: \"246893e5-50e3-402e-8b4f-f21b31af732e\") " pod="openstack/heat-cfnapi-7c696dc874-7894k" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.103016 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jkfv\" (UniqueName: \"kubernetes.io/projected/eda29ee7-1690-43ed-be0f-c8aac34d6daa-kube-api-access-6jkfv\") pod \"heat-api-798947df95-gjr2t\" (UID: \"eda29ee7-1690-43ed-be0f-c8aac34d6daa\") " pod="openstack/heat-api-798947df95-gjr2t" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.103051 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/246893e5-50e3-402e-8b4f-f21b31af732e-config-data-custom\") pod \"heat-cfnapi-7c696dc874-7894k\" (UID: \"246893e5-50e3-402e-8b4f-f21b31af732e\") " pod="openstack/heat-cfnapi-7c696dc874-7894k" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.103073 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/246893e5-50e3-402e-8b4f-f21b31af732e-config-data\") pod \"heat-cfnapi-7c696dc874-7894k\" (UID: \"246893e5-50e3-402e-8b4f-f21b31af732e\") " pod="openstack/heat-cfnapi-7c696dc874-7894k" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.103106 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vhbg\" (UniqueName: \"kubernetes.io/projected/a86ea592-5d80-4415-8c1f-5471274999d4-kube-api-access-4vhbg\") pod \"heat-engine-55fc64bfdd-779rm\" (UID: \"a86ea592-5d80-4415-8c1f-5471274999d4\") " pod="openstack/heat-engine-55fc64bfdd-779rm" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.103160 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a86ea592-5d80-4415-8c1f-5471274999d4-config-data-custom\") pod \"heat-engine-55fc64bfdd-779rm\" (UID: \"a86ea592-5d80-4415-8c1f-5471274999d4\") " pod="openstack/heat-engine-55fc64bfdd-779rm" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.103245 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a86ea592-5d80-4415-8c1f-5471274999d4-combined-ca-bundle\") pod \"heat-engine-55fc64bfdd-779rm\" (UID: \"a86ea592-5d80-4415-8c1f-5471274999d4\") " pod="openstack/heat-engine-55fc64bfdd-779rm" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.113704 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a86ea592-5d80-4415-8c1f-5471274999d4-config-data\") pod \"heat-engine-55fc64bfdd-779rm\" (UID: \"a86ea592-5d80-4415-8c1f-5471274999d4\") " pod="openstack/heat-engine-55fc64bfdd-779rm" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.113786 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eda29ee7-1690-43ed-be0f-c8aac34d6daa-config-data-custom\") pod \"heat-api-798947df95-gjr2t\" (UID: \"eda29ee7-1690-43ed-be0f-c8aac34d6daa\") " pod="openstack/heat-api-798947df95-gjr2t" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.113983 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a86ea592-5d80-4415-8c1f-5471274999d4-combined-ca-bundle\") pod \"heat-engine-55fc64bfdd-779rm\" (UID: \"a86ea592-5d80-4415-8c1f-5471274999d4\") " pod="openstack/heat-engine-55fc64bfdd-779rm" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.115044 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eda29ee7-1690-43ed-be0f-c8aac34d6daa-combined-ca-bundle\") pod \"heat-api-798947df95-gjr2t\" (UID: \"eda29ee7-1690-43ed-be0f-c8aac34d6daa\") " pod="openstack/heat-api-798947df95-gjr2t" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.115250 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a86ea592-5d80-4415-8c1f-5471274999d4-config-data-custom\") pod \"heat-engine-55fc64bfdd-779rm\" (UID: \"a86ea592-5d80-4415-8c1f-5471274999d4\") " pod="openstack/heat-engine-55fc64bfdd-779rm" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.118617 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eda29ee7-1690-43ed-be0f-c8aac34d6daa-config-data\") pod \"heat-api-798947df95-gjr2t\" (UID: \"eda29ee7-1690-43ed-be0f-c8aac34d6daa\") " pod="openstack/heat-api-798947df95-gjr2t" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.123484 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vhbg\" (UniqueName: \"kubernetes.io/projected/a86ea592-5d80-4415-8c1f-5471274999d4-kube-api-access-4vhbg\") pod \"heat-engine-55fc64bfdd-779rm\" (UID: \"a86ea592-5d80-4415-8c1f-5471274999d4\") " pod="openstack/heat-engine-55fc64bfdd-779rm" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.123842 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jkfv\" (UniqueName: \"kubernetes.io/projected/eda29ee7-1690-43ed-be0f-c8aac34d6daa-kube-api-access-6jkfv\") pod \"heat-api-798947df95-gjr2t\" (UID: \"eda29ee7-1690-43ed-be0f-c8aac34d6daa\") " pod="openstack/heat-api-798947df95-gjr2t" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.200003 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-55fc64bfdd-779rm" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.205145 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/246893e5-50e3-402e-8b4f-f21b31af732e-combined-ca-bundle\") pod \"heat-cfnapi-7c696dc874-7894k\" (UID: \"246893e5-50e3-402e-8b4f-f21b31af732e\") " pod="openstack/heat-cfnapi-7c696dc874-7894k" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.205214 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/246893e5-50e3-402e-8b4f-f21b31af732e-config-data-custom\") pod \"heat-cfnapi-7c696dc874-7894k\" (UID: \"246893e5-50e3-402e-8b4f-f21b31af732e\") " pod="openstack/heat-cfnapi-7c696dc874-7894k" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.205239 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/246893e5-50e3-402e-8b4f-f21b31af732e-config-data\") pod \"heat-cfnapi-7c696dc874-7894k\" (UID: \"246893e5-50e3-402e-8b4f-f21b31af732e\") " pod="openstack/heat-cfnapi-7c696dc874-7894k" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.205380 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5bzt\" (UniqueName: \"kubernetes.io/projected/246893e5-50e3-402e-8b4f-f21b31af732e-kube-api-access-s5bzt\") pod \"heat-cfnapi-7c696dc874-7894k\" (UID: \"246893e5-50e3-402e-8b4f-f21b31af732e\") " pod="openstack/heat-cfnapi-7c696dc874-7894k" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.212687 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/246893e5-50e3-402e-8b4f-f21b31af732e-config-data-custom\") pod \"heat-cfnapi-7c696dc874-7894k\" (UID: \"246893e5-50e3-402e-8b4f-f21b31af732e\") " pod="openstack/heat-cfnapi-7c696dc874-7894k" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.214567 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/246893e5-50e3-402e-8b4f-f21b31af732e-config-data\") pod \"heat-cfnapi-7c696dc874-7894k\" (UID: \"246893e5-50e3-402e-8b4f-f21b31af732e\") " pod="openstack/heat-cfnapi-7c696dc874-7894k" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.224588 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/246893e5-50e3-402e-8b4f-f21b31af732e-combined-ca-bundle\") pod \"heat-cfnapi-7c696dc874-7894k\" (UID: \"246893e5-50e3-402e-8b4f-f21b31af732e\") " pod="openstack/heat-cfnapi-7c696dc874-7894k" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.226154 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5bzt\" (UniqueName: \"kubernetes.io/projected/246893e5-50e3-402e-8b4f-f21b31af732e-kube-api-access-s5bzt\") pod \"heat-cfnapi-7c696dc874-7894k\" (UID: \"246893e5-50e3-402e-8b4f-f21b31af732e\") " pod="openstack/heat-cfnapi-7c696dc874-7894k" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.281062 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-798947df95-gjr2t" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.345555 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7c696dc874-7894k" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.354281 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.408867 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bfefa56f-3099-4f29-b43b-47e1fcb7ee3d-config-data-custom\") pod \"bfefa56f-3099-4f29-b43b-47e1fcb7ee3d\" (UID: \"bfefa56f-3099-4f29-b43b-47e1fcb7ee3d\") " Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.409068 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bfefa56f-3099-4f29-b43b-47e1fcb7ee3d-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "bfefa56f-3099-4f29-b43b-47e1fcb7ee3d" (UID: "bfefa56f-3099-4f29-b43b-47e1fcb7ee3d"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.409400 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bfefa56f-3099-4f29-b43b-47e1fcb7ee3d-etc-machine-id\") pod \"bfefa56f-3099-4f29-b43b-47e1fcb7ee3d\" (UID: \"bfefa56f-3099-4f29-b43b-47e1fcb7ee3d\") " Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.409444 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfefa56f-3099-4f29-b43b-47e1fcb7ee3d-scripts\") pod \"bfefa56f-3099-4f29-b43b-47e1fcb7ee3d\" (UID: \"bfefa56f-3099-4f29-b43b-47e1fcb7ee3d\") " Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.409478 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfefa56f-3099-4f29-b43b-47e1fcb7ee3d-combined-ca-bundle\") pod \"bfefa56f-3099-4f29-b43b-47e1fcb7ee3d\" (UID: \"bfefa56f-3099-4f29-b43b-47e1fcb7ee3d\") " Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.409501 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-79vhb\" (UniqueName: \"kubernetes.io/projected/bfefa56f-3099-4f29-b43b-47e1fcb7ee3d-kube-api-access-79vhb\") pod \"bfefa56f-3099-4f29-b43b-47e1fcb7ee3d\" (UID: \"bfefa56f-3099-4f29-b43b-47e1fcb7ee3d\") " Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.409519 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfefa56f-3099-4f29-b43b-47e1fcb7ee3d-config-data\") pod \"bfefa56f-3099-4f29-b43b-47e1fcb7ee3d\" (UID: \"bfefa56f-3099-4f29-b43b-47e1fcb7ee3d\") " Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.410098 4802 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bfefa56f-3099-4f29-b43b-47e1fcb7ee3d-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.429388 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfefa56f-3099-4f29-b43b-47e1fcb7ee3d-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "bfefa56f-3099-4f29-b43b-47e1fcb7ee3d" (UID: "bfefa56f-3099-4f29-b43b-47e1fcb7ee3d"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.429808 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfefa56f-3099-4f29-b43b-47e1fcb7ee3d-kube-api-access-79vhb" (OuterVolumeSpecName: "kube-api-access-79vhb") pod "bfefa56f-3099-4f29-b43b-47e1fcb7ee3d" (UID: "bfefa56f-3099-4f29-b43b-47e1fcb7ee3d"). InnerVolumeSpecName "kube-api-access-79vhb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.433991 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfefa56f-3099-4f29-b43b-47e1fcb7ee3d-scripts" (OuterVolumeSpecName: "scripts") pod "bfefa56f-3099-4f29-b43b-47e1fcb7ee3d" (UID: "bfefa56f-3099-4f29-b43b-47e1fcb7ee3d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.512629 4802 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bfefa56f-3099-4f29-b43b-47e1fcb7ee3d-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.512666 4802 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfefa56f-3099-4f29-b43b-47e1fcb7ee3d-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.512675 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-79vhb\" (UniqueName: \"kubernetes.io/projected/bfefa56f-3099-4f29-b43b-47e1fcb7ee3d-kube-api-access-79vhb\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.526888 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfefa56f-3099-4f29-b43b-47e1fcb7ee3d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bfefa56f-3099-4f29-b43b-47e1fcb7ee3d" (UID: "bfefa56f-3099-4f29-b43b-47e1fcb7ee3d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.571447 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfefa56f-3099-4f29-b43b-47e1fcb7ee3d-config-data" (OuterVolumeSpecName: "config-data") pod "bfefa56f-3099-4f29-b43b-47e1fcb7ee3d" (UID: "bfefa56f-3099-4f29-b43b-47e1fcb7ee3d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.614673 4802 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfefa56f-3099-4f29-b43b-47e1fcb7ee3d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.614705 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfefa56f-3099-4f29-b43b-47e1fcb7ee3d-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.637116 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-686fbdfd7f-sfrn5" event={"ID":"4985762a-1d53-40ad-85da-c6728bdbaa3d","Type":"ContainerStarted","Data":"e5656ca5aaedc4352c3be8165d038ff2da3df4f4f688ee63ef6054db4e70dc41"} Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.658144 4802 generic.go:334] "Generic (PLEG): container finished" podID="22294441-50c6-4dc0-80f1-4b14c1fc62a9" containerID="90662e183b0e95670b64555522541d9155b259f309bcf2d1939c8051e4b9fb83" exitCode=0 Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.658224 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"22294441-50c6-4dc0-80f1-4b14c1fc62a9","Type":"ContainerDied","Data":"90662e183b0e95670b64555522541d9155b259f309bcf2d1939c8051e4b9fb83"} Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.677726 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bfefa56f-3099-4f29-b43b-47e1fcb7ee3d","Type":"ContainerDied","Data":"6d5d5901c49a822b29593438266b9a9d1998e4d00ab361e84e0cc72ddddfcd0d"} Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.677788 4802 scope.go:117] "RemoveContainer" containerID="44682c5fb47ae56c3da4fdc038c58337805efffb60c98e2e3c8fdba5f7f4afdd" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.677940 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.743859 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.749506 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.785252 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.788886 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 04:00:04 crc kubenswrapper[4802]: E1206 04:00:04.789372 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d4f4fff-a6ec-4fef-82b4-6ba392fd297c" containerName="ceilometer-notification-agent" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.789388 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d4f4fff-a6ec-4fef-82b4-6ba392fd297c" containerName="ceilometer-notification-agent" Dec 06 04:00:04 crc kubenswrapper[4802]: E1206 04:00:04.789406 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfefa56f-3099-4f29-b43b-47e1fcb7ee3d" containerName="probe" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.789411 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfefa56f-3099-4f29-b43b-47e1fcb7ee3d" containerName="probe" Dec 06 04:00:04 crc kubenswrapper[4802]: E1206 04:00:04.789424 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfefa56f-3099-4f29-b43b-47e1fcb7ee3d" containerName="cinder-scheduler" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.789430 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfefa56f-3099-4f29-b43b-47e1fcb7ee3d" containerName="cinder-scheduler" Dec 06 04:00:04 crc kubenswrapper[4802]: E1206 04:00:04.789446 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d4f4fff-a6ec-4fef-82b4-6ba392fd297c" containerName="proxy-httpd" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.789451 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d4f4fff-a6ec-4fef-82b4-6ba392fd297c" containerName="proxy-httpd" Dec 06 04:00:04 crc kubenswrapper[4802]: E1206 04:00:04.789468 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d4f4fff-a6ec-4fef-82b4-6ba392fd297c" containerName="sg-core" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.789473 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d4f4fff-a6ec-4fef-82b4-6ba392fd297c" containerName="sg-core" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.789693 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d4f4fff-a6ec-4fef-82b4-6ba392fd297c" containerName="ceilometer-notification-agent" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.789709 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d4f4fff-a6ec-4fef-82b4-6ba392fd297c" containerName="sg-core" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.789726 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfefa56f-3099-4f29-b43b-47e1fcb7ee3d" containerName="cinder-scheduler" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.789790 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfefa56f-3099-4f29-b43b-47e1fcb7ee3d" containerName="probe" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.789804 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d4f4fff-a6ec-4fef-82b4-6ba392fd297c" containerName="proxy-httpd" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.791006 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.795386 4802 scope.go:117] "RemoveContainer" containerID="dcca102ecce1b33a2926573c61c19c6fe74c70e2b98318750c10253d54f15f5c" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.795875 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.816350 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4d4f4fff-a6ec-4fef-82b4-6ba392fd297c-sg-core-conf-yaml\") pod \"4d4f4fff-a6ec-4fef-82b4-6ba392fd297c\" (UID: \"4d4f4fff-a6ec-4fef-82b4-6ba392fd297c\") " Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.816404 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d4f4fff-a6ec-4fef-82b4-6ba392fd297c-config-data\") pod \"4d4f4fff-a6ec-4fef-82b4-6ba392fd297c\" (UID: \"4d4f4fff-a6ec-4fef-82b4-6ba392fd297c\") " Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.816449 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4d4f4fff-a6ec-4fef-82b4-6ba392fd297c-log-httpd\") pod \"4d4f4fff-a6ec-4fef-82b4-6ba392fd297c\" (UID: \"4d4f4fff-a6ec-4fef-82b4-6ba392fd297c\") " Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.816473 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fjnp6\" (UniqueName: \"kubernetes.io/projected/4d4f4fff-a6ec-4fef-82b4-6ba392fd297c-kube-api-access-fjnp6\") pod \"4d4f4fff-a6ec-4fef-82b4-6ba392fd297c\" (UID: \"4d4f4fff-a6ec-4fef-82b4-6ba392fd297c\") " Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.816514 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d4f4fff-a6ec-4fef-82b4-6ba392fd297c-combined-ca-bundle\") pod \"4d4f4fff-a6ec-4fef-82b4-6ba392fd297c\" (UID: \"4d4f4fff-a6ec-4fef-82b4-6ba392fd297c\") " Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.816562 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4d4f4fff-a6ec-4fef-82b4-6ba392fd297c-run-httpd\") pod \"4d4f4fff-a6ec-4fef-82b4-6ba392fd297c\" (UID: \"4d4f4fff-a6ec-4fef-82b4-6ba392fd297c\") " Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.816616 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d4f4fff-a6ec-4fef-82b4-6ba392fd297c-scripts\") pod \"4d4f4fff-a6ec-4fef-82b4-6ba392fd297c\" (UID: \"4d4f4fff-a6ec-4fef-82b4-6ba392fd297c\") " Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.821642 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.822283 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d4f4fff-a6ec-4fef-82b4-6ba392fd297c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "4d4f4fff-a6ec-4fef-82b4-6ba392fd297c" (UID: "4d4f4fff-a6ec-4fef-82b4-6ba392fd297c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.822432 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d4f4fff-a6ec-4fef-82b4-6ba392fd297c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "4d4f4fff-a6ec-4fef-82b4-6ba392fd297c" (UID: "4d4f4fff-a6ec-4fef-82b4-6ba392fd297c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.854704 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d4f4fff-a6ec-4fef-82b4-6ba392fd297c-kube-api-access-fjnp6" (OuterVolumeSpecName: "kube-api-access-fjnp6") pod "4d4f4fff-a6ec-4fef-82b4-6ba392fd297c" (UID: "4d4f4fff-a6ec-4fef-82b4-6ba392fd297c"). InnerVolumeSpecName "kube-api-access-fjnp6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.865793 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d4f4fff-a6ec-4fef-82b4-6ba392fd297c-scripts" (OuterVolumeSpecName: "scripts") pod "4d4f4fff-a6ec-4fef-82b4-6ba392fd297c" (UID: "4d4f4fff-a6ec-4fef-82b4-6ba392fd297c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.918140 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqtck\" (UniqueName: \"kubernetes.io/projected/92c245ae-d893-44d7-99e3-69a9e58735cf-kube-api-access-jqtck\") pod \"cinder-scheduler-0\" (UID: \"92c245ae-d893-44d7-99e3-69a9e58735cf\") " pod="openstack/cinder-scheduler-0" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.918190 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92c245ae-d893-44d7-99e3-69a9e58735cf-scripts\") pod \"cinder-scheduler-0\" (UID: \"92c245ae-d893-44d7-99e3-69a9e58735cf\") " pod="openstack/cinder-scheduler-0" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.918261 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/92c245ae-d893-44d7-99e3-69a9e58735cf-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"92c245ae-d893-44d7-99e3-69a9e58735cf\") " pod="openstack/cinder-scheduler-0" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.918292 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92c245ae-d893-44d7-99e3-69a9e58735cf-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"92c245ae-d893-44d7-99e3-69a9e58735cf\") " pod="openstack/cinder-scheduler-0" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.918307 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/92c245ae-d893-44d7-99e3-69a9e58735cf-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"92c245ae-d893-44d7-99e3-69a9e58735cf\") " pod="openstack/cinder-scheduler-0" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.918354 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92c245ae-d893-44d7-99e3-69a9e58735cf-config-data\") pod \"cinder-scheduler-0\" (UID: \"92c245ae-d893-44d7-99e3-69a9e58735cf\") " pod="openstack/cinder-scheduler-0" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.918441 4802 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4d4f4fff-a6ec-4fef-82b4-6ba392fd297c-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.918454 4802 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d4f4fff-a6ec-4fef-82b4-6ba392fd297c-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.918464 4802 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4d4f4fff-a6ec-4fef-82b4-6ba392fd297c-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.918475 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fjnp6\" (UniqueName: \"kubernetes.io/projected/4d4f4fff-a6ec-4fef-82b4-6ba392fd297c-kube-api-access-fjnp6\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.977704 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d4f4fff-a6ec-4fef-82b4-6ba392fd297c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "4d4f4fff-a6ec-4fef-82b4-6ba392fd297c" (UID: "4d4f4fff-a6ec-4fef-82b4-6ba392fd297c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:00:04 crc kubenswrapper[4802]: I1206 04:00:04.979370 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.006081 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.019967 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f8c7901e-3653-4d82-a15f-dd12cf4ab8c5-httpd-run\") pod \"f8c7901e-3653-4d82-a15f-dd12cf4ab8c5\" (UID: \"f8c7901e-3653-4d82-a15f-dd12cf4ab8c5\") " Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.020076 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8c7901e-3653-4d82-a15f-dd12cf4ab8c5-combined-ca-bundle\") pod \"f8c7901e-3653-4d82-a15f-dd12cf4ab8c5\" (UID: \"f8c7901e-3653-4d82-a15f-dd12cf4ab8c5\") " Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.020122 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hbkgw\" (UniqueName: \"kubernetes.io/projected/f8c7901e-3653-4d82-a15f-dd12cf4ab8c5-kube-api-access-hbkgw\") pod \"f8c7901e-3653-4d82-a15f-dd12cf4ab8c5\" (UID: \"f8c7901e-3653-4d82-a15f-dd12cf4ab8c5\") " Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.020170 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8c7901e-3653-4d82-a15f-dd12cf4ab8c5-public-tls-certs\") pod \"f8c7901e-3653-4d82-a15f-dd12cf4ab8c5\" (UID: \"f8c7901e-3653-4d82-a15f-dd12cf4ab8c5\") " Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.020192 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8c7901e-3653-4d82-a15f-dd12cf4ab8c5-logs\") pod \"f8c7901e-3653-4d82-a15f-dd12cf4ab8c5\" (UID: \"f8c7901e-3653-4d82-a15f-dd12cf4ab8c5\") " Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.020325 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8c7901e-3653-4d82-a15f-dd12cf4ab8c5-scripts\") pod \"f8c7901e-3653-4d82-a15f-dd12cf4ab8c5\" (UID: \"f8c7901e-3653-4d82-a15f-dd12cf4ab8c5\") " Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.020358 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"f8c7901e-3653-4d82-a15f-dd12cf4ab8c5\" (UID: \"f8c7901e-3653-4d82-a15f-dd12cf4ab8c5\") " Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.020386 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8c7901e-3653-4d82-a15f-dd12cf4ab8c5-config-data\") pod \"f8c7901e-3653-4d82-a15f-dd12cf4ab8c5\" (UID: \"f8c7901e-3653-4d82-a15f-dd12cf4ab8c5\") " Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.020572 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92c245ae-d893-44d7-99e3-69a9e58735cf-config-data\") pod \"cinder-scheduler-0\" (UID: \"92c245ae-d893-44d7-99e3-69a9e58735cf\") " pod="openstack/cinder-scheduler-0" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.020626 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqtck\" (UniqueName: \"kubernetes.io/projected/92c245ae-d893-44d7-99e3-69a9e58735cf-kube-api-access-jqtck\") pod \"cinder-scheduler-0\" (UID: \"92c245ae-d893-44d7-99e3-69a9e58735cf\") " pod="openstack/cinder-scheduler-0" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.020649 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92c245ae-d893-44d7-99e3-69a9e58735cf-scripts\") pod \"cinder-scheduler-0\" (UID: \"92c245ae-d893-44d7-99e3-69a9e58735cf\") " pod="openstack/cinder-scheduler-0" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.020710 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/92c245ae-d893-44d7-99e3-69a9e58735cf-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"92c245ae-d893-44d7-99e3-69a9e58735cf\") " pod="openstack/cinder-scheduler-0" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.020743 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92c245ae-d893-44d7-99e3-69a9e58735cf-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"92c245ae-d893-44d7-99e3-69a9e58735cf\") " pod="openstack/cinder-scheduler-0" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.020772 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/92c245ae-d893-44d7-99e3-69a9e58735cf-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"92c245ae-d893-44d7-99e3-69a9e58735cf\") " pod="openstack/cinder-scheduler-0" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.020829 4802 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4d4f4fff-a6ec-4fef-82b4-6ba392fd297c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.020876 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/92c245ae-d893-44d7-99e3-69a9e58735cf-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"92c245ae-d893-44d7-99e3-69a9e58735cf\") " pod="openstack/cinder-scheduler-0" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.021697 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8c7901e-3653-4d82-a15f-dd12cf4ab8c5-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f8c7901e-3653-4d82-a15f-dd12cf4ab8c5" (UID: "f8c7901e-3653-4d82-a15f-dd12cf4ab8c5"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.034007 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8c7901e-3653-4d82-a15f-dd12cf4ab8c5-logs" (OuterVolumeSpecName: "logs") pod "f8c7901e-3653-4d82-a15f-dd12cf4ab8c5" (UID: "f8c7901e-3653-4d82-a15f-dd12cf4ab8c5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.056425 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/92c245ae-d893-44d7-99e3-69a9e58735cf-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"92c245ae-d893-44d7-99e3-69a9e58735cf\") " pod="openstack/cinder-scheduler-0" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.059498 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92c245ae-d893-44d7-99e3-69a9e58735cf-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"92c245ae-d893-44d7-99e3-69a9e58735cf\") " pod="openstack/cinder-scheduler-0" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.064713 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8c7901e-3653-4d82-a15f-dd12cf4ab8c5-scripts" (OuterVolumeSpecName: "scripts") pod "f8c7901e-3653-4d82-a15f-dd12cf4ab8c5" (UID: "f8c7901e-3653-4d82-a15f-dd12cf4ab8c5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.076294 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92c245ae-d893-44d7-99e3-69a9e58735cf-scripts\") pod \"cinder-scheduler-0\" (UID: \"92c245ae-d893-44d7-99e3-69a9e58735cf\") " pod="openstack/cinder-scheduler-0" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.077080 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92c245ae-d893-44d7-99e3-69a9e58735cf-config-data\") pod \"cinder-scheduler-0\" (UID: \"92c245ae-d893-44d7-99e3-69a9e58735cf\") " pod="openstack/cinder-scheduler-0" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.084966 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8c7901e-3653-4d82-a15f-dd12cf4ab8c5-kube-api-access-hbkgw" (OuterVolumeSpecName: "kube-api-access-hbkgw") pod "f8c7901e-3653-4d82-a15f-dd12cf4ab8c5" (UID: "f8c7901e-3653-4d82-a15f-dd12cf4ab8c5"). InnerVolumeSpecName "kube-api-access-hbkgw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.085778 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "f8c7901e-3653-4d82-a15f-dd12cf4ab8c5" (UID: "f8c7901e-3653-4d82-a15f-dd12cf4ab8c5"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.088824 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d4f4fff-a6ec-4fef-82b4-6ba392fd297c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4d4f4fff-a6ec-4fef-82b4-6ba392fd297c" (UID: "4d4f4fff-a6ec-4fef-82b4-6ba392fd297c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.103515 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-5d96658447-rg8zs"] Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.111582 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqtck\" (UniqueName: \"kubernetes.io/projected/92c245ae-d893-44d7-99e3-69a9e58735cf-kube-api-access-jqtck\") pod \"cinder-scheduler-0\" (UID: \"92c245ae-d893-44d7-99e3-69a9e58735cf\") " pod="openstack/cinder-scheduler-0" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.115529 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8c7901e-3653-4d82-a15f-dd12cf4ab8c5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f8c7901e-3653-4d82-a15f-dd12cf4ab8c5" (UID: "f8c7901e-3653-4d82-a15f-dd12cf4ab8c5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.121920 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/22294441-50c6-4dc0-80f1-4b14c1fc62a9-httpd-run\") pod \"22294441-50c6-4dc0-80f1-4b14c1fc62a9\" (UID: \"22294441-50c6-4dc0-80f1-4b14c1fc62a9\") " Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.122002 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9csvl\" (UniqueName: \"kubernetes.io/projected/22294441-50c6-4dc0-80f1-4b14c1fc62a9-kube-api-access-9csvl\") pod \"22294441-50c6-4dc0-80f1-4b14c1fc62a9\" (UID: \"22294441-50c6-4dc0-80f1-4b14c1fc62a9\") " Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.122072 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/22294441-50c6-4dc0-80f1-4b14c1fc62a9-internal-tls-certs\") pod \"22294441-50c6-4dc0-80f1-4b14c1fc62a9\" (UID: \"22294441-50c6-4dc0-80f1-4b14c1fc62a9\") " Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.122097 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"22294441-50c6-4dc0-80f1-4b14c1fc62a9\" (UID: \"22294441-50c6-4dc0-80f1-4b14c1fc62a9\") " Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.122140 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22294441-50c6-4dc0-80f1-4b14c1fc62a9-config-data\") pod \"22294441-50c6-4dc0-80f1-4b14c1fc62a9\" (UID: \"22294441-50c6-4dc0-80f1-4b14c1fc62a9\") " Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.122239 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/22294441-50c6-4dc0-80f1-4b14c1fc62a9-scripts\") pod \"22294441-50c6-4dc0-80f1-4b14c1fc62a9\" (UID: \"22294441-50c6-4dc0-80f1-4b14c1fc62a9\") " Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.122270 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22294441-50c6-4dc0-80f1-4b14c1fc62a9-combined-ca-bundle\") pod \"22294441-50c6-4dc0-80f1-4b14c1fc62a9\" (UID: \"22294441-50c6-4dc0-80f1-4b14c1fc62a9\") " Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.122304 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22294441-50c6-4dc0-80f1-4b14c1fc62a9-logs\") pod \"22294441-50c6-4dc0-80f1-4b14c1fc62a9\" (UID: \"22294441-50c6-4dc0-80f1-4b14c1fc62a9\") " Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.122706 4802 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8c7901e-3653-4d82-a15f-dd12cf4ab8c5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.122722 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hbkgw\" (UniqueName: \"kubernetes.io/projected/f8c7901e-3653-4d82-a15f-dd12cf4ab8c5-kube-api-access-hbkgw\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.122733 4802 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8c7901e-3653-4d82-a15f-dd12cf4ab8c5-logs\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.122743 4802 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8c7901e-3653-4d82-a15f-dd12cf4ab8c5-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.122784 4802 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.122795 4802 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f8c7901e-3653-4d82-a15f-dd12cf4ab8c5-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.122806 4802 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d4f4fff-a6ec-4fef-82b4-6ba392fd297c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.126292 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22294441-50c6-4dc0-80f1-4b14c1fc62a9-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "22294441-50c6-4dc0-80f1-4b14c1fc62a9" (UID: "22294441-50c6-4dc0-80f1-4b14c1fc62a9"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.128127 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22294441-50c6-4dc0-80f1-4b14c1fc62a9-logs" (OuterVolumeSpecName: "logs") pod "22294441-50c6-4dc0-80f1-4b14c1fc62a9" (UID: "22294441-50c6-4dc0-80f1-4b14c1fc62a9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.129856 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22294441-50c6-4dc0-80f1-4b14c1fc62a9-kube-api-access-9csvl" (OuterVolumeSpecName: "kube-api-access-9csvl") pod "22294441-50c6-4dc0-80f1-4b14c1fc62a9" (UID: "22294441-50c6-4dc0-80f1-4b14c1fc62a9"). InnerVolumeSpecName "kube-api-access-9csvl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.131285 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d4f4fff-a6ec-4fef-82b4-6ba392fd297c-config-data" (OuterVolumeSpecName: "config-data") pod "4d4f4fff-a6ec-4fef-82b4-6ba392fd297c" (UID: "4d4f4fff-a6ec-4fef-82b4-6ba392fd297c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.144171 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.146115 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22294441-50c6-4dc0-80f1-4b14c1fc62a9-scripts" (OuterVolumeSpecName: "scripts") pod "22294441-50c6-4dc0-80f1-4b14c1fc62a9" (UID: "22294441-50c6-4dc0-80f1-4b14c1fc62a9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.153927 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "22294441-50c6-4dc0-80f1-4b14c1fc62a9" (UID: "22294441-50c6-4dc0-80f1-4b14c1fc62a9"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.218273 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22294441-50c6-4dc0-80f1-4b14c1fc62a9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "22294441-50c6-4dc0-80f1-4b14c1fc62a9" (UID: "22294441-50c6-4dc0-80f1-4b14c1fc62a9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.227891 4802 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/22294441-50c6-4dc0-80f1-4b14c1fc62a9-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.227951 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9csvl\" (UniqueName: \"kubernetes.io/projected/22294441-50c6-4dc0-80f1-4b14c1fc62a9-kube-api-access-9csvl\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.227976 4802 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.227986 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d4f4fff-a6ec-4fef-82b4-6ba392fd297c-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.228000 4802 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/22294441-50c6-4dc0-80f1-4b14c1fc62a9-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.228013 4802 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22294441-50c6-4dc0-80f1-4b14c1fc62a9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.228023 4802 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22294441-50c6-4dc0-80f1-4b14c1fc62a9-logs\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.255323 4802 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.258948 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8c7901e-3653-4d82-a15f-dd12cf4ab8c5-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "f8c7901e-3653-4d82-a15f-dd12cf4ab8c5" (UID: "f8c7901e-3653-4d82-a15f-dd12cf4ab8c5"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.270692 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8c7901e-3653-4d82-a15f-dd12cf4ab8c5-config-data" (OuterVolumeSpecName: "config-data") pod "f8c7901e-3653-4d82-a15f-dd12cf4ab8c5" (UID: "f8c7901e-3653-4d82-a15f-dd12cf4ab8c5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.277793 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22294441-50c6-4dc0-80f1-4b14c1fc62a9-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "22294441-50c6-4dc0-80f1-4b14c1fc62a9" (UID: "22294441-50c6-4dc0-80f1-4b14c1fc62a9"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.284872 4802 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.295850 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22294441-50c6-4dc0-80f1-4b14c1fc62a9-config-data" (OuterVolumeSpecName: "config-data") pod "22294441-50c6-4dc0-80f1-4b14c1fc62a9" (UID: "22294441-50c6-4dc0-80f1-4b14c1fc62a9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.329202 4802 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/22294441-50c6-4dc0-80f1-4b14c1fc62a9-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.329234 4802 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.329243 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22294441-50c6-4dc0-80f1-4b14c1fc62a9-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.329252 4802 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.329260 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8c7901e-3653-4d82-a15f-dd12cf4ab8c5-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.329268 4802 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8c7901e-3653-4d82-a15f-dd12cf4ab8c5-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.345089 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-68bd774b7c-9sjk9" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.419857 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-844df74668-r698z"] Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.420215 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-844df74668-r698z" podUID="b364aef0-231f-472a-a8c9-51a5199c0997" containerName="neutron-api" containerID="cri-o://1bd700182ce971dd2c603b23e1a7c8370de9caf2e1ce639678041c207ab36f16" gracePeriod=30 Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.420564 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-844df74668-r698z" podUID="b364aef0-231f-472a-a8c9-51a5199c0997" containerName="neutron-httpd" containerID="cri-o://31dc097fa5cf9c76a286ae10dd83297f83da4c4ead70f0e1d9e17473d8ccab81" gracePeriod=30 Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.473331 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bfefa56f-3099-4f29-b43b-47e1fcb7ee3d" path="/var/lib/kubelet/pods/bfefa56f-3099-4f29-b43b-47e1fcb7ee3d/volumes" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.694126 4802 generic.go:334] "Generic (PLEG): container finished" podID="b364aef0-231f-472a-a8c9-51a5199c0997" containerID="31dc097fa5cf9c76a286ae10dd83297f83da4c4ead70f0e1d9e17473d8ccab81" exitCode=0 Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.694199 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-844df74668-r698z" event={"ID":"b364aef0-231f-472a-a8c9-51a5199c0997","Type":"ContainerDied","Data":"31dc097fa5cf9c76a286ae10dd83297f83da4c4ead70f0e1d9e17473d8ccab81"} Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.701509 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"22294441-50c6-4dc0-80f1-4b14c1fc62a9","Type":"ContainerDied","Data":"58648fb9db1348bd1d6f78c8f2f582f88632a5f63e2e397444c22fc545cb0cf0"} Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.701578 4802 scope.go:117] "RemoveContainer" containerID="90662e183b0e95670b64555522541d9155b259f309bcf2d1939c8051e4b9fb83" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.701597 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.708595 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-5d96658447-rg8zs" event={"ID":"d3da4180-e8fe-4785-8d4f-6cd172fb6399","Type":"ContainerStarted","Data":"15461749ec2cccc9a0f2a06a22252592288f81b9b432dde2bc6f7f9917e9bb8a"} Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.708645 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-5d96658447-rg8zs" event={"ID":"d3da4180-e8fe-4785-8d4f-6cd172fb6399","Type":"ContainerStarted","Data":"663395a8db84a3f93d8c284a83195cda768663bccaaf7e4a1ab8ce35a11fcb70"} Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.709072 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-5d96658447-rg8zs" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.714642 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"20c275f9-f37c-427a-b8f0-77c35fc880b5","Type":"ContainerStarted","Data":"91feaf9e236acdb592afee593227d9182b072b191ed0b4a19cfe7a72ae654e1e"} Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.720057 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-686fbdfd7f-sfrn5" event={"ID":"4985762a-1d53-40ad-85da-c6728bdbaa3d","Type":"ContainerStarted","Data":"1214da525932950dad5cf4fe69660eee47d5a8d92130ca9deffd19e1470973da"} Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.720284 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-686fbdfd7f-sfrn5" event={"ID":"4985762a-1d53-40ad-85da-c6728bdbaa3d","Type":"ContainerStarted","Data":"f4e95804be7b8ab2dc5ff40f35b2ed12e52ae086598591563a8f7d40fe31039c"} Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.722392 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-686fbdfd7f-sfrn5" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.722560 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-686fbdfd7f-sfrn5" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.730598 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f8c7901e-3653-4d82-a15f-dd12cf4ab8c5","Type":"ContainerDied","Data":"92c412ca7d508b88eea090e0fed09145e674f09e003ef82d0e103b23d4b190b6"} Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.730867 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.736132 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4d4f4fff-a6ec-4fef-82b4-6ba392fd297c","Type":"ContainerDied","Data":"5ebc226f052371f321b24045dae747ef258f0a92b1f6b491170158e35a15d1ca"} Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.736391 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.746597 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.751895 4802 scope.go:117] "RemoveContainer" containerID="c6a619cd1b3bf4fef4f5bd45c896b6e8b20d75cfc21ef84bac1fb69e9b572e9a" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.773653 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.799603 4802 scope.go:117] "RemoveContainer" containerID="43393bb092a35409606f84f7092e94a2e0eec77e324a88aed2126f8d32b543b1" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.802268 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 04:00:05 crc kubenswrapper[4802]: E1206 04:00:05.802740 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22294441-50c6-4dc0-80f1-4b14c1fc62a9" containerName="glance-httpd" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.802764 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="22294441-50c6-4dc0-80f1-4b14c1fc62a9" containerName="glance-httpd" Dec 06 04:00:05 crc kubenswrapper[4802]: E1206 04:00:05.802779 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8c7901e-3653-4d82-a15f-dd12cf4ab8c5" containerName="glance-log" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.802785 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8c7901e-3653-4d82-a15f-dd12cf4ab8c5" containerName="glance-log" Dec 06 04:00:05 crc kubenswrapper[4802]: E1206 04:00:05.802795 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22294441-50c6-4dc0-80f1-4b14c1fc62a9" containerName="glance-log" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.802802 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="22294441-50c6-4dc0-80f1-4b14c1fc62a9" containerName="glance-log" Dec 06 04:00:05 crc kubenswrapper[4802]: E1206 04:00:05.802823 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8c7901e-3653-4d82-a15f-dd12cf4ab8c5" containerName="glance-httpd" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.802829 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8c7901e-3653-4d82-a15f-dd12cf4ab8c5" containerName="glance-httpd" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.802996 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8c7901e-3653-4d82-a15f-dd12cf4ab8c5" containerName="glance-httpd" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.803013 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="22294441-50c6-4dc0-80f1-4b14c1fc62a9" containerName="glance-log" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.803025 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="22294441-50c6-4dc0-80f1-4b14c1fc62a9" containerName="glance-httpd" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.803037 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8c7901e-3653-4d82-a15f-dd12cf4ab8c5" containerName="glance-log" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.804044 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.811319 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-n458c" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.811482 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.811617 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.811732 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.820006 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-686fbdfd7f-sfrn5" podStartSLOduration=9.819984371 podStartE2EDuration="9.819984371s" podCreationTimestamp="2025-12-06 03:59:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:00:05.760397279 +0000 UTC m=+1198.632306431" watchObservedRunningTime="2025-12-06 04:00:05.819984371 +0000 UTC m=+1198.691893523" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.836573 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.841524 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/19c7331e-7c80-4aa6-8a9f-2a7f878c79f2-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"19c7331e-7c80-4aa6-8a9f-2a7f878c79f2\") " pod="openstack/glance-default-internal-api-0" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.841634 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19c7331e-7c80-4aa6-8a9f-2a7f878c79f2-config-data\") pod \"glance-default-internal-api-0\" (UID: \"19c7331e-7c80-4aa6-8a9f-2a7f878c79f2\") " pod="openstack/glance-default-internal-api-0" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.841666 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/19c7331e-7c80-4aa6-8a9f-2a7f878c79f2-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"19c7331e-7c80-4aa6-8a9f-2a7f878c79f2\") " pod="openstack/glance-default-internal-api-0" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.841701 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"19c7331e-7c80-4aa6-8a9f-2a7f878c79f2\") " pod="openstack/glance-default-internal-api-0" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.841873 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19c7331e-7c80-4aa6-8a9f-2a7f878c79f2-logs\") pod \"glance-default-internal-api-0\" (UID: \"19c7331e-7c80-4aa6-8a9f-2a7f878c79f2\") " pod="openstack/glance-default-internal-api-0" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.841905 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjqpf\" (UniqueName: \"kubernetes.io/projected/19c7331e-7c80-4aa6-8a9f-2a7f878c79f2-kube-api-access-qjqpf\") pod \"glance-default-internal-api-0\" (UID: \"19c7331e-7c80-4aa6-8a9f-2a7f878c79f2\") " pod="openstack/glance-default-internal-api-0" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.841983 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19c7331e-7c80-4aa6-8a9f-2a7f878c79f2-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"19c7331e-7c80-4aa6-8a9f-2a7f878c79f2\") " pod="openstack/glance-default-internal-api-0" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.842001 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19c7331e-7c80-4aa6-8a9f-2a7f878c79f2-scripts\") pod \"glance-default-internal-api-0\" (UID: \"19c7331e-7c80-4aa6-8a9f-2a7f878c79f2\") " pod="openstack/glance-default-internal-api-0" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.844108 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=3.035857297 podStartE2EDuration="14.844087426s" podCreationTimestamp="2025-12-06 03:59:51 +0000 UTC" firstStartedPulling="2025-12-06 03:59:52.4904227 +0000 UTC m=+1185.362331852" lastFinishedPulling="2025-12-06 04:00:04.298652829 +0000 UTC m=+1197.170561981" observedRunningTime="2025-12-06 04:00:05.789045979 +0000 UTC m=+1198.660955131" watchObservedRunningTime="2025-12-06 04:00:05.844087426 +0000 UTC m=+1198.715996578" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.880474 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-7c696dc874-7894k"] Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.891265 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416560-zzktw"] Dec 06 04:00:05 crc kubenswrapper[4802]: W1206 04:00:05.899643 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeda29ee7_1690_43ed_be0f_c8aac34d6daa.slice/crio-07d0778674f0592d75374d5761ba6925eda84b3b0bde117a99e41e6eea18d700 WatchSource:0}: Error finding container 07d0778674f0592d75374d5761ba6925eda84b3b0bde117a99e41e6eea18d700: Status 404 returned error can't find the container with id 07d0778674f0592d75374d5761ba6925eda84b3b0bde117a99e41e6eea18d700 Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.925126 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-688b9f5b49-vxlxf"] Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.939927 4802 scope.go:117] "RemoveContainer" containerID="bd7d4fa9d3441f11c0fea392f92f249b999c25be443db3a969902839362960b2" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.943144 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19c7331e-7c80-4aa6-8a9f-2a7f878c79f2-config-data\") pod \"glance-default-internal-api-0\" (UID: \"19c7331e-7c80-4aa6-8a9f-2a7f878c79f2\") " pod="openstack/glance-default-internal-api-0" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.945782 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/19c7331e-7c80-4aa6-8a9f-2a7f878c79f2-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"19c7331e-7c80-4aa6-8a9f-2a7f878c79f2\") " pod="openstack/glance-default-internal-api-0" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.945918 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"19c7331e-7c80-4aa6-8a9f-2a7f878c79f2\") " pod="openstack/glance-default-internal-api-0" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.946034 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19c7331e-7c80-4aa6-8a9f-2a7f878c79f2-logs\") pod \"glance-default-internal-api-0\" (UID: \"19c7331e-7c80-4aa6-8a9f-2a7f878c79f2\") " pod="openstack/glance-default-internal-api-0" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.946129 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjqpf\" (UniqueName: \"kubernetes.io/projected/19c7331e-7c80-4aa6-8a9f-2a7f878c79f2-kube-api-access-qjqpf\") pod \"glance-default-internal-api-0\" (UID: \"19c7331e-7c80-4aa6-8a9f-2a7f878c79f2\") " pod="openstack/glance-default-internal-api-0" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.946286 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19c7331e-7c80-4aa6-8a9f-2a7f878c79f2-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"19c7331e-7c80-4aa6-8a9f-2a7f878c79f2\") " pod="openstack/glance-default-internal-api-0" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.946373 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19c7331e-7c80-4aa6-8a9f-2a7f878c79f2-scripts\") pod \"glance-default-internal-api-0\" (UID: \"19c7331e-7c80-4aa6-8a9f-2a7f878c79f2\") " pod="openstack/glance-default-internal-api-0" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.947989 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/19c7331e-7c80-4aa6-8a9f-2a7f878c79f2-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"19c7331e-7c80-4aa6-8a9f-2a7f878c79f2\") " pod="openstack/glance-default-internal-api-0" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.947319 4802 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"19c7331e-7c80-4aa6-8a9f-2a7f878c79f2\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-internal-api-0" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.949347 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/19c7331e-7c80-4aa6-8a9f-2a7f878c79f2-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"19c7331e-7c80-4aa6-8a9f-2a7f878c79f2\") " pod="openstack/glance-default-internal-api-0" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.947070 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19c7331e-7c80-4aa6-8a9f-2a7f878c79f2-logs\") pod \"glance-default-internal-api-0\" (UID: \"19c7331e-7c80-4aa6-8a9f-2a7f878c79f2\") " pod="openstack/glance-default-internal-api-0" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.956336 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/19c7331e-7c80-4aa6-8a9f-2a7f878c79f2-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"19c7331e-7c80-4aa6-8a9f-2a7f878c79f2\") " pod="openstack/glance-default-internal-api-0" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.956937 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19c7331e-7c80-4aa6-8a9f-2a7f878c79f2-config-data\") pod \"glance-default-internal-api-0\" (UID: \"19c7331e-7c80-4aa6-8a9f-2a7f878c79f2\") " pod="openstack/glance-default-internal-api-0" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.960091 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-55fc64bfdd-779rm"] Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.963782 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19c7331e-7c80-4aa6-8a9f-2a7f878c79f2-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"19c7331e-7c80-4aa6-8a9f-2a7f878c79f2\") " pod="openstack/glance-default-internal-api-0" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.976905 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19c7331e-7c80-4aa6-8a9f-2a7f878c79f2-scripts\") pod \"glance-default-internal-api-0\" (UID: \"19c7331e-7c80-4aa6-8a9f-2a7f878c79f2\") " pod="openstack/glance-default-internal-api-0" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.978529 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjqpf\" (UniqueName: \"kubernetes.io/projected/19c7331e-7c80-4aa6-8a9f-2a7f878c79f2-kube-api-access-qjqpf\") pod \"glance-default-internal-api-0\" (UID: \"19c7331e-7c80-4aa6-8a9f-2a7f878c79f2\") " pod="openstack/glance-default-internal-api-0" Dec 06 04:00:05 crc kubenswrapper[4802]: I1206 04:00:05.996643 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-798947df95-gjr2t"] Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.010919 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-6fc9b98cc8-7swm4"] Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.012671 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-5d96658447-rg8zs" podStartSLOduration=10.012651196 podStartE2EDuration="10.012651196s" podCreationTimestamp="2025-12-06 03:59:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:00:05.840645243 +0000 UTC m=+1198.712554395" watchObservedRunningTime="2025-12-06 04:00:06.012651196 +0000 UTC m=+1198.884560348" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.015398 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"19c7331e-7c80-4aa6-8a9f-2a7f878c79f2\") " pod="openstack/glance-default-internal-api-0" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.019783 4802 scope.go:117] "RemoveContainer" containerID="ec333f217f947fe8e066593d0ae87d8fd22fa9027bc4f3bc00f5f9ae7a30b116" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.134149 4802 scope.go:117] "RemoveContainer" containerID="fff16873781a4a16ca4ed5fa3a576de4e59fcca96bb97e804e8f0e742df14984" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.147772 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.191160 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.204626 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.255288 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.265303 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 04:00:06 crc kubenswrapper[4802]: W1206 04:00:06.271090 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod92c245ae_d893_44d7_99e3_69a9e58735cf.slice/crio-2d4a90ce9e1302a8c3c06035a1e04123227fc8058b5fb558713cbc770776246e WatchSource:0}: Error finding container 2d4a90ce9e1302a8c3c06035a1e04123227fc8058b5fb558713cbc770776246e: Status 404 returned error can't find the container with id 2d4a90ce9e1302a8c3c06035a1e04123227fc8058b5fb558713cbc770776246e Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.271306 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.271530 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.304677 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.328380 4802 scope.go:117] "RemoveContainer" containerID="fa037e55af9d2c4256d307c1f7e958d5e921062fe281b0182ef8ec24b83b1e30" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.341989 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.345333 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.364617 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vwk5\" (UniqueName: \"kubernetes.io/projected/11b0ceda-c31c-4bc3-ad63-849047b7f80c-kube-api-access-4vwk5\") pod \"ceilometer-0\" (UID: \"11b0ceda-c31c-4bc3-ad63-849047b7f80c\") " pod="openstack/ceilometer-0" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.365016 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/11b0ceda-c31c-4bc3-ad63-849047b7f80c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"11b0ceda-c31c-4bc3-ad63-849047b7f80c\") " pod="openstack/ceilometer-0" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.365051 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/11b0ceda-c31c-4bc3-ad63-849047b7f80c-run-httpd\") pod \"ceilometer-0\" (UID: \"11b0ceda-c31c-4bc3-ad63-849047b7f80c\") " pod="openstack/ceilometer-0" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.365105 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11b0ceda-c31c-4bc3-ad63-849047b7f80c-scripts\") pod \"ceilometer-0\" (UID: \"11b0ceda-c31c-4bc3-ad63-849047b7f80c\") " pod="openstack/ceilometer-0" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.365131 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11b0ceda-c31c-4bc3-ad63-849047b7f80c-config-data\") pod \"ceilometer-0\" (UID: \"11b0ceda-c31c-4bc3-ad63-849047b7f80c\") " pod="openstack/ceilometer-0" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.365172 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11b0ceda-c31c-4bc3-ad63-849047b7f80c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"11b0ceda-c31c-4bc3-ad63-849047b7f80c\") " pod="openstack/ceilometer-0" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.365216 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/11b0ceda-c31c-4bc3-ad63-849047b7f80c-log-httpd\") pod \"ceilometer-0\" (UID: \"11b0ceda-c31c-4bc3-ad63-849047b7f80c\") " pod="openstack/ceilometer-0" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.372109 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.374182 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.380538 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.380902 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.394163 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.407182 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-67b6f8b59b-ghkb7"] Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.439867 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.470162 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11b0ceda-c31c-4bc3-ad63-849047b7f80c-config-data\") pod \"ceilometer-0\" (UID: \"11b0ceda-c31c-4bc3-ad63-849047b7f80c\") " pod="openstack/ceilometer-0" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.470203 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11b0ceda-c31c-4bc3-ad63-849047b7f80c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"11b0ceda-c31c-4bc3-ad63-849047b7f80c\") " pod="openstack/ceilometer-0" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.470270 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/11b0ceda-c31c-4bc3-ad63-849047b7f80c-log-httpd\") pod \"ceilometer-0\" (UID: \"11b0ceda-c31c-4bc3-ad63-849047b7f80c\") " pod="openstack/ceilometer-0" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.470350 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vwk5\" (UniqueName: \"kubernetes.io/projected/11b0ceda-c31c-4bc3-ad63-849047b7f80c-kube-api-access-4vwk5\") pod \"ceilometer-0\" (UID: \"11b0ceda-c31c-4bc3-ad63-849047b7f80c\") " pod="openstack/ceilometer-0" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.470384 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/11b0ceda-c31c-4bc3-ad63-849047b7f80c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"11b0ceda-c31c-4bc3-ad63-849047b7f80c\") " pod="openstack/ceilometer-0" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.470404 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/11b0ceda-c31c-4bc3-ad63-849047b7f80c-run-httpd\") pod \"ceilometer-0\" (UID: \"11b0ceda-c31c-4bc3-ad63-849047b7f80c\") " pod="openstack/ceilometer-0" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.470459 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11b0ceda-c31c-4bc3-ad63-849047b7f80c-scripts\") pod \"ceilometer-0\" (UID: \"11b0ceda-c31c-4bc3-ad63-849047b7f80c\") " pod="openstack/ceilometer-0" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.472890 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/11b0ceda-c31c-4bc3-ad63-849047b7f80c-log-httpd\") pod \"ceilometer-0\" (UID: \"11b0ceda-c31c-4bc3-ad63-849047b7f80c\") " pod="openstack/ceilometer-0" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.473351 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/11b0ceda-c31c-4bc3-ad63-849047b7f80c-run-httpd\") pod \"ceilometer-0\" (UID: \"11b0ceda-c31c-4bc3-ad63-849047b7f80c\") " pod="openstack/ceilometer-0" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.476055 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11b0ceda-c31c-4bc3-ad63-849047b7f80c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"11b0ceda-c31c-4bc3-ad63-849047b7f80c\") " pod="openstack/ceilometer-0" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.476611 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/11b0ceda-c31c-4bc3-ad63-849047b7f80c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"11b0ceda-c31c-4bc3-ad63-849047b7f80c\") " pod="openstack/ceilometer-0" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.483012 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11b0ceda-c31c-4bc3-ad63-849047b7f80c-config-data\") pod \"ceilometer-0\" (UID: \"11b0ceda-c31c-4bc3-ad63-849047b7f80c\") " pod="openstack/ceilometer-0" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.485895 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11b0ceda-c31c-4bc3-ad63-849047b7f80c-scripts\") pod \"ceilometer-0\" (UID: \"11b0ceda-c31c-4bc3-ad63-849047b7f80c\") " pod="openstack/ceilometer-0" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.510937 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vwk5\" (UniqueName: \"kubernetes.io/projected/11b0ceda-c31c-4bc3-ad63-849047b7f80c-kube-api-access-4vwk5\") pod \"ceilometer-0\" (UID: \"11b0ceda-c31c-4bc3-ad63-849047b7f80c\") " pod="openstack/ceilometer-0" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.573936 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02888e6b-6185-45b8-84b7-8b3654270b77-logs\") pod \"glance-default-external-api-0\" (UID: \"02888e6b-6185-45b8-84b7-8b3654270b77\") " pod="openstack/glance-default-external-api-0" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.573985 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzqbf\" (UniqueName: \"kubernetes.io/projected/02888e6b-6185-45b8-84b7-8b3654270b77-kube-api-access-pzqbf\") pod \"glance-default-external-api-0\" (UID: \"02888e6b-6185-45b8-84b7-8b3654270b77\") " pod="openstack/glance-default-external-api-0" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.574081 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/02888e6b-6185-45b8-84b7-8b3654270b77-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"02888e6b-6185-45b8-84b7-8b3654270b77\") " pod="openstack/glance-default-external-api-0" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.574103 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02888e6b-6185-45b8-84b7-8b3654270b77-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"02888e6b-6185-45b8-84b7-8b3654270b77\") " pod="openstack/glance-default-external-api-0" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.574131 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/02888e6b-6185-45b8-84b7-8b3654270b77-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"02888e6b-6185-45b8-84b7-8b3654270b77\") " pod="openstack/glance-default-external-api-0" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.574174 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02888e6b-6185-45b8-84b7-8b3654270b77-config-data\") pod \"glance-default-external-api-0\" (UID: \"02888e6b-6185-45b8-84b7-8b3654270b77\") " pod="openstack/glance-default-external-api-0" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.574199 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/02888e6b-6185-45b8-84b7-8b3654270b77-scripts\") pod \"glance-default-external-api-0\" (UID: \"02888e6b-6185-45b8-84b7-8b3654270b77\") " pod="openstack/glance-default-external-api-0" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.574261 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"02888e6b-6185-45b8-84b7-8b3654270b77\") " pod="openstack/glance-default-external-api-0" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.678159 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02888e6b-6185-45b8-84b7-8b3654270b77-logs\") pod \"glance-default-external-api-0\" (UID: \"02888e6b-6185-45b8-84b7-8b3654270b77\") " pod="openstack/glance-default-external-api-0" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.678219 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzqbf\" (UniqueName: \"kubernetes.io/projected/02888e6b-6185-45b8-84b7-8b3654270b77-kube-api-access-pzqbf\") pod \"glance-default-external-api-0\" (UID: \"02888e6b-6185-45b8-84b7-8b3654270b77\") " pod="openstack/glance-default-external-api-0" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.678309 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/02888e6b-6185-45b8-84b7-8b3654270b77-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"02888e6b-6185-45b8-84b7-8b3654270b77\") " pod="openstack/glance-default-external-api-0" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.678335 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02888e6b-6185-45b8-84b7-8b3654270b77-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"02888e6b-6185-45b8-84b7-8b3654270b77\") " pod="openstack/glance-default-external-api-0" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.678361 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/02888e6b-6185-45b8-84b7-8b3654270b77-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"02888e6b-6185-45b8-84b7-8b3654270b77\") " pod="openstack/glance-default-external-api-0" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.678400 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02888e6b-6185-45b8-84b7-8b3654270b77-config-data\") pod \"glance-default-external-api-0\" (UID: \"02888e6b-6185-45b8-84b7-8b3654270b77\") " pod="openstack/glance-default-external-api-0" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.678428 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/02888e6b-6185-45b8-84b7-8b3654270b77-scripts\") pod \"glance-default-external-api-0\" (UID: \"02888e6b-6185-45b8-84b7-8b3654270b77\") " pod="openstack/glance-default-external-api-0" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.678475 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"02888e6b-6185-45b8-84b7-8b3654270b77\") " pod="openstack/glance-default-external-api-0" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.678888 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/02888e6b-6185-45b8-84b7-8b3654270b77-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"02888e6b-6185-45b8-84b7-8b3654270b77\") " pod="openstack/glance-default-external-api-0" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.678949 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02888e6b-6185-45b8-84b7-8b3654270b77-logs\") pod \"glance-default-external-api-0\" (UID: \"02888e6b-6185-45b8-84b7-8b3654270b77\") " pod="openstack/glance-default-external-api-0" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.679005 4802 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"02888e6b-6185-45b8-84b7-8b3654270b77\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-external-api-0" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.687190 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/02888e6b-6185-45b8-84b7-8b3654270b77-scripts\") pod \"glance-default-external-api-0\" (UID: \"02888e6b-6185-45b8-84b7-8b3654270b77\") " pod="openstack/glance-default-external-api-0" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.689692 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/02888e6b-6185-45b8-84b7-8b3654270b77-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"02888e6b-6185-45b8-84b7-8b3654270b77\") " pod="openstack/glance-default-external-api-0" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.690576 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02888e6b-6185-45b8-84b7-8b3654270b77-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"02888e6b-6185-45b8-84b7-8b3654270b77\") " pod="openstack/glance-default-external-api-0" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.691198 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02888e6b-6185-45b8-84b7-8b3654270b77-config-data\") pod \"glance-default-external-api-0\" (UID: \"02888e6b-6185-45b8-84b7-8b3654270b77\") " pod="openstack/glance-default-external-api-0" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.705880 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzqbf\" (UniqueName: \"kubernetes.io/projected/02888e6b-6185-45b8-84b7-8b3654270b77-kube-api-access-pzqbf\") pod \"glance-default-external-api-0\" (UID: \"02888e6b-6185-45b8-84b7-8b3654270b77\") " pod="openstack/glance-default-external-api-0" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.707538 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.754409 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"02888e6b-6185-45b8-84b7-8b3654270b77\") " pod="openstack/glance-default-external-api-0" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.774159 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"92c245ae-d893-44d7-99e3-69a9e58735cf","Type":"ContainerStarted","Data":"2d4a90ce9e1302a8c3c06035a1e04123227fc8058b5fb558713cbc770776246e"} Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.797111 4802 generic.go:334] "Generic (PLEG): container finished" podID="5898b86d-3906-4c87-8509-8c37fe50f544" containerID="b44f058acf9a6945816306974c5c4f9921892f61e3f4067def14e9ff7e313600" exitCode=0 Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.797540 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416560-zzktw" event={"ID":"5898b86d-3906-4c87-8509-8c37fe50f544","Type":"ContainerDied","Data":"b44f058acf9a6945816306974c5c4f9921892f61e3f4067def14e9ff7e313600"} Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.797564 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416560-zzktw" event={"ID":"5898b86d-3906-4c87-8509-8c37fe50f544","Type":"ContainerStarted","Data":"1e045a62ce0c233322d0a16c3e011eae154f7e843cf2c10a69ff225ab2777f70"} Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.804330 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-55fc64bfdd-779rm" event={"ID":"a86ea592-5d80-4415-8c1f-5471274999d4","Type":"ContainerStarted","Data":"6cd096cd663e8968613997e7663397fc57bacc64e65d0774ca3b27bd54251cf0"} Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.804372 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-55fc64bfdd-779rm" event={"ID":"a86ea592-5d80-4415-8c1f-5471274999d4","Type":"ContainerStarted","Data":"d3893fc79990dd3419bf6736a3fa0bf8afbf69c440247b6edc087231abaae801"} Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.805210 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-55fc64bfdd-779rm" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.834521 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7c696dc874-7894k" event={"ID":"246893e5-50e3-402e-8b4f-f21b31af732e","Type":"ContainerStarted","Data":"c2e97f7a5af1583a24465c8f20a8e5e8781bf5ba31141dc4070617a176c41652"} Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.836642 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-798947df95-gjr2t" event={"ID":"eda29ee7-1690-43ed-be0f-c8aac34d6daa","Type":"ContainerStarted","Data":"07d0778674f0592d75374d5761ba6925eda84b3b0bde117a99e41e6eea18d700"} Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.839297 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6fc9b98cc8-7swm4" event={"ID":"16c7bcbb-1a4f-4c68-ade9-3cafed6d6876","Type":"ContainerStarted","Data":"6fc048149218fe366aa4f0c654c99f43aa3fa60d98b649a65547004b197b9362"} Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.843033 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-55fc64bfdd-779rm" podStartSLOduration=3.8430160989999997 podStartE2EDuration="3.843016099s" podCreationTimestamp="2025-12-06 04:00:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:00:06.833238283 +0000 UTC m=+1199.705147435" watchObservedRunningTime="2025-12-06 04:00:06.843016099 +0000 UTC m=+1199.714925251" Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.846175 4802 generic.go:334] "Generic (PLEG): container finished" podID="aa11f460-8357-4771-9941-0d1682ed0b74" containerID="f13ba3f09772ae13d7557ec46f47e01c1e51ec8e4be764cd3a1bd49bf62a18df" exitCode=0 Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.846270 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688b9f5b49-vxlxf" event={"ID":"aa11f460-8357-4771-9941-0d1682ed0b74","Type":"ContainerDied","Data":"f13ba3f09772ae13d7557ec46f47e01c1e51ec8e4be764cd3a1bd49bf62a18df"} Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.846343 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688b9f5b49-vxlxf" event={"ID":"aa11f460-8357-4771-9941-0d1682ed0b74","Type":"ContainerStarted","Data":"04057ba7dda93084cee2e5e79d0b64105810b94c48c817669bcabe78bfc85aeb"} Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.893236 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-67b6f8b59b-ghkb7" event={"ID":"a355c0a5-b839-4928-b7ce-0b2d0389190d","Type":"ContainerStarted","Data":"ddcfb67eb69643ac7b9d9c9903298efc1e945f43d8a22bf773d8f95f04598de3"} Dec 06 04:00:06 crc kubenswrapper[4802]: I1206 04:00:06.985012 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-67b6f8b59b-ghkb7"] Dec 06 04:00:07 crc kubenswrapper[4802]: I1206 04:00:07.009894 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-6fc9b98cc8-7swm4"] Dec 06 04:00:07 crc kubenswrapper[4802]: I1206 04:00:07.012474 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 04:00:07 crc kubenswrapper[4802]: I1206 04:00:07.031819 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-788b9d8f4c-rk4vn"] Dec 06 04:00:07 crc kubenswrapper[4802]: I1206 04:00:07.033014 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-788b9d8f4c-rk4vn" Dec 06 04:00:07 crc kubenswrapper[4802]: I1206 04:00:07.036083 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-internal-svc" Dec 06 04:00:07 crc kubenswrapper[4802]: I1206 04:00:07.036315 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-public-svc" Dec 06 04:00:07 crc kubenswrapper[4802]: I1206 04:00:07.056132 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-788b9d8f4c-rk4vn"] Dec 06 04:00:07 crc kubenswrapper[4802]: I1206 04:00:07.074145 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-6ddd7d8889-js75h"] Dec 06 04:00:07 crc kubenswrapper[4802]: I1206 04:00:07.075603 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6ddd7d8889-js75h" Dec 06 04:00:07 crc kubenswrapper[4802]: I1206 04:00:07.083898 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-internal-svc" Dec 06 04:00:07 crc kubenswrapper[4802]: I1206 04:00:07.084562 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-public-svc" Dec 06 04:00:07 crc kubenswrapper[4802]: I1206 04:00:07.104371 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-6ddd7d8889-js75h"] Dec 06 04:00:07 crc kubenswrapper[4802]: I1206 04:00:07.207874 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/71fc6de6-20c6-49fe-9780-3cfdb0002c62-config-data-custom\") pod \"heat-api-6ddd7d8889-js75h\" (UID: \"71fc6de6-20c6-49fe-9780-3cfdb0002c62\") " pod="openstack/heat-api-6ddd7d8889-js75h" Dec 06 04:00:07 crc kubenswrapper[4802]: I1206 04:00:07.208302 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhzmn\" (UniqueName: \"kubernetes.io/projected/71fc6de6-20c6-49fe-9780-3cfdb0002c62-kube-api-access-xhzmn\") pod \"heat-api-6ddd7d8889-js75h\" (UID: \"71fc6de6-20c6-49fe-9780-3cfdb0002c62\") " pod="openstack/heat-api-6ddd7d8889-js75h" Dec 06 04:00:07 crc kubenswrapper[4802]: I1206 04:00:07.208454 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71fc6de6-20c6-49fe-9780-3cfdb0002c62-config-data\") pod \"heat-api-6ddd7d8889-js75h\" (UID: \"71fc6de6-20c6-49fe-9780-3cfdb0002c62\") " pod="openstack/heat-api-6ddd7d8889-js75h" Dec 06 04:00:07 crc kubenswrapper[4802]: I1206 04:00:07.209195 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/71fc6de6-20c6-49fe-9780-3cfdb0002c62-public-tls-certs\") pod \"heat-api-6ddd7d8889-js75h\" (UID: \"71fc6de6-20c6-49fe-9780-3cfdb0002c62\") " pod="openstack/heat-api-6ddd7d8889-js75h" Dec 06 04:00:07 crc kubenswrapper[4802]: I1206 04:00:07.209256 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/10210aa2-6d6b-4422-ad35-32b04e1444f2-config-data-custom\") pod \"heat-cfnapi-788b9d8f4c-rk4vn\" (UID: \"10210aa2-6d6b-4422-ad35-32b04e1444f2\") " pod="openstack/heat-cfnapi-788b9d8f4c-rk4vn" Dec 06 04:00:07 crc kubenswrapper[4802]: I1206 04:00:07.209316 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/71fc6de6-20c6-49fe-9780-3cfdb0002c62-internal-tls-certs\") pod \"heat-api-6ddd7d8889-js75h\" (UID: \"71fc6de6-20c6-49fe-9780-3cfdb0002c62\") " pod="openstack/heat-api-6ddd7d8889-js75h" Dec 06 04:00:07 crc kubenswrapper[4802]: I1206 04:00:07.209656 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10210aa2-6d6b-4422-ad35-32b04e1444f2-combined-ca-bundle\") pod \"heat-cfnapi-788b9d8f4c-rk4vn\" (UID: \"10210aa2-6d6b-4422-ad35-32b04e1444f2\") " pod="openstack/heat-cfnapi-788b9d8f4c-rk4vn" Dec 06 04:00:07 crc kubenswrapper[4802]: I1206 04:00:07.210010 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49jd8\" (UniqueName: \"kubernetes.io/projected/10210aa2-6d6b-4422-ad35-32b04e1444f2-kube-api-access-49jd8\") pod \"heat-cfnapi-788b9d8f4c-rk4vn\" (UID: \"10210aa2-6d6b-4422-ad35-32b04e1444f2\") " pod="openstack/heat-cfnapi-788b9d8f4c-rk4vn" Dec 06 04:00:07 crc kubenswrapper[4802]: I1206 04:00:07.210044 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/10210aa2-6d6b-4422-ad35-32b04e1444f2-public-tls-certs\") pod \"heat-cfnapi-788b9d8f4c-rk4vn\" (UID: \"10210aa2-6d6b-4422-ad35-32b04e1444f2\") " pod="openstack/heat-cfnapi-788b9d8f4c-rk4vn" Dec 06 04:00:07 crc kubenswrapper[4802]: I1206 04:00:07.210360 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/10210aa2-6d6b-4422-ad35-32b04e1444f2-internal-tls-certs\") pod \"heat-cfnapi-788b9d8f4c-rk4vn\" (UID: \"10210aa2-6d6b-4422-ad35-32b04e1444f2\") " pod="openstack/heat-cfnapi-788b9d8f4c-rk4vn" Dec 06 04:00:07 crc kubenswrapper[4802]: I1206 04:00:07.210523 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71fc6de6-20c6-49fe-9780-3cfdb0002c62-combined-ca-bundle\") pod \"heat-api-6ddd7d8889-js75h\" (UID: \"71fc6de6-20c6-49fe-9780-3cfdb0002c62\") " pod="openstack/heat-api-6ddd7d8889-js75h" Dec 06 04:00:07 crc kubenswrapper[4802]: I1206 04:00:07.210553 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10210aa2-6d6b-4422-ad35-32b04e1444f2-config-data\") pod \"heat-cfnapi-788b9d8f4c-rk4vn\" (UID: \"10210aa2-6d6b-4422-ad35-32b04e1444f2\") " pod="openstack/heat-cfnapi-788b9d8f4c-rk4vn" Dec 06 04:00:07 crc kubenswrapper[4802]: I1206 04:00:07.212134 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 04:00:07 crc kubenswrapper[4802]: I1206 04:00:07.311888 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49jd8\" (UniqueName: \"kubernetes.io/projected/10210aa2-6d6b-4422-ad35-32b04e1444f2-kube-api-access-49jd8\") pod \"heat-cfnapi-788b9d8f4c-rk4vn\" (UID: \"10210aa2-6d6b-4422-ad35-32b04e1444f2\") " pod="openstack/heat-cfnapi-788b9d8f4c-rk4vn" Dec 06 04:00:07 crc kubenswrapper[4802]: I1206 04:00:07.311927 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/10210aa2-6d6b-4422-ad35-32b04e1444f2-public-tls-certs\") pod \"heat-cfnapi-788b9d8f4c-rk4vn\" (UID: \"10210aa2-6d6b-4422-ad35-32b04e1444f2\") " pod="openstack/heat-cfnapi-788b9d8f4c-rk4vn" Dec 06 04:00:07 crc kubenswrapper[4802]: I1206 04:00:07.311983 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/10210aa2-6d6b-4422-ad35-32b04e1444f2-internal-tls-certs\") pod \"heat-cfnapi-788b9d8f4c-rk4vn\" (UID: \"10210aa2-6d6b-4422-ad35-32b04e1444f2\") " pod="openstack/heat-cfnapi-788b9d8f4c-rk4vn" Dec 06 04:00:07 crc kubenswrapper[4802]: I1206 04:00:07.312006 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71fc6de6-20c6-49fe-9780-3cfdb0002c62-combined-ca-bundle\") pod \"heat-api-6ddd7d8889-js75h\" (UID: \"71fc6de6-20c6-49fe-9780-3cfdb0002c62\") " pod="openstack/heat-api-6ddd7d8889-js75h" Dec 06 04:00:07 crc kubenswrapper[4802]: I1206 04:00:07.312024 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10210aa2-6d6b-4422-ad35-32b04e1444f2-config-data\") pod \"heat-cfnapi-788b9d8f4c-rk4vn\" (UID: \"10210aa2-6d6b-4422-ad35-32b04e1444f2\") " pod="openstack/heat-cfnapi-788b9d8f4c-rk4vn" Dec 06 04:00:07 crc kubenswrapper[4802]: I1206 04:00:07.312062 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/71fc6de6-20c6-49fe-9780-3cfdb0002c62-config-data-custom\") pod \"heat-api-6ddd7d8889-js75h\" (UID: \"71fc6de6-20c6-49fe-9780-3cfdb0002c62\") " pod="openstack/heat-api-6ddd7d8889-js75h" Dec 06 04:00:07 crc kubenswrapper[4802]: I1206 04:00:07.312083 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhzmn\" (UniqueName: \"kubernetes.io/projected/71fc6de6-20c6-49fe-9780-3cfdb0002c62-kube-api-access-xhzmn\") pod \"heat-api-6ddd7d8889-js75h\" (UID: \"71fc6de6-20c6-49fe-9780-3cfdb0002c62\") " pod="openstack/heat-api-6ddd7d8889-js75h" Dec 06 04:00:07 crc kubenswrapper[4802]: I1206 04:00:07.312098 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71fc6de6-20c6-49fe-9780-3cfdb0002c62-config-data\") pod \"heat-api-6ddd7d8889-js75h\" (UID: \"71fc6de6-20c6-49fe-9780-3cfdb0002c62\") " pod="openstack/heat-api-6ddd7d8889-js75h" Dec 06 04:00:07 crc kubenswrapper[4802]: I1206 04:00:07.312113 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/71fc6de6-20c6-49fe-9780-3cfdb0002c62-public-tls-certs\") pod \"heat-api-6ddd7d8889-js75h\" (UID: \"71fc6de6-20c6-49fe-9780-3cfdb0002c62\") " pod="openstack/heat-api-6ddd7d8889-js75h" Dec 06 04:00:07 crc kubenswrapper[4802]: I1206 04:00:07.312129 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/10210aa2-6d6b-4422-ad35-32b04e1444f2-config-data-custom\") pod \"heat-cfnapi-788b9d8f4c-rk4vn\" (UID: \"10210aa2-6d6b-4422-ad35-32b04e1444f2\") " pod="openstack/heat-cfnapi-788b9d8f4c-rk4vn" Dec 06 04:00:07 crc kubenswrapper[4802]: I1206 04:00:07.312151 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/71fc6de6-20c6-49fe-9780-3cfdb0002c62-internal-tls-certs\") pod \"heat-api-6ddd7d8889-js75h\" (UID: \"71fc6de6-20c6-49fe-9780-3cfdb0002c62\") " pod="openstack/heat-api-6ddd7d8889-js75h" Dec 06 04:00:07 crc kubenswrapper[4802]: I1206 04:00:07.312171 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10210aa2-6d6b-4422-ad35-32b04e1444f2-combined-ca-bundle\") pod \"heat-cfnapi-788b9d8f4c-rk4vn\" (UID: \"10210aa2-6d6b-4422-ad35-32b04e1444f2\") " pod="openstack/heat-cfnapi-788b9d8f4c-rk4vn" Dec 06 04:00:07 crc kubenswrapper[4802]: I1206 04:00:07.320314 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10210aa2-6d6b-4422-ad35-32b04e1444f2-combined-ca-bundle\") pod \"heat-cfnapi-788b9d8f4c-rk4vn\" (UID: \"10210aa2-6d6b-4422-ad35-32b04e1444f2\") " pod="openstack/heat-cfnapi-788b9d8f4c-rk4vn" Dec 06 04:00:07 crc kubenswrapper[4802]: I1206 04:00:07.321423 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/71fc6de6-20c6-49fe-9780-3cfdb0002c62-internal-tls-certs\") pod \"heat-api-6ddd7d8889-js75h\" (UID: \"71fc6de6-20c6-49fe-9780-3cfdb0002c62\") " pod="openstack/heat-api-6ddd7d8889-js75h" Dec 06 04:00:07 crc kubenswrapper[4802]: I1206 04:00:07.325354 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/71fc6de6-20c6-49fe-9780-3cfdb0002c62-config-data-custom\") pod \"heat-api-6ddd7d8889-js75h\" (UID: \"71fc6de6-20c6-49fe-9780-3cfdb0002c62\") " pod="openstack/heat-api-6ddd7d8889-js75h" Dec 06 04:00:07 crc kubenswrapper[4802]: I1206 04:00:07.325941 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/71fc6de6-20c6-49fe-9780-3cfdb0002c62-public-tls-certs\") pod \"heat-api-6ddd7d8889-js75h\" (UID: \"71fc6de6-20c6-49fe-9780-3cfdb0002c62\") " pod="openstack/heat-api-6ddd7d8889-js75h" Dec 06 04:00:07 crc kubenswrapper[4802]: I1206 04:00:07.326664 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/10210aa2-6d6b-4422-ad35-32b04e1444f2-public-tls-certs\") pod \"heat-cfnapi-788b9d8f4c-rk4vn\" (UID: \"10210aa2-6d6b-4422-ad35-32b04e1444f2\") " pod="openstack/heat-cfnapi-788b9d8f4c-rk4vn" Dec 06 04:00:07 crc kubenswrapper[4802]: I1206 04:00:07.326930 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71fc6de6-20c6-49fe-9780-3cfdb0002c62-config-data\") pod \"heat-api-6ddd7d8889-js75h\" (UID: \"71fc6de6-20c6-49fe-9780-3cfdb0002c62\") " pod="openstack/heat-api-6ddd7d8889-js75h" Dec 06 04:00:07 crc kubenswrapper[4802]: I1206 04:00:07.327159 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10210aa2-6d6b-4422-ad35-32b04e1444f2-config-data\") pod \"heat-cfnapi-788b9d8f4c-rk4vn\" (UID: \"10210aa2-6d6b-4422-ad35-32b04e1444f2\") " pod="openstack/heat-cfnapi-788b9d8f4c-rk4vn" Dec 06 04:00:07 crc kubenswrapper[4802]: I1206 04:00:07.327720 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/10210aa2-6d6b-4422-ad35-32b04e1444f2-internal-tls-certs\") pod \"heat-cfnapi-788b9d8f4c-rk4vn\" (UID: \"10210aa2-6d6b-4422-ad35-32b04e1444f2\") " pod="openstack/heat-cfnapi-788b9d8f4c-rk4vn" Dec 06 04:00:07 crc kubenswrapper[4802]: I1206 04:00:07.328097 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/10210aa2-6d6b-4422-ad35-32b04e1444f2-config-data-custom\") pod \"heat-cfnapi-788b9d8f4c-rk4vn\" (UID: \"10210aa2-6d6b-4422-ad35-32b04e1444f2\") " pod="openstack/heat-cfnapi-788b9d8f4c-rk4vn" Dec 06 04:00:07 crc kubenswrapper[4802]: I1206 04:00:07.330902 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71fc6de6-20c6-49fe-9780-3cfdb0002c62-combined-ca-bundle\") pod \"heat-api-6ddd7d8889-js75h\" (UID: \"71fc6de6-20c6-49fe-9780-3cfdb0002c62\") " pod="openstack/heat-api-6ddd7d8889-js75h" Dec 06 04:00:07 crc kubenswrapper[4802]: I1206 04:00:07.332483 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49jd8\" (UniqueName: \"kubernetes.io/projected/10210aa2-6d6b-4422-ad35-32b04e1444f2-kube-api-access-49jd8\") pod \"heat-cfnapi-788b9d8f4c-rk4vn\" (UID: \"10210aa2-6d6b-4422-ad35-32b04e1444f2\") " pod="openstack/heat-cfnapi-788b9d8f4c-rk4vn" Dec 06 04:00:07 crc kubenswrapper[4802]: I1206 04:00:07.335564 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhzmn\" (UniqueName: \"kubernetes.io/projected/71fc6de6-20c6-49fe-9780-3cfdb0002c62-kube-api-access-xhzmn\") pod \"heat-api-6ddd7d8889-js75h\" (UID: \"71fc6de6-20c6-49fe-9780-3cfdb0002c62\") " pod="openstack/heat-api-6ddd7d8889-js75h" Dec 06 04:00:07 crc kubenswrapper[4802]: I1206 04:00:07.497917 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-788b9d8f4c-rk4vn" Dec 06 04:00:07 crc kubenswrapper[4802]: I1206 04:00:07.542435 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6ddd7d8889-js75h" Dec 06 04:00:07 crc kubenswrapper[4802]: I1206 04:00:07.825785 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22294441-50c6-4dc0-80f1-4b14c1fc62a9" path="/var/lib/kubelet/pods/22294441-50c6-4dc0-80f1-4b14c1fc62a9/volumes" Dec 06 04:00:07 crc kubenswrapper[4802]: I1206 04:00:07.826993 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d4f4fff-a6ec-4fef-82b4-6ba392fd297c" path="/var/lib/kubelet/pods/4d4f4fff-a6ec-4fef-82b4-6ba392fd297c/volumes" Dec 06 04:00:07 crc kubenswrapper[4802]: I1206 04:00:07.828689 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8c7901e-3653-4d82-a15f-dd12cf4ab8c5" path="/var/lib/kubelet/pods/f8c7901e-3653-4d82-a15f-dd12cf4ab8c5/volumes" Dec 06 04:00:07 crc kubenswrapper[4802]: I1206 04:00:07.829373 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 04:00:07 crc kubenswrapper[4802]: I1206 04:00:07.968679 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 04:00:08 crc kubenswrapper[4802]: I1206 04:00:08.029971 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"11b0ceda-c31c-4bc3-ad63-849047b7f80c","Type":"ContainerStarted","Data":"090052942c5bf6ac980af4ff0c0af25216c513749a0bf82817c01d320f4e67fd"} Dec 06 04:00:08 crc kubenswrapper[4802]: I1206 04:00:08.101032 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688b9f5b49-vxlxf" event={"ID":"aa11f460-8357-4771-9941-0d1682ed0b74","Type":"ContainerStarted","Data":"97f98ebf726b3b17ee80d198485eb9354459be38d160912c70ba7b948dd66281"} Dec 06 04:00:08 crc kubenswrapper[4802]: I1206 04:00:08.102211 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-688b9f5b49-vxlxf" Dec 06 04:00:08 crc kubenswrapper[4802]: I1206 04:00:08.109845 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"19c7331e-7c80-4aa6-8a9f-2a7f878c79f2","Type":"ContainerStarted","Data":"5756584fa694e804d657078130717bb74aa837694b2b010a729bca6f353f306b"} Dec 06 04:00:08 crc kubenswrapper[4802]: I1206 04:00:08.134789 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-688b9f5b49-vxlxf" podStartSLOduration=12.134771412 podStartE2EDuration="12.134771412s" podCreationTimestamp="2025-12-06 03:59:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:00:08.12476956 +0000 UTC m=+1200.996678722" watchObservedRunningTime="2025-12-06 04:00:08.134771412 +0000 UTC m=+1201.006680564" Dec 06 04:00:08 crc kubenswrapper[4802]: I1206 04:00:08.291134 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-6ddd7d8889-js75h"] Dec 06 04:00:08 crc kubenswrapper[4802]: I1206 04:00:08.323919 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-788b9d8f4c-rk4vn"] Dec 06 04:00:08 crc kubenswrapper[4802]: I1206 04:00:08.620406 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416560-zzktw" Dec 06 04:00:08 crc kubenswrapper[4802]: I1206 04:00:08.764230 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5898b86d-3906-4c87-8509-8c37fe50f544-secret-volume\") pod \"5898b86d-3906-4c87-8509-8c37fe50f544\" (UID: \"5898b86d-3906-4c87-8509-8c37fe50f544\") " Dec 06 04:00:08 crc kubenswrapper[4802]: I1206 04:00:08.764294 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5898b86d-3906-4c87-8509-8c37fe50f544-config-volume\") pod \"5898b86d-3906-4c87-8509-8c37fe50f544\" (UID: \"5898b86d-3906-4c87-8509-8c37fe50f544\") " Dec 06 04:00:08 crc kubenswrapper[4802]: I1206 04:00:08.764327 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h6qgq\" (UniqueName: \"kubernetes.io/projected/5898b86d-3906-4c87-8509-8c37fe50f544-kube-api-access-h6qgq\") pod \"5898b86d-3906-4c87-8509-8c37fe50f544\" (UID: \"5898b86d-3906-4c87-8509-8c37fe50f544\") " Dec 06 04:00:08 crc kubenswrapper[4802]: I1206 04:00:08.766163 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5898b86d-3906-4c87-8509-8c37fe50f544-config-volume" (OuterVolumeSpecName: "config-volume") pod "5898b86d-3906-4c87-8509-8c37fe50f544" (UID: "5898b86d-3906-4c87-8509-8c37fe50f544"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:00:08 crc kubenswrapper[4802]: I1206 04:00:08.768315 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5898b86d-3906-4c87-8509-8c37fe50f544-kube-api-access-h6qgq" (OuterVolumeSpecName: "kube-api-access-h6qgq") pod "5898b86d-3906-4c87-8509-8c37fe50f544" (UID: "5898b86d-3906-4c87-8509-8c37fe50f544"). InnerVolumeSpecName "kube-api-access-h6qgq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:00:08 crc kubenswrapper[4802]: I1206 04:00:08.770164 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5898b86d-3906-4c87-8509-8c37fe50f544-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "5898b86d-3906-4c87-8509-8c37fe50f544" (UID: "5898b86d-3906-4c87-8509-8c37fe50f544"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:00:08 crc kubenswrapper[4802]: I1206 04:00:08.870905 4802 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5898b86d-3906-4c87-8509-8c37fe50f544-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:08 crc kubenswrapper[4802]: I1206 04:00:08.871203 4802 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5898b86d-3906-4c87-8509-8c37fe50f544-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:08 crc kubenswrapper[4802]: I1206 04:00:08.871219 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h6qgq\" (UniqueName: \"kubernetes.io/projected/5898b86d-3906-4c87-8509-8c37fe50f544-kube-api-access-h6qgq\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:09 crc kubenswrapper[4802]: I1206 04:00:09.141368 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"19c7331e-7c80-4aa6-8a9f-2a7f878c79f2","Type":"ContainerStarted","Data":"ddcbe75780b628d7fb787fd3fffb98ece18b77de90dd1c66e128728bcd08679c"} Dec 06 04:00:09 crc kubenswrapper[4802]: I1206 04:00:09.163116 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"92c245ae-d893-44d7-99e3-69a9e58735cf","Type":"ContainerStarted","Data":"e13a683862c43802278e7dcbeef40aaa7b8d68a9d39f494ab99cf4f2aa011fb6"} Dec 06 04:00:09 crc kubenswrapper[4802]: I1206 04:00:09.169409 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416560-zzktw" Dec 06 04:00:09 crc kubenswrapper[4802]: I1206 04:00:09.173181 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416560-zzktw" event={"ID":"5898b86d-3906-4c87-8509-8c37fe50f544","Type":"ContainerDied","Data":"1e045a62ce0c233322d0a16c3e011eae154f7e843cf2c10a69ff225ab2777f70"} Dec 06 04:00:09 crc kubenswrapper[4802]: I1206 04:00:09.173249 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1e045a62ce0c233322d0a16c3e011eae154f7e843cf2c10a69ff225ab2777f70" Dec 06 04:00:09 crc kubenswrapper[4802]: I1206 04:00:09.186869 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"02888e6b-6185-45b8-84b7-8b3654270b77","Type":"ContainerStarted","Data":"46cdfc3ab04be534c92b4f2e30a819c09c186108775ce61af9ebdd040bda4686"} Dec 06 04:00:09 crc kubenswrapper[4802]: I1206 04:00:09.191567 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-788b9d8f4c-rk4vn" event={"ID":"10210aa2-6d6b-4422-ad35-32b04e1444f2","Type":"ContainerStarted","Data":"60633260f747fba795b7f807dd8bcef133ce620e9a388360cc2d2ca9c86bdfd4"} Dec 06 04:00:09 crc kubenswrapper[4802]: I1206 04:00:09.219438 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6ddd7d8889-js75h" event={"ID":"71fc6de6-20c6-49fe-9780-3cfdb0002c62","Type":"ContainerStarted","Data":"2685521eb228c0895fb0735786ff10b5fba397e14886d5bb7c7b8ed799ff8aa9"} Dec 06 04:00:10 crc kubenswrapper[4802]: I1206 04:00:10.230382 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"92c245ae-d893-44d7-99e3-69a9e58735cf","Type":"ContainerStarted","Data":"127bf7f0fc32a61ae6e1b8d256bc5b84593d18b795e3ab5473c19d2276f6913a"} Dec 06 04:00:10 crc kubenswrapper[4802]: I1206 04:00:10.232628 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"02888e6b-6185-45b8-84b7-8b3654270b77","Type":"ContainerStarted","Data":"f3f5299661e5b8912566fe5b4f41b4fe58c5ab7ec6cfde9b828a03ea4cf4450d"} Dec 06 04:00:10 crc kubenswrapper[4802]: I1206 04:00:10.261207 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=6.261187345 podStartE2EDuration="6.261187345s" podCreationTimestamp="2025-12-06 04:00:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:00:10.249786214 +0000 UTC m=+1203.121695366" watchObservedRunningTime="2025-12-06 04:00:10.261187345 +0000 UTC m=+1203.133096497" Dec 06 04:00:11 crc kubenswrapper[4802]: I1206 04:00:11.248376 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"19c7331e-7c80-4aa6-8a9f-2a7f878c79f2","Type":"ContainerStarted","Data":"70750c75be0ca21966959ce15b4ff21535e9a43f2be719b9c34dbcce6d6ef4d2"} Dec 06 04:00:11 crc kubenswrapper[4802]: I1206 04:00:11.272793 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"11b0ceda-c31c-4bc3-ad63-849047b7f80c","Type":"ContainerStarted","Data":"64ed0713fa1f49302593eb067506d0753fe92e92791f4939fbc92371d8519cf1"} Dec 06 04:00:11 crc kubenswrapper[4802]: I1206 04:00:11.288917 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.28889853 podStartE2EDuration="6.28889853s" podCreationTimestamp="2025-12-06 04:00:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:00:11.28157345 +0000 UTC m=+1204.153482602" watchObservedRunningTime="2025-12-06 04:00:11.28889853 +0000 UTC m=+1204.160807682" Dec 06 04:00:11 crc kubenswrapper[4802]: I1206 04:00:11.479105 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 04:00:12 crc kubenswrapper[4802]: I1206 04:00:12.082428 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-686fbdfd7f-sfrn5" Dec 06 04:00:12 crc kubenswrapper[4802]: I1206 04:00:12.085525 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-686fbdfd7f-sfrn5" Dec 06 04:00:12 crc kubenswrapper[4802]: I1206 04:00:12.310894 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-api-6fc9b98cc8-7swm4" podUID="16c7bcbb-1a4f-4c68-ade9-3cafed6d6876" containerName="heat-api" containerID="cri-o://eac8dc56792c93de7fdbf9fe1bd17b57d01e6e1351e6557d48f57e2a4994ddb6" gracePeriod=60 Dec 06 04:00:12 crc kubenswrapper[4802]: I1206 04:00:12.311232 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6fc9b98cc8-7swm4" event={"ID":"16c7bcbb-1a4f-4c68-ade9-3cafed6d6876","Type":"ContainerStarted","Data":"eac8dc56792c93de7fdbf9fe1bd17b57d01e6e1351e6557d48f57e2a4994ddb6"} Dec 06 04:00:12 crc kubenswrapper[4802]: I1206 04:00:12.311292 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-6fc9b98cc8-7swm4" Dec 06 04:00:12 crc kubenswrapper[4802]: I1206 04:00:12.323357 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"11b0ceda-c31c-4bc3-ad63-849047b7f80c","Type":"ContainerStarted","Data":"c55ccc53f4da5768babf5a455a6f4c6821e3ce0f9dd3a5706c5e2a72b5f38e9e"} Dec 06 04:00:12 crc kubenswrapper[4802]: I1206 04:00:12.338215 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-6fc9b98cc8-7swm4" podStartSLOduration=11.159229777 podStartE2EDuration="16.338196973s" podCreationTimestamp="2025-12-06 03:59:56 +0000 UTC" firstStartedPulling="2025-12-06 04:00:05.939773072 +0000 UTC m=+1198.811682224" lastFinishedPulling="2025-12-06 04:00:11.118740268 +0000 UTC m=+1203.990649420" observedRunningTime="2025-12-06 04:00:12.33662504 +0000 UTC m=+1205.208534212" watchObservedRunningTime="2025-12-06 04:00:12.338196973 +0000 UTC m=+1205.210106125" Dec 06 04:00:12 crc kubenswrapper[4802]: I1206 04:00:12.343530 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"02888e6b-6185-45b8-84b7-8b3654270b77","Type":"ContainerStarted","Data":"ba86e26ef668f707cbdd8321ef39b4f8dcf247b0dae661c9f926d2b193f3a01c"} Dec 06 04:00:12 crc kubenswrapper[4802]: I1206 04:00:12.347943 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-788b9d8f4c-rk4vn" event={"ID":"10210aa2-6d6b-4422-ad35-32b04e1444f2","Type":"ContainerStarted","Data":"bcc509fc78a7d21afad0b17b4f49cd86aae591338802418a3dd368aec1e62402"} Dec 06 04:00:12 crc kubenswrapper[4802]: I1206 04:00:12.348691 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-788b9d8f4c-rk4vn" Dec 06 04:00:12 crc kubenswrapper[4802]: I1206 04:00:12.351113 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6ddd7d8889-js75h" event={"ID":"71fc6de6-20c6-49fe-9780-3cfdb0002c62","Type":"ContainerStarted","Data":"d4dd185864bc747bb1f1d6cb968e8795c15337a4e7691a22379304774c2e4c92"} Dec 06 04:00:12 crc kubenswrapper[4802]: I1206 04:00:12.351324 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-6ddd7d8889-js75h" Dec 06 04:00:12 crc kubenswrapper[4802]: I1206 04:00:12.352283 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-67b6f8b59b-ghkb7" event={"ID":"a355c0a5-b839-4928-b7ce-0b2d0389190d","Type":"ContainerStarted","Data":"1ceeff302da4a4b548a82ccd407f19272be54b13244b5e74c534d653c4230eb4"} Dec 06 04:00:12 crc kubenswrapper[4802]: I1206 04:00:12.352422 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-cfnapi-67b6f8b59b-ghkb7" podUID="a355c0a5-b839-4928-b7ce-0b2d0389190d" containerName="heat-cfnapi" containerID="cri-o://1ceeff302da4a4b548a82ccd407f19272be54b13244b5e74c534d653c4230eb4" gracePeriod=60 Dec 06 04:00:12 crc kubenswrapper[4802]: I1206 04:00:12.352701 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-67b6f8b59b-ghkb7" Dec 06 04:00:12 crc kubenswrapper[4802]: I1206 04:00:12.356068 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7c696dc874-7894k" event={"ID":"246893e5-50e3-402e-8b4f-f21b31af732e","Type":"ContainerStarted","Data":"0002b0aa701e624f219f99a56bb534bc5d7a426b92f77190811c73a9247cf198"} Dec 06 04:00:12 crc kubenswrapper[4802]: I1206 04:00:12.356217 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-7c696dc874-7894k" Dec 06 04:00:12 crc kubenswrapper[4802]: I1206 04:00:12.367971 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-798947df95-gjr2t" event={"ID":"eda29ee7-1690-43ed-be0f-c8aac34d6daa","Type":"ContainerStarted","Data":"e9d44e94aca910a1047cfa3151a5f1b3e0b2522dcda077ddbaacde5cb0c97103"} Dec 06 04:00:12 crc kubenswrapper[4802]: I1206 04:00:12.368474 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-798947df95-gjr2t" Dec 06 04:00:12 crc kubenswrapper[4802]: I1206 04:00:12.383008 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=7.382986972 podStartE2EDuration="7.382986972s" podCreationTimestamp="2025-12-06 04:00:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:00:12.362044622 +0000 UTC m=+1205.233953774" watchObservedRunningTime="2025-12-06 04:00:12.382986972 +0000 UTC m=+1205.254896124" Dec 06 04:00:12 crc kubenswrapper[4802]: I1206 04:00:12.389208 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-788b9d8f4c-rk4vn" podStartSLOduration=3.59094239 podStartE2EDuration="6.389189801s" podCreationTimestamp="2025-12-06 04:00:06 +0000 UTC" firstStartedPulling="2025-12-06 04:00:08.329574175 +0000 UTC m=+1201.201483327" lastFinishedPulling="2025-12-06 04:00:11.127821586 +0000 UTC m=+1203.999730738" observedRunningTime="2025-12-06 04:00:12.388679357 +0000 UTC m=+1205.260588509" watchObservedRunningTime="2025-12-06 04:00:12.389189801 +0000 UTC m=+1205.261098953" Dec 06 04:00:12 crc kubenswrapper[4802]: I1206 04:00:12.432428 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-67b6f8b59b-ghkb7" podStartSLOduration=11.349194318 podStartE2EDuration="16.432407627s" podCreationTimestamp="2025-12-06 03:59:56 +0000 UTC" firstStartedPulling="2025-12-06 04:00:06.04366156 +0000 UTC m=+1198.915570712" lastFinishedPulling="2025-12-06 04:00:11.126874869 +0000 UTC m=+1203.998784021" observedRunningTime="2025-12-06 04:00:12.404527258 +0000 UTC m=+1205.276436410" watchObservedRunningTime="2025-12-06 04:00:12.432407627 +0000 UTC m=+1205.304316779" Dec 06 04:00:12 crc kubenswrapper[4802]: I1206 04:00:12.475058 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-688b9f5b49-vxlxf" Dec 06 04:00:12 crc kubenswrapper[4802]: I1206 04:00:12.483622 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-6ddd7d8889-js75h" podStartSLOduration=2.69413371 podStartE2EDuration="5.483600891s" podCreationTimestamp="2025-12-06 04:00:07 +0000 UTC" firstStartedPulling="2025-12-06 04:00:08.338119988 +0000 UTC m=+1201.210029140" lastFinishedPulling="2025-12-06 04:00:11.127587179 +0000 UTC m=+1203.999496321" observedRunningTime="2025-12-06 04:00:12.421588522 +0000 UTC m=+1205.293497674" watchObservedRunningTime="2025-12-06 04:00:12.483600891 +0000 UTC m=+1205.355510043" Dec 06 04:00:12 crc kubenswrapper[4802]: I1206 04:00:12.496857 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-7c696dc874-7894k" podStartSLOduration=4.208044245 podStartE2EDuration="9.496838371s" podCreationTimestamp="2025-12-06 04:00:03 +0000 UTC" firstStartedPulling="2025-12-06 04:00:05.8298785 +0000 UTC m=+1198.701787652" lastFinishedPulling="2025-12-06 04:00:11.118672626 +0000 UTC m=+1203.990581778" observedRunningTime="2025-12-06 04:00:12.450365856 +0000 UTC m=+1205.322275008" watchObservedRunningTime="2025-12-06 04:00:12.496838371 +0000 UTC m=+1205.368747523" Dec 06 04:00:12 crc kubenswrapper[4802]: I1206 04:00:12.509850 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-798947df95-gjr2t" podStartSLOduration=4.343464592 podStartE2EDuration="9.509832265s" podCreationTimestamp="2025-12-06 04:00:03 +0000 UTC" firstStartedPulling="2025-12-06 04:00:05.959692874 +0000 UTC m=+1198.831602026" lastFinishedPulling="2025-12-06 04:00:11.126060547 +0000 UTC m=+1203.997969699" observedRunningTime="2025-12-06 04:00:12.476184169 +0000 UTC m=+1205.348093321" watchObservedRunningTime="2025-12-06 04:00:12.509832265 +0000 UTC m=+1205.381741427" Dec 06 04:00:12 crc kubenswrapper[4802]: I1206 04:00:12.540575 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-h9zv8"] Dec 06 04:00:12 crc kubenswrapper[4802]: I1206 04:00:12.540842 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6578955fd5-h9zv8" podUID="5a5637e5-3258-4eab-9438-479d0aac8099" containerName="dnsmasq-dns" containerID="cri-o://26c2f7af0077ebae0eb9505c78dae937555d12b3ce110a02311e89cbbfae5345" gracePeriod=10 Dec 06 04:00:13 crc kubenswrapper[4802]: I1206 04:00:13.375731 4802 generic.go:334] "Generic (PLEG): container finished" podID="5a5637e5-3258-4eab-9438-479d0aac8099" containerID="26c2f7af0077ebae0eb9505c78dae937555d12b3ce110a02311e89cbbfae5345" exitCode=0 Dec 06 04:00:13 crc kubenswrapper[4802]: I1206 04:00:13.375793 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-h9zv8" event={"ID":"5a5637e5-3258-4eab-9438-479d0aac8099","Type":"ContainerDied","Data":"26c2f7af0077ebae0eb9505c78dae937555d12b3ce110a02311e89cbbfae5345"} Dec 06 04:00:13 crc kubenswrapper[4802]: I1206 04:00:13.377848 4802 generic.go:334] "Generic (PLEG): container finished" podID="246893e5-50e3-402e-8b4f-f21b31af732e" containerID="0002b0aa701e624f219f99a56bb534bc5d7a426b92f77190811c73a9247cf198" exitCode=1 Dec 06 04:00:13 crc kubenswrapper[4802]: I1206 04:00:13.377920 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7c696dc874-7894k" event={"ID":"246893e5-50e3-402e-8b4f-f21b31af732e","Type":"ContainerDied","Data":"0002b0aa701e624f219f99a56bb534bc5d7a426b92f77190811c73a9247cf198"} Dec 06 04:00:13 crc kubenswrapper[4802]: I1206 04:00:13.378317 4802 scope.go:117] "RemoveContainer" containerID="0002b0aa701e624f219f99a56bb534bc5d7a426b92f77190811c73a9247cf198" Dec 06 04:00:13 crc kubenswrapper[4802]: I1206 04:00:13.379662 4802 generic.go:334] "Generic (PLEG): container finished" podID="eda29ee7-1690-43ed-be0f-c8aac34d6daa" containerID="e9d44e94aca910a1047cfa3151a5f1b3e0b2522dcda077ddbaacde5cb0c97103" exitCode=1 Dec 06 04:00:13 crc kubenswrapper[4802]: I1206 04:00:13.379781 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-798947df95-gjr2t" event={"ID":"eda29ee7-1690-43ed-be0f-c8aac34d6daa","Type":"ContainerDied","Data":"e9d44e94aca910a1047cfa3151a5f1b3e0b2522dcda077ddbaacde5cb0c97103"} Dec 06 04:00:13 crc kubenswrapper[4802]: I1206 04:00:13.380539 4802 scope.go:117] "RemoveContainer" containerID="e9d44e94aca910a1047cfa3151a5f1b3e0b2522dcda077ddbaacde5cb0c97103" Dec 06 04:00:13 crc kubenswrapper[4802]: I1206 04:00:13.989165 4802 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6578955fd5-h9zv8" podUID="5a5637e5-3258-4eab-9438-479d0aac8099" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.161:5353: connect: connection refused" Dec 06 04:00:14 crc kubenswrapper[4802]: I1206 04:00:14.282505 4802 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-api-798947df95-gjr2t" Dec 06 04:00:14 crc kubenswrapper[4802]: I1206 04:00:14.346541 4802 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-cfnapi-7c696dc874-7894k" Dec 06 04:00:15 crc kubenswrapper[4802]: I1206 04:00:15.144982 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 06 04:00:15 crc kubenswrapper[4802]: I1206 04:00:15.396963 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-798947df95-gjr2t" event={"ID":"eda29ee7-1690-43ed-be0f-c8aac34d6daa","Type":"ContainerStarted","Data":"9d9867310e246ceb74a94766b213bec68224ed5bf248ef2c905af51fa55f0841"} Dec 06 04:00:15 crc kubenswrapper[4802]: I1206 04:00:15.404080 4802 generic.go:334] "Generic (PLEG): container finished" podID="16c7bcbb-1a4f-4c68-ade9-3cafed6d6876" containerID="eac8dc56792c93de7fdbf9fe1bd17b57d01e6e1351e6557d48f57e2a4994ddb6" exitCode=0 Dec 06 04:00:15 crc kubenswrapper[4802]: I1206 04:00:15.404165 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6fc9b98cc8-7swm4" event={"ID":"16c7bcbb-1a4f-4c68-ade9-3cafed6d6876","Type":"ContainerDied","Data":"eac8dc56792c93de7fdbf9fe1bd17b57d01e6e1351e6557d48f57e2a4994ddb6"} Dec 06 04:00:15 crc kubenswrapper[4802]: I1206 04:00:15.411672 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"11b0ceda-c31c-4bc3-ad63-849047b7f80c","Type":"ContainerStarted","Data":"3c540921637b4fd0bcd4f4eb90cc1ec51a09d48dbcc5e4205f5bca162358eae7"} Dec 06 04:00:15 crc kubenswrapper[4802]: I1206 04:00:15.414641 4802 generic.go:334] "Generic (PLEG): container finished" podID="a355c0a5-b839-4928-b7ce-0b2d0389190d" containerID="1ceeff302da4a4b548a82ccd407f19272be54b13244b5e74c534d653c4230eb4" exitCode=0 Dec 06 04:00:15 crc kubenswrapper[4802]: I1206 04:00:15.414743 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-67b6f8b59b-ghkb7" event={"ID":"a355c0a5-b839-4928-b7ce-0b2d0389190d","Type":"ContainerDied","Data":"1ceeff302da4a4b548a82ccd407f19272be54b13244b5e74c534d653c4230eb4"} Dec 06 04:00:15 crc kubenswrapper[4802]: I1206 04:00:15.417244 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7c696dc874-7894k" event={"ID":"246893e5-50e3-402e-8b4f-f21b31af732e","Type":"ContainerStarted","Data":"bead59f8707df720f20d9b3824500ec2f9ea32ae0f3fbde1604209abd9f3a330"} Dec 06 04:00:15 crc kubenswrapper[4802]: I1206 04:00:15.543941 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 06 04:00:15 crc kubenswrapper[4802]: I1206 04:00:15.723072 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-h9zv8" Dec 06 04:00:15 crc kubenswrapper[4802]: I1206 04:00:15.823175 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hbzws\" (UniqueName: \"kubernetes.io/projected/5a5637e5-3258-4eab-9438-479d0aac8099-kube-api-access-hbzws\") pod \"5a5637e5-3258-4eab-9438-479d0aac8099\" (UID: \"5a5637e5-3258-4eab-9438-479d0aac8099\") " Dec 06 04:00:15 crc kubenswrapper[4802]: I1206 04:00:15.823217 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5a5637e5-3258-4eab-9438-479d0aac8099-dns-svc\") pod \"5a5637e5-3258-4eab-9438-479d0aac8099\" (UID: \"5a5637e5-3258-4eab-9438-479d0aac8099\") " Dec 06 04:00:15 crc kubenswrapper[4802]: I1206 04:00:15.823296 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5a5637e5-3258-4eab-9438-479d0aac8099-ovsdbserver-nb\") pod \"5a5637e5-3258-4eab-9438-479d0aac8099\" (UID: \"5a5637e5-3258-4eab-9438-479d0aac8099\") " Dec 06 04:00:15 crc kubenswrapper[4802]: I1206 04:00:15.823348 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5a5637e5-3258-4eab-9438-479d0aac8099-dns-swift-storage-0\") pod \"5a5637e5-3258-4eab-9438-479d0aac8099\" (UID: \"5a5637e5-3258-4eab-9438-479d0aac8099\") " Dec 06 04:00:15 crc kubenswrapper[4802]: I1206 04:00:15.823443 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a5637e5-3258-4eab-9438-479d0aac8099-config\") pod \"5a5637e5-3258-4eab-9438-479d0aac8099\" (UID: \"5a5637e5-3258-4eab-9438-479d0aac8099\") " Dec 06 04:00:15 crc kubenswrapper[4802]: I1206 04:00:15.823532 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5a5637e5-3258-4eab-9438-479d0aac8099-ovsdbserver-sb\") pod \"5a5637e5-3258-4eab-9438-479d0aac8099\" (UID: \"5a5637e5-3258-4eab-9438-479d0aac8099\") " Dec 06 04:00:15 crc kubenswrapper[4802]: I1206 04:00:15.863350 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a5637e5-3258-4eab-9438-479d0aac8099-kube-api-access-hbzws" (OuterVolumeSpecName: "kube-api-access-hbzws") pod "5a5637e5-3258-4eab-9438-479d0aac8099" (UID: "5a5637e5-3258-4eab-9438-479d0aac8099"). InnerVolumeSpecName "kube-api-access-hbzws". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:00:15 crc kubenswrapper[4802]: I1206 04:00:15.909998 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a5637e5-3258-4eab-9438-479d0aac8099-config" (OuterVolumeSpecName: "config") pod "5a5637e5-3258-4eab-9438-479d0aac8099" (UID: "5a5637e5-3258-4eab-9438-479d0aac8099"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:00:15 crc kubenswrapper[4802]: I1206 04:00:15.926569 4802 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a5637e5-3258-4eab-9438-479d0aac8099-config\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:15 crc kubenswrapper[4802]: I1206 04:00:15.926609 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hbzws\" (UniqueName: \"kubernetes.io/projected/5a5637e5-3258-4eab-9438-479d0aac8099-kube-api-access-hbzws\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:15 crc kubenswrapper[4802]: I1206 04:00:15.940714 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a5637e5-3258-4eab-9438-479d0aac8099-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5a5637e5-3258-4eab-9438-479d0aac8099" (UID: "5a5637e5-3258-4eab-9438-479d0aac8099"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:00:15 crc kubenswrapper[4802]: I1206 04:00:15.965234 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a5637e5-3258-4eab-9438-479d0aac8099-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5a5637e5-3258-4eab-9438-479d0aac8099" (UID: "5a5637e5-3258-4eab-9438-479d0aac8099"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:00:15 crc kubenswrapper[4802]: I1206 04:00:15.983199 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a5637e5-3258-4eab-9438-479d0aac8099-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "5a5637e5-3258-4eab-9438-479d0aac8099" (UID: "5a5637e5-3258-4eab-9438-479d0aac8099"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.007252 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a5637e5-3258-4eab-9438-479d0aac8099-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5a5637e5-3258-4eab-9438-479d0aac8099" (UID: "5a5637e5-3258-4eab-9438-479d0aac8099"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.027811 4802 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5a5637e5-3258-4eab-9438-479d0aac8099-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.027840 4802 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5a5637e5-3258-4eab-9438-479d0aac8099-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.027849 4802 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5a5637e5-3258-4eab-9438-479d0aac8099-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.027859 4802 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5a5637e5-3258-4eab-9438-479d0aac8099-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.148968 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.149204 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.173213 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6fc9b98cc8-7swm4" Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.214194 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.214302 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.266034 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-67b6f8b59b-ghkb7" Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.336535 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-thv7m\" (UniqueName: \"kubernetes.io/projected/16c7bcbb-1a4f-4c68-ade9-3cafed6d6876-kube-api-access-thv7m\") pod \"16c7bcbb-1a4f-4c68-ade9-3cafed6d6876\" (UID: \"16c7bcbb-1a4f-4c68-ade9-3cafed6d6876\") " Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.338904 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a355c0a5-b839-4928-b7ce-0b2d0389190d-combined-ca-bundle\") pod \"a355c0a5-b839-4928-b7ce-0b2d0389190d\" (UID: \"a355c0a5-b839-4928-b7ce-0b2d0389190d\") " Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.338998 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16c7bcbb-1a4f-4c68-ade9-3cafed6d6876-combined-ca-bundle\") pod \"16c7bcbb-1a4f-4c68-ade9-3cafed6d6876\" (UID: \"16c7bcbb-1a4f-4c68-ade9-3cafed6d6876\") " Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.339025 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a355c0a5-b839-4928-b7ce-0b2d0389190d-config-data-custom\") pod \"a355c0a5-b839-4928-b7ce-0b2d0389190d\" (UID: \"a355c0a5-b839-4928-b7ce-0b2d0389190d\") " Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.339057 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wl2gt\" (UniqueName: \"kubernetes.io/projected/a355c0a5-b839-4928-b7ce-0b2d0389190d-kube-api-access-wl2gt\") pod \"a355c0a5-b839-4928-b7ce-0b2d0389190d\" (UID: \"a355c0a5-b839-4928-b7ce-0b2d0389190d\") " Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.339090 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a355c0a5-b839-4928-b7ce-0b2d0389190d-config-data\") pod \"a355c0a5-b839-4928-b7ce-0b2d0389190d\" (UID: \"a355c0a5-b839-4928-b7ce-0b2d0389190d\") " Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.339113 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16c7bcbb-1a4f-4c68-ade9-3cafed6d6876-config-data\") pod \"16c7bcbb-1a4f-4c68-ade9-3cafed6d6876\" (UID: \"16c7bcbb-1a4f-4c68-ade9-3cafed6d6876\") " Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.342235 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/16c7bcbb-1a4f-4c68-ade9-3cafed6d6876-config-data-custom\") pod \"16c7bcbb-1a4f-4c68-ade9-3cafed6d6876\" (UID: \"16c7bcbb-1a4f-4c68-ade9-3cafed6d6876\") " Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.344707 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a355c0a5-b839-4928-b7ce-0b2d0389190d-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a355c0a5-b839-4928-b7ce-0b2d0389190d" (UID: "a355c0a5-b839-4928-b7ce-0b2d0389190d"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.345353 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16c7bcbb-1a4f-4c68-ade9-3cafed6d6876-kube-api-access-thv7m" (OuterVolumeSpecName: "kube-api-access-thv7m") pod "16c7bcbb-1a4f-4c68-ade9-3cafed6d6876" (UID: "16c7bcbb-1a4f-4c68-ade9-3cafed6d6876"). InnerVolumeSpecName "kube-api-access-thv7m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.363112 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16c7bcbb-1a4f-4c68-ade9-3cafed6d6876-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "16c7bcbb-1a4f-4c68-ade9-3cafed6d6876" (UID: "16c7bcbb-1a4f-4c68-ade9-3cafed6d6876"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.369651 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a355c0a5-b839-4928-b7ce-0b2d0389190d-kube-api-access-wl2gt" (OuterVolumeSpecName: "kube-api-access-wl2gt") pod "a355c0a5-b839-4928-b7ce-0b2d0389190d" (UID: "a355c0a5-b839-4928-b7ce-0b2d0389190d"). InnerVolumeSpecName "kube-api-access-wl2gt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.383112 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16c7bcbb-1a4f-4c68-ade9-3cafed6d6876-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "16c7bcbb-1a4f-4c68-ade9-3cafed6d6876" (UID: "16c7bcbb-1a4f-4c68-ade9-3cafed6d6876"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.397024 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a355c0a5-b839-4928-b7ce-0b2d0389190d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a355c0a5-b839-4928-b7ce-0b2d0389190d" (UID: "a355c0a5-b839-4928-b7ce-0b2d0389190d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.403295 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a355c0a5-b839-4928-b7ce-0b2d0389190d-config-data" (OuterVolumeSpecName: "config-data") pod "a355c0a5-b839-4928-b7ce-0b2d0389190d" (UID: "a355c0a5-b839-4928-b7ce-0b2d0389190d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.424176 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16c7bcbb-1a4f-4c68-ade9-3cafed6d6876-config-data" (OuterVolumeSpecName: "config-data") pod "16c7bcbb-1a4f-4c68-ade9-3cafed6d6876" (UID: "16c7bcbb-1a4f-4c68-ade9-3cafed6d6876"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.425543 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-67b6f8b59b-ghkb7" event={"ID":"a355c0a5-b839-4928-b7ce-0b2d0389190d","Type":"ContainerDied","Data":"ddcfb67eb69643ac7b9d9c9903298efc1e945f43d8a22bf773d8f95f04598de3"} Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.425566 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-67b6f8b59b-ghkb7" Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.425588 4802 scope.go:117] "RemoveContainer" containerID="1ceeff302da4a4b548a82ccd407f19272be54b13244b5e74c534d653c4230eb4" Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.431914 4802 generic.go:334] "Generic (PLEG): container finished" podID="246893e5-50e3-402e-8b4f-f21b31af732e" containerID="bead59f8707df720f20d9b3824500ec2f9ea32ae0f3fbde1604209abd9f3a330" exitCode=1 Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.431976 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7c696dc874-7894k" event={"ID":"246893e5-50e3-402e-8b4f-f21b31af732e","Type":"ContainerDied","Data":"bead59f8707df720f20d9b3824500ec2f9ea32ae0f3fbde1604209abd9f3a330"} Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.432419 4802 scope.go:117] "RemoveContainer" containerID="bead59f8707df720f20d9b3824500ec2f9ea32ae0f3fbde1604209abd9f3a330" Dec 06 04:00:16 crc kubenswrapper[4802]: E1206 04:00:16.432612 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-7c696dc874-7894k_openstack(246893e5-50e3-402e-8b4f-f21b31af732e)\"" pod="openstack/heat-cfnapi-7c696dc874-7894k" podUID="246893e5-50e3-402e-8b4f-f21b31af732e" Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.435593 4802 generic.go:334] "Generic (PLEG): container finished" podID="eda29ee7-1690-43ed-be0f-c8aac34d6daa" containerID="9d9867310e246ceb74a94766b213bec68224ed5bf248ef2c905af51fa55f0841" exitCode=1 Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.435643 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-798947df95-gjr2t" event={"ID":"eda29ee7-1690-43ed-be0f-c8aac34d6daa","Type":"ContainerDied","Data":"9d9867310e246ceb74a94766b213bec68224ed5bf248ef2c905af51fa55f0841"} Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.436314 4802 scope.go:117] "RemoveContainer" containerID="9d9867310e246ceb74a94766b213bec68224ed5bf248ef2c905af51fa55f0841" Dec 06 04:00:16 crc kubenswrapper[4802]: E1206 04:00:16.436517 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-798947df95-gjr2t_openstack(eda29ee7-1690-43ed-be0f-c8aac34d6daa)\"" pod="openstack/heat-api-798947df95-gjr2t" podUID="eda29ee7-1690-43ed-be0f-c8aac34d6daa" Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.440352 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6fc9b98cc8-7swm4" event={"ID":"16c7bcbb-1a4f-4c68-ade9-3cafed6d6876","Type":"ContainerDied","Data":"6fc048149218fe366aa4f0c654c99f43aa3fa60d98b649a65547004b197b9362"} Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.440446 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6fc9b98cc8-7swm4" Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.442977 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-h9zv8" Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.443200 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-h9zv8" event={"ID":"5a5637e5-3258-4eab-9438-479d0aac8099","Type":"ContainerDied","Data":"d0dafd87f3f702437aa75029b5742ef894c8f8b0d040c8a67e065c303ed988cf"} Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.444108 4802 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16c7bcbb-1a4f-4c68-ade9-3cafed6d6876-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.444128 4802 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a355c0a5-b839-4928-b7ce-0b2d0389190d-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.444138 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wl2gt\" (UniqueName: \"kubernetes.io/projected/a355c0a5-b839-4928-b7ce-0b2d0389190d-kube-api-access-wl2gt\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.444149 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a355c0a5-b839-4928-b7ce-0b2d0389190d-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.444158 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16c7bcbb-1a4f-4c68-ade9-3cafed6d6876-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.444166 4802 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/16c7bcbb-1a4f-4c68-ade9-3cafed6d6876-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.444175 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-thv7m\" (UniqueName: \"kubernetes.io/projected/16c7bcbb-1a4f-4c68-ade9-3cafed6d6876-kube-api-access-thv7m\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.444183 4802 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a355c0a5-b839-4928-b7ce-0b2d0389190d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.444272 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.445904 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.479436 4802 scope.go:117] "RemoveContainer" containerID="0002b0aa701e624f219f99a56bb534bc5d7a426b92f77190811c73a9247cf198" Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.537852 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-67b6f8b59b-ghkb7"] Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.553974 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-67b6f8b59b-ghkb7"] Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.569671 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-h9zv8"] Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.578090 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-h9zv8"] Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.587627 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-6fc9b98cc8-7swm4"] Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.595406 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-6fc9b98cc8-7swm4"] Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.830553 4802 scope.go:117] "RemoveContainer" containerID="e9d44e94aca910a1047cfa3151a5f1b3e0b2522dcda077ddbaacde5cb0c97103" Dec 06 04:00:16 crc kubenswrapper[4802]: I1206 04:00:16.926999 4802 scope.go:117] "RemoveContainer" containerID="eac8dc56792c93de7fdbf9fe1bd17b57d01e6e1351e6557d48f57e2a4994ddb6" Dec 06 04:00:17 crc kubenswrapper[4802]: I1206 04:00:17.013783 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 06 04:00:17 crc kubenswrapper[4802]: I1206 04:00:17.013827 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 06 04:00:17 crc kubenswrapper[4802]: I1206 04:00:17.044100 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 06 04:00:17 crc kubenswrapper[4802]: I1206 04:00:17.079979 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 06 04:00:17 crc kubenswrapper[4802]: I1206 04:00:17.102477 4802 scope.go:117] "RemoveContainer" containerID="26c2f7af0077ebae0eb9505c78dae937555d12b3ce110a02311e89cbbfae5345" Dec 06 04:00:17 crc kubenswrapper[4802]: I1206 04:00:17.280065 4802 scope.go:117] "RemoveContainer" containerID="2f8f82847a070c0100fa0db88d5b3b94d65dd663421a778e3015430b4291f86e" Dec 06 04:00:17 crc kubenswrapper[4802]: I1206 04:00:17.480973 4802 scope.go:117] "RemoveContainer" containerID="9d9867310e246ceb74a94766b213bec68224ed5bf248ef2c905af51fa55f0841" Dec 06 04:00:17 crc kubenswrapper[4802]: E1206 04:00:17.481638 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-798947df95-gjr2t_openstack(eda29ee7-1690-43ed-be0f-c8aac34d6daa)\"" pod="openstack/heat-api-798947df95-gjr2t" podUID="eda29ee7-1690-43ed-be0f-c8aac34d6daa" Dec 06 04:00:17 crc kubenswrapper[4802]: I1206 04:00:17.527829 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16c7bcbb-1a4f-4c68-ade9-3cafed6d6876" path="/var/lib/kubelet/pods/16c7bcbb-1a4f-4c68-ade9-3cafed6d6876/volumes" Dec 06 04:00:17 crc kubenswrapper[4802]: I1206 04:00:17.528729 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a5637e5-3258-4eab-9438-479d0aac8099" path="/var/lib/kubelet/pods/5a5637e5-3258-4eab-9438-479d0aac8099/volumes" Dec 06 04:00:17 crc kubenswrapper[4802]: I1206 04:00:17.530680 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a355c0a5-b839-4928-b7ce-0b2d0389190d" path="/var/lib/kubelet/pods/a355c0a5-b839-4928-b7ce-0b2d0389190d/volumes" Dec 06 04:00:17 crc kubenswrapper[4802]: I1206 04:00:17.531466 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-5d96658447-rg8zs" Dec 06 04:00:17 crc kubenswrapper[4802]: I1206 04:00:17.536978 4802 scope.go:117] "RemoveContainer" containerID="bead59f8707df720f20d9b3824500ec2f9ea32ae0f3fbde1604209abd9f3a330" Dec 06 04:00:17 crc kubenswrapper[4802]: E1206 04:00:17.537175 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-7c696dc874-7894k_openstack(246893e5-50e3-402e-8b4f-f21b31af732e)\"" pod="openstack/heat-cfnapi-7c696dc874-7894k" podUID="246893e5-50e3-402e-8b4f-f21b31af732e" Dec 06 04:00:17 crc kubenswrapper[4802]: I1206 04:00:17.553106 4802 generic.go:334] "Generic (PLEG): container finished" podID="b364aef0-231f-472a-a8c9-51a5199c0997" containerID="1bd700182ce971dd2c603b23e1a7c8370de9caf2e1ce639678041c207ab36f16" exitCode=0 Dec 06 04:00:17 crc kubenswrapper[4802]: I1206 04:00:17.554729 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-844df74668-r698z" event={"ID":"b364aef0-231f-472a-a8c9-51a5199c0997","Type":"ContainerDied","Data":"1bd700182ce971dd2c603b23e1a7c8370de9caf2e1ce639678041c207ab36f16"} Dec 06 04:00:17 crc kubenswrapper[4802]: I1206 04:00:17.554791 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 06 04:00:17 crc kubenswrapper[4802]: I1206 04:00:17.554810 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 06 04:00:17 crc kubenswrapper[4802]: I1206 04:00:17.669520 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-844df74668-r698z" Dec 06 04:00:17 crc kubenswrapper[4802]: I1206 04:00:17.806967 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d2599\" (UniqueName: \"kubernetes.io/projected/b364aef0-231f-472a-a8c9-51a5199c0997-kube-api-access-d2599\") pod \"b364aef0-231f-472a-a8c9-51a5199c0997\" (UID: \"b364aef0-231f-472a-a8c9-51a5199c0997\") " Dec 06 04:00:17 crc kubenswrapper[4802]: I1206 04:00:17.807264 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b364aef0-231f-472a-a8c9-51a5199c0997-httpd-config\") pod \"b364aef0-231f-472a-a8c9-51a5199c0997\" (UID: \"b364aef0-231f-472a-a8c9-51a5199c0997\") " Dec 06 04:00:17 crc kubenswrapper[4802]: I1206 04:00:17.807304 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b364aef0-231f-472a-a8c9-51a5199c0997-combined-ca-bundle\") pod \"b364aef0-231f-472a-a8c9-51a5199c0997\" (UID: \"b364aef0-231f-472a-a8c9-51a5199c0997\") " Dec 06 04:00:17 crc kubenswrapper[4802]: I1206 04:00:17.807923 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b364aef0-231f-472a-a8c9-51a5199c0997-ovndb-tls-certs\") pod \"b364aef0-231f-472a-a8c9-51a5199c0997\" (UID: \"b364aef0-231f-472a-a8c9-51a5199c0997\") " Dec 06 04:00:17 crc kubenswrapper[4802]: I1206 04:00:17.807998 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b364aef0-231f-472a-a8c9-51a5199c0997-config\") pod \"b364aef0-231f-472a-a8c9-51a5199c0997\" (UID: \"b364aef0-231f-472a-a8c9-51a5199c0997\") " Dec 06 04:00:17 crc kubenswrapper[4802]: I1206 04:00:17.834824 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b364aef0-231f-472a-a8c9-51a5199c0997-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "b364aef0-231f-472a-a8c9-51a5199c0997" (UID: "b364aef0-231f-472a-a8c9-51a5199c0997"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:00:17 crc kubenswrapper[4802]: I1206 04:00:17.843031 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b364aef0-231f-472a-a8c9-51a5199c0997-kube-api-access-d2599" (OuterVolumeSpecName: "kube-api-access-d2599") pod "b364aef0-231f-472a-a8c9-51a5199c0997" (UID: "b364aef0-231f-472a-a8c9-51a5199c0997"). InnerVolumeSpecName "kube-api-access-d2599". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:00:17 crc kubenswrapper[4802]: I1206 04:00:17.909907 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d2599\" (UniqueName: \"kubernetes.io/projected/b364aef0-231f-472a-a8c9-51a5199c0997-kube-api-access-d2599\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:17 crc kubenswrapper[4802]: I1206 04:00:17.909932 4802 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b364aef0-231f-472a-a8c9-51a5199c0997-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:17 crc kubenswrapper[4802]: I1206 04:00:17.914923 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b364aef0-231f-472a-a8c9-51a5199c0997-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "b364aef0-231f-472a-a8c9-51a5199c0997" (UID: "b364aef0-231f-472a-a8c9-51a5199c0997"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:00:17 crc kubenswrapper[4802]: I1206 04:00:17.936992 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b364aef0-231f-472a-a8c9-51a5199c0997-config" (OuterVolumeSpecName: "config") pod "b364aef0-231f-472a-a8c9-51a5199c0997" (UID: "b364aef0-231f-472a-a8c9-51a5199c0997"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:00:17 crc kubenswrapper[4802]: I1206 04:00:17.955061 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b364aef0-231f-472a-a8c9-51a5199c0997-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b364aef0-231f-472a-a8c9-51a5199c0997" (UID: "b364aef0-231f-472a-a8c9-51a5199c0997"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:00:18 crc kubenswrapper[4802]: I1206 04:00:18.011955 4802 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/b364aef0-231f-472a-a8c9-51a5199c0997-config\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:18 crc kubenswrapper[4802]: I1206 04:00:18.012037 4802 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b364aef0-231f-472a-a8c9-51a5199c0997-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:18 crc kubenswrapper[4802]: I1206 04:00:18.012056 4802 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b364aef0-231f-472a-a8c9-51a5199c0997-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:18 crc kubenswrapper[4802]: I1206 04:00:18.566053 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-844df74668-r698z" Dec 06 04:00:18 crc kubenswrapper[4802]: I1206 04:00:18.566103 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-844df74668-r698z" event={"ID":"b364aef0-231f-472a-a8c9-51a5199c0997","Type":"ContainerDied","Data":"894633cc495354c37aef490a08fa3810f4101afaa0484f5beaa33ccfc89b9bf9"} Dec 06 04:00:18 crc kubenswrapper[4802]: I1206 04:00:18.567777 4802 scope.go:117] "RemoveContainer" containerID="31dc097fa5cf9c76a286ae10dd83297f83da4c4ead70f0e1d9e17473d8ccab81" Dec 06 04:00:18 crc kubenswrapper[4802]: I1206 04:00:18.581981 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"11b0ceda-c31c-4bc3-ad63-849047b7f80c","Type":"ContainerStarted","Data":"72204b39cf3ad8d51d488287d976a47130d179e020d16c458d0c471c1ec12b87"} Dec 06 04:00:18 crc kubenswrapper[4802]: I1206 04:00:18.582385 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="11b0ceda-c31c-4bc3-ad63-849047b7f80c" containerName="proxy-httpd" containerID="cri-o://72204b39cf3ad8d51d488287d976a47130d179e020d16c458d0c471c1ec12b87" gracePeriod=30 Dec 06 04:00:18 crc kubenswrapper[4802]: I1206 04:00:18.582416 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="11b0ceda-c31c-4bc3-ad63-849047b7f80c" containerName="ceilometer-notification-agent" containerID="cri-o://c55ccc53f4da5768babf5a455a6f4c6821e3ce0f9dd3a5706c5e2a72b5f38e9e" gracePeriod=30 Dec 06 04:00:18 crc kubenswrapper[4802]: I1206 04:00:18.582398 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="11b0ceda-c31c-4bc3-ad63-849047b7f80c" containerName="sg-core" containerID="cri-o://3c540921637b4fd0bcd4f4eb90cc1ec51a09d48dbcc5e4205f5bca162358eae7" gracePeriod=30 Dec 06 04:00:18 crc kubenswrapper[4802]: I1206 04:00:18.582284 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="11b0ceda-c31c-4bc3-ad63-849047b7f80c" containerName="ceilometer-central-agent" containerID="cri-o://64ed0713fa1f49302593eb067506d0753fe92e92791f4939fbc92371d8519cf1" gracePeriod=30 Dec 06 04:00:18 crc kubenswrapper[4802]: I1206 04:00:18.586620 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 06 04:00:18 crc kubenswrapper[4802]: I1206 04:00:18.613998 4802 scope.go:117] "RemoveContainer" containerID="1bd700182ce971dd2c603b23e1a7c8370de9caf2e1ce639678041c207ab36f16" Dec 06 04:00:18 crc kubenswrapper[4802]: I1206 04:00:18.614715 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=4.278871294 podStartE2EDuration="13.614693813s" podCreationTimestamp="2025-12-06 04:00:05 +0000 UTC" firstStartedPulling="2025-12-06 04:00:07.81765372 +0000 UTC m=+1200.689562872" lastFinishedPulling="2025-12-06 04:00:17.153476239 +0000 UTC m=+1210.025385391" observedRunningTime="2025-12-06 04:00:18.610685765 +0000 UTC m=+1211.482594907" watchObservedRunningTime="2025-12-06 04:00:18.614693813 +0000 UTC m=+1211.486602955" Dec 06 04:00:18 crc kubenswrapper[4802]: I1206 04:00:18.663664 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-844df74668-r698z"] Dec 06 04:00:18 crc kubenswrapper[4802]: I1206 04:00:18.674733 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-844df74668-r698z"] Dec 06 04:00:18 crc kubenswrapper[4802]: I1206 04:00:18.820854 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 06 04:00:18 crc kubenswrapper[4802]: I1206 04:00:18.820969 4802 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 04:00:19 crc kubenswrapper[4802]: I1206 04:00:19.282399 4802 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-api-798947df95-gjr2t" Dec 06 04:00:19 crc kubenswrapper[4802]: I1206 04:00:19.282442 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-798947df95-gjr2t" Dec 06 04:00:19 crc kubenswrapper[4802]: I1206 04:00:19.283106 4802 scope.go:117] "RemoveContainer" containerID="9d9867310e246ceb74a94766b213bec68224ed5bf248ef2c905af51fa55f0841" Dec 06 04:00:19 crc kubenswrapper[4802]: E1206 04:00:19.283384 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-798947df95-gjr2t_openstack(eda29ee7-1690-43ed-be0f-c8aac34d6daa)\"" pod="openstack/heat-api-798947df95-gjr2t" podUID="eda29ee7-1690-43ed-be0f-c8aac34d6daa" Dec 06 04:00:19 crc kubenswrapper[4802]: I1206 04:00:19.346727 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-7c696dc874-7894k" Dec 06 04:00:19 crc kubenswrapper[4802]: I1206 04:00:19.347096 4802 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-cfnapi-7c696dc874-7894k" Dec 06 04:00:19 crc kubenswrapper[4802]: I1206 04:00:19.347842 4802 scope.go:117] "RemoveContainer" containerID="bead59f8707df720f20d9b3824500ec2f9ea32ae0f3fbde1604209abd9f3a330" Dec 06 04:00:19 crc kubenswrapper[4802]: E1206 04:00:19.348086 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-7c696dc874-7894k_openstack(246893e5-50e3-402e-8b4f-f21b31af732e)\"" pod="openstack/heat-cfnapi-7c696dc874-7894k" podUID="246893e5-50e3-402e-8b4f-f21b31af732e" Dec 06 04:00:19 crc kubenswrapper[4802]: I1206 04:00:19.406670 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 06 04:00:19 crc kubenswrapper[4802]: I1206 04:00:19.462064 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b364aef0-231f-472a-a8c9-51a5199c0997" path="/var/lib/kubelet/pods/b364aef0-231f-472a-a8c9-51a5199c0997/volumes" Dec 06 04:00:19 crc kubenswrapper[4802]: I1206 04:00:19.580257 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-788b9d8f4c-rk4vn" Dec 06 04:00:19 crc kubenswrapper[4802]: I1206 04:00:19.593587 4802 generic.go:334] "Generic (PLEG): container finished" podID="11b0ceda-c31c-4bc3-ad63-849047b7f80c" containerID="72204b39cf3ad8d51d488287d976a47130d179e020d16c458d0c471c1ec12b87" exitCode=0 Dec 06 04:00:19 crc kubenswrapper[4802]: I1206 04:00:19.593626 4802 generic.go:334] "Generic (PLEG): container finished" podID="11b0ceda-c31c-4bc3-ad63-849047b7f80c" containerID="3c540921637b4fd0bcd4f4eb90cc1ec51a09d48dbcc5e4205f5bca162358eae7" exitCode=2 Dec 06 04:00:19 crc kubenswrapper[4802]: I1206 04:00:19.593635 4802 generic.go:334] "Generic (PLEG): container finished" podID="11b0ceda-c31c-4bc3-ad63-849047b7f80c" containerID="c55ccc53f4da5768babf5a455a6f4c6821e3ce0f9dd3a5706c5e2a72b5f38e9e" exitCode=0 Dec 06 04:00:19 crc kubenswrapper[4802]: I1206 04:00:19.593683 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"11b0ceda-c31c-4bc3-ad63-849047b7f80c","Type":"ContainerDied","Data":"72204b39cf3ad8d51d488287d976a47130d179e020d16c458d0c471c1ec12b87"} Dec 06 04:00:19 crc kubenswrapper[4802]: I1206 04:00:19.593740 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"11b0ceda-c31c-4bc3-ad63-849047b7f80c","Type":"ContainerDied","Data":"3c540921637b4fd0bcd4f4eb90cc1ec51a09d48dbcc5e4205f5bca162358eae7"} Dec 06 04:00:19 crc kubenswrapper[4802]: I1206 04:00:19.593762 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"11b0ceda-c31c-4bc3-ad63-849047b7f80c","Type":"ContainerDied","Data":"c55ccc53f4da5768babf5a455a6f4c6821e3ce0f9dd3a5706c5e2a72b5f38e9e"} Dec 06 04:00:19 crc kubenswrapper[4802]: I1206 04:00:19.595019 4802 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 04:00:19 crc kubenswrapper[4802]: I1206 04:00:19.595033 4802 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 04:00:19 crc kubenswrapper[4802]: I1206 04:00:19.636779 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-7c696dc874-7894k"] Dec 06 04:00:20 crc kubenswrapper[4802]: I1206 04:00:20.032192 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-6ddd7d8889-js75h" Dec 06 04:00:20 crc kubenswrapper[4802]: I1206 04:00:20.082407 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-798947df95-gjr2t"] Dec 06 04:00:20 crc kubenswrapper[4802]: I1206 04:00:20.157594 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7c696dc874-7894k" Dec 06 04:00:20 crc kubenswrapper[4802]: I1206 04:00:20.198247 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 06 04:00:20 crc kubenswrapper[4802]: I1206 04:00:20.260184 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/246893e5-50e3-402e-8b4f-f21b31af732e-combined-ca-bundle\") pod \"246893e5-50e3-402e-8b4f-f21b31af732e\" (UID: \"246893e5-50e3-402e-8b4f-f21b31af732e\") " Dec 06 04:00:20 crc kubenswrapper[4802]: I1206 04:00:20.260340 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s5bzt\" (UniqueName: \"kubernetes.io/projected/246893e5-50e3-402e-8b4f-f21b31af732e-kube-api-access-s5bzt\") pod \"246893e5-50e3-402e-8b4f-f21b31af732e\" (UID: \"246893e5-50e3-402e-8b4f-f21b31af732e\") " Dec 06 04:00:20 crc kubenswrapper[4802]: I1206 04:00:20.260562 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/246893e5-50e3-402e-8b4f-f21b31af732e-config-data\") pod \"246893e5-50e3-402e-8b4f-f21b31af732e\" (UID: \"246893e5-50e3-402e-8b4f-f21b31af732e\") " Dec 06 04:00:20 crc kubenswrapper[4802]: I1206 04:00:20.260605 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/246893e5-50e3-402e-8b4f-f21b31af732e-config-data-custom\") pod \"246893e5-50e3-402e-8b4f-f21b31af732e\" (UID: \"246893e5-50e3-402e-8b4f-f21b31af732e\") " Dec 06 04:00:20 crc kubenswrapper[4802]: I1206 04:00:20.271801 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/246893e5-50e3-402e-8b4f-f21b31af732e-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "246893e5-50e3-402e-8b4f-f21b31af732e" (UID: "246893e5-50e3-402e-8b4f-f21b31af732e"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:00:20 crc kubenswrapper[4802]: I1206 04:00:20.272461 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/246893e5-50e3-402e-8b4f-f21b31af732e-kube-api-access-s5bzt" (OuterVolumeSpecName: "kube-api-access-s5bzt") pod "246893e5-50e3-402e-8b4f-f21b31af732e" (UID: "246893e5-50e3-402e-8b4f-f21b31af732e"). InnerVolumeSpecName "kube-api-access-s5bzt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:00:20 crc kubenswrapper[4802]: I1206 04:00:20.313107 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/246893e5-50e3-402e-8b4f-f21b31af732e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "246893e5-50e3-402e-8b4f-f21b31af732e" (UID: "246893e5-50e3-402e-8b4f-f21b31af732e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:00:20 crc kubenswrapper[4802]: I1206 04:00:20.338076 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/246893e5-50e3-402e-8b4f-f21b31af732e-config-data" (OuterVolumeSpecName: "config-data") pod "246893e5-50e3-402e-8b4f-f21b31af732e" (UID: "246893e5-50e3-402e-8b4f-f21b31af732e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:00:20 crc kubenswrapper[4802]: I1206 04:00:20.363656 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/246893e5-50e3-402e-8b4f-f21b31af732e-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:20 crc kubenswrapper[4802]: I1206 04:00:20.363683 4802 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/246893e5-50e3-402e-8b4f-f21b31af732e-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:20 crc kubenswrapper[4802]: I1206 04:00:20.363696 4802 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/246893e5-50e3-402e-8b4f-f21b31af732e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:20 crc kubenswrapper[4802]: I1206 04:00:20.363705 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s5bzt\" (UniqueName: \"kubernetes.io/projected/246893e5-50e3-402e-8b4f-f21b31af732e-kube-api-access-s5bzt\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:20 crc kubenswrapper[4802]: I1206 04:00:20.379324 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 06 04:00:20 crc kubenswrapper[4802]: I1206 04:00:20.468376 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-798947df95-gjr2t" Dec 06 04:00:20 crc kubenswrapper[4802]: I1206 04:00:20.566594 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6jkfv\" (UniqueName: \"kubernetes.io/projected/eda29ee7-1690-43ed-be0f-c8aac34d6daa-kube-api-access-6jkfv\") pod \"eda29ee7-1690-43ed-be0f-c8aac34d6daa\" (UID: \"eda29ee7-1690-43ed-be0f-c8aac34d6daa\") " Dec 06 04:00:20 crc kubenswrapper[4802]: I1206 04:00:20.566633 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eda29ee7-1690-43ed-be0f-c8aac34d6daa-config-data-custom\") pod \"eda29ee7-1690-43ed-be0f-c8aac34d6daa\" (UID: \"eda29ee7-1690-43ed-be0f-c8aac34d6daa\") " Dec 06 04:00:20 crc kubenswrapper[4802]: I1206 04:00:20.566761 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eda29ee7-1690-43ed-be0f-c8aac34d6daa-config-data\") pod \"eda29ee7-1690-43ed-be0f-c8aac34d6daa\" (UID: \"eda29ee7-1690-43ed-be0f-c8aac34d6daa\") " Dec 06 04:00:20 crc kubenswrapper[4802]: I1206 04:00:20.566789 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eda29ee7-1690-43ed-be0f-c8aac34d6daa-combined-ca-bundle\") pod \"eda29ee7-1690-43ed-be0f-c8aac34d6daa\" (UID: \"eda29ee7-1690-43ed-be0f-c8aac34d6daa\") " Dec 06 04:00:20 crc kubenswrapper[4802]: I1206 04:00:20.570314 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eda29ee7-1690-43ed-be0f-c8aac34d6daa-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "eda29ee7-1690-43ed-be0f-c8aac34d6daa" (UID: "eda29ee7-1690-43ed-be0f-c8aac34d6daa"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:00:20 crc kubenswrapper[4802]: I1206 04:00:20.572149 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eda29ee7-1690-43ed-be0f-c8aac34d6daa-kube-api-access-6jkfv" (OuterVolumeSpecName: "kube-api-access-6jkfv") pod "eda29ee7-1690-43ed-be0f-c8aac34d6daa" (UID: "eda29ee7-1690-43ed-be0f-c8aac34d6daa"). InnerVolumeSpecName "kube-api-access-6jkfv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:00:20 crc kubenswrapper[4802]: I1206 04:00:20.599826 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eda29ee7-1690-43ed-be0f-c8aac34d6daa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eda29ee7-1690-43ed-be0f-c8aac34d6daa" (UID: "eda29ee7-1690-43ed-be0f-c8aac34d6daa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:00:20 crc kubenswrapper[4802]: I1206 04:00:20.605000 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7c696dc874-7894k" event={"ID":"246893e5-50e3-402e-8b4f-f21b31af732e","Type":"ContainerDied","Data":"c2e97f7a5af1583a24465c8f20a8e5e8781bf5ba31141dc4070617a176c41652"} Dec 06 04:00:20 crc kubenswrapper[4802]: I1206 04:00:20.605360 4802 scope.go:117] "RemoveContainer" containerID="bead59f8707df720f20d9b3824500ec2f9ea32ae0f3fbde1604209abd9f3a330" Dec 06 04:00:20 crc kubenswrapper[4802]: I1206 04:00:20.605042 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7c696dc874-7894k" Dec 06 04:00:20 crc kubenswrapper[4802]: I1206 04:00:20.612316 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-798947df95-gjr2t" Dec 06 04:00:20 crc kubenswrapper[4802]: I1206 04:00:20.612460 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-798947df95-gjr2t" event={"ID":"eda29ee7-1690-43ed-be0f-c8aac34d6daa","Type":"ContainerDied","Data":"07d0778674f0592d75374d5761ba6925eda84b3b0bde117a99e41e6eea18d700"} Dec 06 04:00:20 crc kubenswrapper[4802]: I1206 04:00:20.629699 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eda29ee7-1690-43ed-be0f-c8aac34d6daa-config-data" (OuterVolumeSpecName: "config-data") pod "eda29ee7-1690-43ed-be0f-c8aac34d6daa" (UID: "eda29ee7-1690-43ed-be0f-c8aac34d6daa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:00:20 crc kubenswrapper[4802]: I1206 04:00:20.641590 4802 scope.go:117] "RemoveContainer" containerID="9d9867310e246ceb74a94766b213bec68224ed5bf248ef2c905af51fa55f0841" Dec 06 04:00:20 crc kubenswrapper[4802]: I1206 04:00:20.647839 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-7c696dc874-7894k"] Dec 06 04:00:20 crc kubenswrapper[4802]: I1206 04:00:20.655153 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-7c696dc874-7894k"] Dec 06 04:00:20 crc kubenswrapper[4802]: I1206 04:00:20.680111 4802 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eda29ee7-1690-43ed-be0f-c8aac34d6daa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:20 crc kubenswrapper[4802]: I1206 04:00:20.680146 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6jkfv\" (UniqueName: \"kubernetes.io/projected/eda29ee7-1690-43ed-be0f-c8aac34d6daa-kube-api-access-6jkfv\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:20 crc kubenswrapper[4802]: I1206 04:00:20.680158 4802 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eda29ee7-1690-43ed-be0f-c8aac34d6daa-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:20 crc kubenswrapper[4802]: I1206 04:00:20.680170 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eda29ee7-1690-43ed-be0f-c8aac34d6daa-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:20 crc kubenswrapper[4802]: I1206 04:00:20.943689 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-798947df95-gjr2t"] Dec 06 04:00:20 crc kubenswrapper[4802]: I1206 04:00:20.952035 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-798947df95-gjr2t"] Dec 06 04:00:21 crc kubenswrapper[4802]: I1206 04:00:21.459961 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="246893e5-50e3-402e-8b4f-f21b31af732e" path="/var/lib/kubelet/pods/246893e5-50e3-402e-8b4f-f21b31af732e/volumes" Dec 06 04:00:21 crc kubenswrapper[4802]: I1206 04:00:21.460558 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eda29ee7-1690-43ed-be0f-c8aac34d6daa" path="/var/lib/kubelet/pods/eda29ee7-1690-43ed-be0f-c8aac34d6daa/volumes" Dec 06 04:00:24 crc kubenswrapper[4802]: I1206 04:00:24.232080 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-55fc64bfdd-779rm" Dec 06 04:00:24 crc kubenswrapper[4802]: I1206 04:00:24.280072 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-5d96658447-rg8zs"] Dec 06 04:00:24 crc kubenswrapper[4802]: I1206 04:00:24.280287 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-engine-5d96658447-rg8zs" podUID="d3da4180-e8fe-4785-8d4f-6cd172fb6399" containerName="heat-engine" containerID="cri-o://15461749ec2cccc9a0f2a06a22252592288f81b9b432dde2bc6f7f9917e9bb8a" gracePeriod=60 Dec 06 04:00:24 crc kubenswrapper[4802]: I1206 04:00:24.651974 4802 generic.go:334] "Generic (PLEG): container finished" podID="11b0ceda-c31c-4bc3-ad63-849047b7f80c" containerID="64ed0713fa1f49302593eb067506d0753fe92e92791f4939fbc92371d8519cf1" exitCode=0 Dec 06 04:00:24 crc kubenswrapper[4802]: I1206 04:00:24.652335 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"11b0ceda-c31c-4bc3-ad63-849047b7f80c","Type":"ContainerDied","Data":"64ed0713fa1f49302593eb067506d0753fe92e92791f4939fbc92371d8519cf1"} Dec 06 04:00:24 crc kubenswrapper[4802]: I1206 04:00:24.918399 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 04:00:24 crc kubenswrapper[4802]: I1206 04:00:24.956326 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11b0ceda-c31c-4bc3-ad63-849047b7f80c-combined-ca-bundle\") pod \"11b0ceda-c31c-4bc3-ad63-849047b7f80c\" (UID: \"11b0ceda-c31c-4bc3-ad63-849047b7f80c\") " Dec 06 04:00:24 crc kubenswrapper[4802]: I1206 04:00:24.956648 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/11b0ceda-c31c-4bc3-ad63-849047b7f80c-log-httpd\") pod \"11b0ceda-c31c-4bc3-ad63-849047b7f80c\" (UID: \"11b0ceda-c31c-4bc3-ad63-849047b7f80c\") " Dec 06 04:00:24 crc kubenswrapper[4802]: I1206 04:00:24.957130 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11b0ceda-c31c-4bc3-ad63-849047b7f80c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "11b0ceda-c31c-4bc3-ad63-849047b7f80c" (UID: "11b0ceda-c31c-4bc3-ad63-849047b7f80c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:00:24 crc kubenswrapper[4802]: I1206 04:00:24.957364 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11b0ceda-c31c-4bc3-ad63-849047b7f80c-scripts\") pod \"11b0ceda-c31c-4bc3-ad63-849047b7f80c\" (UID: \"11b0ceda-c31c-4bc3-ad63-849047b7f80c\") " Dec 06 04:00:24 crc kubenswrapper[4802]: I1206 04:00:24.957804 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/11b0ceda-c31c-4bc3-ad63-849047b7f80c-run-httpd\") pod \"11b0ceda-c31c-4bc3-ad63-849047b7f80c\" (UID: \"11b0ceda-c31c-4bc3-ad63-849047b7f80c\") " Dec 06 04:00:24 crc kubenswrapper[4802]: I1206 04:00:24.957999 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4vwk5\" (UniqueName: \"kubernetes.io/projected/11b0ceda-c31c-4bc3-ad63-849047b7f80c-kube-api-access-4vwk5\") pod \"11b0ceda-c31c-4bc3-ad63-849047b7f80c\" (UID: \"11b0ceda-c31c-4bc3-ad63-849047b7f80c\") " Dec 06 04:00:24 crc kubenswrapper[4802]: I1206 04:00:24.958095 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11b0ceda-c31c-4bc3-ad63-849047b7f80c-config-data\") pod \"11b0ceda-c31c-4bc3-ad63-849047b7f80c\" (UID: \"11b0ceda-c31c-4bc3-ad63-849047b7f80c\") " Dec 06 04:00:24 crc kubenswrapper[4802]: I1206 04:00:24.958179 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/11b0ceda-c31c-4bc3-ad63-849047b7f80c-sg-core-conf-yaml\") pod \"11b0ceda-c31c-4bc3-ad63-849047b7f80c\" (UID: \"11b0ceda-c31c-4bc3-ad63-849047b7f80c\") " Dec 06 04:00:24 crc kubenswrapper[4802]: I1206 04:00:24.959280 4802 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/11b0ceda-c31c-4bc3-ad63-849047b7f80c-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:24 crc kubenswrapper[4802]: I1206 04:00:24.959934 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11b0ceda-c31c-4bc3-ad63-849047b7f80c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "11b0ceda-c31c-4bc3-ad63-849047b7f80c" (UID: "11b0ceda-c31c-4bc3-ad63-849047b7f80c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:00:24 crc kubenswrapper[4802]: I1206 04:00:24.962865 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11b0ceda-c31c-4bc3-ad63-849047b7f80c-scripts" (OuterVolumeSpecName: "scripts") pod "11b0ceda-c31c-4bc3-ad63-849047b7f80c" (UID: "11b0ceda-c31c-4bc3-ad63-849047b7f80c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:00:24 crc kubenswrapper[4802]: I1206 04:00:24.967898 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11b0ceda-c31c-4bc3-ad63-849047b7f80c-kube-api-access-4vwk5" (OuterVolumeSpecName: "kube-api-access-4vwk5") pod "11b0ceda-c31c-4bc3-ad63-849047b7f80c" (UID: "11b0ceda-c31c-4bc3-ad63-849047b7f80c"). InnerVolumeSpecName "kube-api-access-4vwk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.000179 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11b0ceda-c31c-4bc3-ad63-849047b7f80c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "11b0ceda-c31c-4bc3-ad63-849047b7f80c" (UID: "11b0ceda-c31c-4bc3-ad63-849047b7f80c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.048793 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11b0ceda-c31c-4bc3-ad63-849047b7f80c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "11b0ceda-c31c-4bc3-ad63-849047b7f80c" (UID: "11b0ceda-c31c-4bc3-ad63-849047b7f80c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.066139 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4vwk5\" (UniqueName: \"kubernetes.io/projected/11b0ceda-c31c-4bc3-ad63-849047b7f80c-kube-api-access-4vwk5\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.066189 4802 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/11b0ceda-c31c-4bc3-ad63-849047b7f80c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.066202 4802 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11b0ceda-c31c-4bc3-ad63-849047b7f80c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.066212 4802 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11b0ceda-c31c-4bc3-ad63-849047b7f80c-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.066222 4802 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/11b0ceda-c31c-4bc3-ad63-849047b7f80c-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.101175 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11b0ceda-c31c-4bc3-ad63-849047b7f80c-config-data" (OuterVolumeSpecName: "config-data") pod "11b0ceda-c31c-4bc3-ad63-849047b7f80c" (UID: "11b0ceda-c31c-4bc3-ad63-849047b7f80c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.167472 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11b0ceda-c31c-4bc3-ad63-849047b7f80c-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.667663 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"11b0ceda-c31c-4bc3-ad63-849047b7f80c","Type":"ContainerDied","Data":"090052942c5bf6ac980af4ff0c0af25216c513749a0bf82817c01d320f4e67fd"} Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.667980 4802 scope.go:117] "RemoveContainer" containerID="72204b39cf3ad8d51d488287d976a47130d179e020d16c458d0c471c1ec12b87" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.667947 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.688737 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.700166 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.701193 4802 scope.go:117] "RemoveContainer" containerID="3c540921637b4fd0bcd4f4eb90cc1ec51a09d48dbcc5e4205f5bca162358eae7" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.714340 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 06 04:00:25 crc kubenswrapper[4802]: E1206 04:00:25.714918 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16c7bcbb-1a4f-4c68-ade9-3cafed6d6876" containerName="heat-api" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.714990 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="16c7bcbb-1a4f-4c68-ade9-3cafed6d6876" containerName="heat-api" Dec 06 04:00:25 crc kubenswrapper[4802]: E1206 04:00:25.715063 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a5637e5-3258-4eab-9438-479d0aac8099" containerName="init" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.715114 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a5637e5-3258-4eab-9438-479d0aac8099" containerName="init" Dec 06 04:00:25 crc kubenswrapper[4802]: E1206 04:00:25.715170 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eda29ee7-1690-43ed-be0f-c8aac34d6daa" containerName="heat-api" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.715219 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="eda29ee7-1690-43ed-be0f-c8aac34d6daa" containerName="heat-api" Dec 06 04:00:25 crc kubenswrapper[4802]: E1206 04:00:25.715270 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11b0ceda-c31c-4bc3-ad63-849047b7f80c" containerName="ceilometer-notification-agent" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.715330 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="11b0ceda-c31c-4bc3-ad63-849047b7f80c" containerName="ceilometer-notification-agent" Dec 06 04:00:25 crc kubenswrapper[4802]: E1206 04:00:25.715393 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="246893e5-50e3-402e-8b4f-f21b31af732e" containerName="heat-cfnapi" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.715444 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="246893e5-50e3-402e-8b4f-f21b31af732e" containerName="heat-cfnapi" Dec 06 04:00:25 crc kubenswrapper[4802]: E1206 04:00:25.715535 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11b0ceda-c31c-4bc3-ad63-849047b7f80c" containerName="sg-core" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.715592 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="11b0ceda-c31c-4bc3-ad63-849047b7f80c" containerName="sg-core" Dec 06 04:00:25 crc kubenswrapper[4802]: E1206 04:00:25.715645 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11b0ceda-c31c-4bc3-ad63-849047b7f80c" containerName="ceilometer-central-agent" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.715699 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="11b0ceda-c31c-4bc3-ad63-849047b7f80c" containerName="ceilometer-central-agent" Dec 06 04:00:25 crc kubenswrapper[4802]: E1206 04:00:25.715816 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11b0ceda-c31c-4bc3-ad63-849047b7f80c" containerName="proxy-httpd" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.715889 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="11b0ceda-c31c-4bc3-ad63-849047b7f80c" containerName="proxy-httpd" Dec 06 04:00:25 crc kubenswrapper[4802]: E1206 04:00:25.715960 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b364aef0-231f-472a-a8c9-51a5199c0997" containerName="neutron-httpd" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.716017 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="b364aef0-231f-472a-a8c9-51a5199c0997" containerName="neutron-httpd" Dec 06 04:00:25 crc kubenswrapper[4802]: E1206 04:00:25.716075 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b364aef0-231f-472a-a8c9-51a5199c0997" containerName="neutron-api" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.716129 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="b364aef0-231f-472a-a8c9-51a5199c0997" containerName="neutron-api" Dec 06 04:00:25 crc kubenswrapper[4802]: E1206 04:00:25.716184 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a355c0a5-b839-4928-b7ce-0b2d0389190d" containerName="heat-cfnapi" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.716237 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="a355c0a5-b839-4928-b7ce-0b2d0389190d" containerName="heat-cfnapi" Dec 06 04:00:25 crc kubenswrapper[4802]: E1206 04:00:25.716299 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a5637e5-3258-4eab-9438-479d0aac8099" containerName="dnsmasq-dns" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.716365 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a5637e5-3258-4eab-9438-479d0aac8099" containerName="dnsmasq-dns" Dec 06 04:00:25 crc kubenswrapper[4802]: E1206 04:00:25.716432 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5898b86d-3906-4c87-8509-8c37fe50f544" containerName="collect-profiles" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.716508 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="5898b86d-3906-4c87-8509-8c37fe50f544" containerName="collect-profiles" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.716808 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="11b0ceda-c31c-4bc3-ad63-849047b7f80c" containerName="sg-core" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.716983 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="b364aef0-231f-472a-a8c9-51a5199c0997" containerName="neutron-api" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.717061 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="b364aef0-231f-472a-a8c9-51a5199c0997" containerName="neutron-httpd" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.717132 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="a355c0a5-b839-4928-b7ce-0b2d0389190d" containerName="heat-cfnapi" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.717186 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="246893e5-50e3-402e-8b4f-f21b31af732e" containerName="heat-cfnapi" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.717249 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="eda29ee7-1690-43ed-be0f-c8aac34d6daa" containerName="heat-api" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.717303 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="5898b86d-3906-4c87-8509-8c37fe50f544" containerName="collect-profiles" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.717361 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="11b0ceda-c31c-4bc3-ad63-849047b7f80c" containerName="proxy-httpd" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.717420 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="11b0ceda-c31c-4bc3-ad63-849047b7f80c" containerName="ceilometer-central-agent" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.717472 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="246893e5-50e3-402e-8b4f-f21b31af732e" containerName="heat-cfnapi" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.717531 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a5637e5-3258-4eab-9438-479d0aac8099" containerName="dnsmasq-dns" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.717588 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="16c7bcbb-1a4f-4c68-ade9-3cafed6d6876" containerName="heat-api" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.717649 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="11b0ceda-c31c-4bc3-ad63-849047b7f80c" containerName="ceilometer-notification-agent" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.717769 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="eda29ee7-1690-43ed-be0f-c8aac34d6daa" containerName="heat-api" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.717934 4802 scope.go:117] "RemoveContainer" containerID="c55ccc53f4da5768babf5a455a6f4c6821e3ce0f9dd3a5706c5e2a72b5f38e9e" Dec 06 04:00:25 crc kubenswrapper[4802]: E1206 04:00:25.718147 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="246893e5-50e3-402e-8b4f-f21b31af732e" containerName="heat-cfnapi" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.718216 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="246893e5-50e3-402e-8b4f-f21b31af732e" containerName="heat-cfnapi" Dec 06 04:00:25 crc kubenswrapper[4802]: E1206 04:00:25.718292 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eda29ee7-1690-43ed-be0f-c8aac34d6daa" containerName="heat-api" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.718393 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="eda29ee7-1690-43ed-be0f-c8aac34d6daa" containerName="heat-api" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.720077 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.725633 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.725717 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.726017 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.740353 4802 scope.go:117] "RemoveContainer" containerID="64ed0713fa1f49302593eb067506d0753fe92e92791f4939fbc92371d8519cf1" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.785999 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e13eaebb-eae6-4924-8940-9c899466b9ad-log-httpd\") pod \"ceilometer-0\" (UID: \"e13eaebb-eae6-4924-8940-9c899466b9ad\") " pod="openstack/ceilometer-0" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.786218 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e13eaebb-eae6-4924-8940-9c899466b9ad-config-data\") pod \"ceilometer-0\" (UID: \"e13eaebb-eae6-4924-8940-9c899466b9ad\") " pod="openstack/ceilometer-0" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.786300 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e13eaebb-eae6-4924-8940-9c899466b9ad-run-httpd\") pod \"ceilometer-0\" (UID: \"e13eaebb-eae6-4924-8940-9c899466b9ad\") " pod="openstack/ceilometer-0" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.786403 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e13eaebb-eae6-4924-8940-9c899466b9ad-scripts\") pod \"ceilometer-0\" (UID: \"e13eaebb-eae6-4924-8940-9c899466b9ad\") " pod="openstack/ceilometer-0" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.786469 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e13eaebb-eae6-4924-8940-9c899466b9ad-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e13eaebb-eae6-4924-8940-9c899466b9ad\") " pod="openstack/ceilometer-0" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.786577 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9nzw\" (UniqueName: \"kubernetes.io/projected/e13eaebb-eae6-4924-8940-9c899466b9ad-kube-api-access-q9nzw\") pod \"ceilometer-0\" (UID: \"e13eaebb-eae6-4924-8940-9c899466b9ad\") " pod="openstack/ceilometer-0" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.786686 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e13eaebb-eae6-4924-8940-9c899466b9ad-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e13eaebb-eae6-4924-8940-9c899466b9ad\") " pod="openstack/ceilometer-0" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.888469 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e13eaebb-eae6-4924-8940-9c899466b9ad-run-httpd\") pod \"ceilometer-0\" (UID: \"e13eaebb-eae6-4924-8940-9c899466b9ad\") " pod="openstack/ceilometer-0" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.888558 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e13eaebb-eae6-4924-8940-9c899466b9ad-scripts\") pod \"ceilometer-0\" (UID: \"e13eaebb-eae6-4924-8940-9c899466b9ad\") " pod="openstack/ceilometer-0" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.888579 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e13eaebb-eae6-4924-8940-9c899466b9ad-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e13eaebb-eae6-4924-8940-9c899466b9ad\") " pod="openstack/ceilometer-0" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.888665 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9nzw\" (UniqueName: \"kubernetes.io/projected/e13eaebb-eae6-4924-8940-9c899466b9ad-kube-api-access-q9nzw\") pod \"ceilometer-0\" (UID: \"e13eaebb-eae6-4924-8940-9c899466b9ad\") " pod="openstack/ceilometer-0" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.888717 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e13eaebb-eae6-4924-8940-9c899466b9ad-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e13eaebb-eae6-4924-8940-9c899466b9ad\") " pod="openstack/ceilometer-0" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.888742 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e13eaebb-eae6-4924-8940-9c899466b9ad-log-httpd\") pod \"ceilometer-0\" (UID: \"e13eaebb-eae6-4924-8940-9c899466b9ad\") " pod="openstack/ceilometer-0" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.888851 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e13eaebb-eae6-4924-8940-9c899466b9ad-config-data\") pod \"ceilometer-0\" (UID: \"e13eaebb-eae6-4924-8940-9c899466b9ad\") " pod="openstack/ceilometer-0" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.889446 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e13eaebb-eae6-4924-8940-9c899466b9ad-run-httpd\") pod \"ceilometer-0\" (UID: \"e13eaebb-eae6-4924-8940-9c899466b9ad\") " pod="openstack/ceilometer-0" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.889683 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e13eaebb-eae6-4924-8940-9c899466b9ad-log-httpd\") pod \"ceilometer-0\" (UID: \"e13eaebb-eae6-4924-8940-9c899466b9ad\") " pod="openstack/ceilometer-0" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.893546 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e13eaebb-eae6-4924-8940-9c899466b9ad-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e13eaebb-eae6-4924-8940-9c899466b9ad\") " pod="openstack/ceilometer-0" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.894330 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e13eaebb-eae6-4924-8940-9c899466b9ad-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e13eaebb-eae6-4924-8940-9c899466b9ad\") " pod="openstack/ceilometer-0" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.895273 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e13eaebb-eae6-4924-8940-9c899466b9ad-scripts\") pod \"ceilometer-0\" (UID: \"e13eaebb-eae6-4924-8940-9c899466b9ad\") " pod="openstack/ceilometer-0" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.895465 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e13eaebb-eae6-4924-8940-9c899466b9ad-config-data\") pod \"ceilometer-0\" (UID: \"e13eaebb-eae6-4924-8940-9c899466b9ad\") " pod="openstack/ceilometer-0" Dec 06 04:00:25 crc kubenswrapper[4802]: I1206 04:00:25.907785 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9nzw\" (UniqueName: \"kubernetes.io/projected/e13eaebb-eae6-4924-8940-9c899466b9ad-kube-api-access-q9nzw\") pod \"ceilometer-0\" (UID: \"e13eaebb-eae6-4924-8940-9c899466b9ad\") " pod="openstack/ceilometer-0" Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.040090 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.184362 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-gv9nf"] Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.188593 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-gv9nf" Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.199589 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-gv9nf"] Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.261900 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-rhcrb"] Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.263514 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-rhcrb" Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.293779 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-rhcrb"] Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.302698 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b2e9e18c-0730-4b74-8eaf-941cbeb08246-operator-scripts\") pod \"nova-cell0-db-create-rhcrb\" (UID: \"b2e9e18c-0730-4b74-8eaf-941cbeb08246\") " pod="openstack/nova-cell0-db-create-rhcrb" Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.302735 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2f247bdd-ddf7-4872-a1c5-597fec152cde-operator-scripts\") pod \"nova-api-db-create-gv9nf\" (UID: \"2f247bdd-ddf7-4872-a1c5-597fec152cde\") " pod="openstack/nova-api-db-create-gv9nf" Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.302839 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87l9f\" (UniqueName: \"kubernetes.io/projected/2f247bdd-ddf7-4872-a1c5-597fec152cde-kube-api-access-87l9f\") pod \"nova-api-db-create-gv9nf\" (UID: \"2f247bdd-ddf7-4872-a1c5-597fec152cde\") " pod="openstack/nova-api-db-create-gv9nf" Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.302874 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4gsh\" (UniqueName: \"kubernetes.io/projected/b2e9e18c-0730-4b74-8eaf-941cbeb08246-kube-api-access-q4gsh\") pod \"nova-cell0-db-create-rhcrb\" (UID: \"b2e9e18c-0730-4b74-8eaf-941cbeb08246\") " pod="openstack/nova-cell0-db-create-rhcrb" Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.305240 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-1043-account-create-update-xzd9x"] Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.306081 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-1043-account-create-update-xzd9x" Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.307946 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.330121 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-1043-account-create-update-xzd9x"] Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.383489 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-b8sxp"] Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.384849 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-b8sxp" Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.403930 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87l9f\" (UniqueName: \"kubernetes.io/projected/2f247bdd-ddf7-4872-a1c5-597fec152cde-kube-api-access-87l9f\") pod \"nova-api-db-create-gv9nf\" (UID: \"2f247bdd-ddf7-4872-a1c5-597fec152cde\") " pod="openstack/nova-api-db-create-gv9nf" Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.404003 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4gsh\" (UniqueName: \"kubernetes.io/projected/b2e9e18c-0730-4b74-8eaf-941cbeb08246-kube-api-access-q4gsh\") pod \"nova-cell0-db-create-rhcrb\" (UID: \"b2e9e18c-0730-4b74-8eaf-941cbeb08246\") " pod="openstack/nova-cell0-db-create-rhcrb" Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.404055 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/59d551a5-c1ef-46ee-8ae2-f98526b2ab09-operator-scripts\") pod \"nova-api-1043-account-create-update-xzd9x\" (UID: \"59d551a5-c1ef-46ee-8ae2-f98526b2ab09\") " pod="openstack/nova-api-1043-account-create-update-xzd9x" Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.404129 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b2e9e18c-0730-4b74-8eaf-941cbeb08246-operator-scripts\") pod \"nova-cell0-db-create-rhcrb\" (UID: \"b2e9e18c-0730-4b74-8eaf-941cbeb08246\") " pod="openstack/nova-cell0-db-create-rhcrb" Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.404159 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2f247bdd-ddf7-4872-a1c5-597fec152cde-operator-scripts\") pod \"nova-api-db-create-gv9nf\" (UID: \"2f247bdd-ddf7-4872-a1c5-597fec152cde\") " pod="openstack/nova-api-db-create-gv9nf" Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.404200 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phn2m\" (UniqueName: \"kubernetes.io/projected/59d551a5-c1ef-46ee-8ae2-f98526b2ab09-kube-api-access-phn2m\") pod \"nova-api-1043-account-create-update-xzd9x\" (UID: \"59d551a5-c1ef-46ee-8ae2-f98526b2ab09\") " pod="openstack/nova-api-1043-account-create-update-xzd9x" Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.405283 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b2e9e18c-0730-4b74-8eaf-941cbeb08246-operator-scripts\") pod \"nova-cell0-db-create-rhcrb\" (UID: \"b2e9e18c-0730-4b74-8eaf-941cbeb08246\") " pod="openstack/nova-cell0-db-create-rhcrb" Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.405378 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2f247bdd-ddf7-4872-a1c5-597fec152cde-operator-scripts\") pod \"nova-api-db-create-gv9nf\" (UID: \"2f247bdd-ddf7-4872-a1c5-597fec152cde\") " pod="openstack/nova-api-db-create-gv9nf" Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.410784 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-b8sxp"] Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.427524 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4gsh\" (UniqueName: \"kubernetes.io/projected/b2e9e18c-0730-4b74-8eaf-941cbeb08246-kube-api-access-q4gsh\") pod \"nova-cell0-db-create-rhcrb\" (UID: \"b2e9e18c-0730-4b74-8eaf-941cbeb08246\") " pod="openstack/nova-cell0-db-create-rhcrb" Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.428177 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87l9f\" (UniqueName: \"kubernetes.io/projected/2f247bdd-ddf7-4872-a1c5-597fec152cde-kube-api-access-87l9f\") pod \"nova-api-db-create-gv9nf\" (UID: \"2f247bdd-ddf7-4872-a1c5-597fec152cde\") " pod="openstack/nova-api-db-create-gv9nf" Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.472908 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cdd3-account-create-update-bjk6c"] Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.474029 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cdd3-account-create-update-bjk6c" Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.478962 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.488997 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cdd3-account-create-update-bjk6c"] Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.506769 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phn2m\" (UniqueName: \"kubernetes.io/projected/59d551a5-c1ef-46ee-8ae2-f98526b2ab09-kube-api-access-phn2m\") pod \"nova-api-1043-account-create-update-xzd9x\" (UID: \"59d551a5-c1ef-46ee-8ae2-f98526b2ab09\") " pod="openstack/nova-api-1043-account-create-update-xzd9x" Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.506847 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czqc9\" (UniqueName: \"kubernetes.io/projected/04a970e8-495a-491a-9f83-1b25159450b1-kube-api-access-czqc9\") pod \"nova-cell1-db-create-b8sxp\" (UID: \"04a970e8-495a-491a-9f83-1b25159450b1\") " pod="openstack/nova-cell1-db-create-b8sxp" Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.506873 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/04a970e8-495a-491a-9f83-1b25159450b1-operator-scripts\") pod \"nova-cell1-db-create-b8sxp\" (UID: \"04a970e8-495a-491a-9f83-1b25159450b1\") " pod="openstack/nova-cell1-db-create-b8sxp" Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.506934 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/59d551a5-c1ef-46ee-8ae2-f98526b2ab09-operator-scripts\") pod \"nova-api-1043-account-create-update-xzd9x\" (UID: \"59d551a5-c1ef-46ee-8ae2-f98526b2ab09\") " pod="openstack/nova-api-1043-account-create-update-xzd9x" Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.507655 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/59d551a5-c1ef-46ee-8ae2-f98526b2ab09-operator-scripts\") pod \"nova-api-1043-account-create-update-xzd9x\" (UID: \"59d551a5-c1ef-46ee-8ae2-f98526b2ab09\") " pod="openstack/nova-api-1043-account-create-update-xzd9x" Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.523314 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-gv9nf" Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.537517 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phn2m\" (UniqueName: \"kubernetes.io/projected/59d551a5-c1ef-46ee-8ae2-f98526b2ab09-kube-api-access-phn2m\") pod \"nova-api-1043-account-create-update-xzd9x\" (UID: \"59d551a5-c1ef-46ee-8ae2-f98526b2ab09\") " pod="openstack/nova-api-1043-account-create-update-xzd9x" Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.598233 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-rhcrb" Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.608321 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czqc9\" (UniqueName: \"kubernetes.io/projected/04a970e8-495a-491a-9f83-1b25159450b1-kube-api-access-czqc9\") pod \"nova-cell1-db-create-b8sxp\" (UID: \"04a970e8-495a-491a-9f83-1b25159450b1\") " pod="openstack/nova-cell1-db-create-b8sxp" Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.608367 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/04a970e8-495a-491a-9f83-1b25159450b1-operator-scripts\") pod \"nova-cell1-db-create-b8sxp\" (UID: \"04a970e8-495a-491a-9f83-1b25159450b1\") " pod="openstack/nova-cell1-db-create-b8sxp" Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.608430 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sk6ph\" (UniqueName: \"kubernetes.io/projected/164b8cc5-aab5-4e71-af01-d39679d4b3ba-kube-api-access-sk6ph\") pod \"nova-cell0-cdd3-account-create-update-bjk6c\" (UID: \"164b8cc5-aab5-4e71-af01-d39679d4b3ba\") " pod="openstack/nova-cell0-cdd3-account-create-update-bjk6c" Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.608455 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/164b8cc5-aab5-4e71-af01-d39679d4b3ba-operator-scripts\") pod \"nova-cell0-cdd3-account-create-update-bjk6c\" (UID: \"164b8cc5-aab5-4e71-af01-d39679d4b3ba\") " pod="openstack/nova-cell0-cdd3-account-create-update-bjk6c" Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.609435 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/04a970e8-495a-491a-9f83-1b25159450b1-operator-scripts\") pod \"nova-cell1-db-create-b8sxp\" (UID: \"04a970e8-495a-491a-9f83-1b25159450b1\") " pod="openstack/nova-cell1-db-create-b8sxp" Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.628153 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-1043-account-create-update-xzd9x" Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.631315 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czqc9\" (UniqueName: \"kubernetes.io/projected/04a970e8-495a-491a-9f83-1b25159450b1-kube-api-access-czqc9\") pod \"nova-cell1-db-create-b8sxp\" (UID: \"04a970e8-495a-491a-9f83-1b25159450b1\") " pod="openstack/nova-cell1-db-create-b8sxp" Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.655427 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.675083 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-66f6-account-create-update-th68q"] Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.677493 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-66f6-account-create-update-th68q" Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.679979 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.690218 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-66f6-account-create-update-th68q"] Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.703163 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-b8sxp" Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.710431 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sk6ph\" (UniqueName: \"kubernetes.io/projected/164b8cc5-aab5-4e71-af01-d39679d4b3ba-kube-api-access-sk6ph\") pod \"nova-cell0-cdd3-account-create-update-bjk6c\" (UID: \"164b8cc5-aab5-4e71-af01-d39679d4b3ba\") " pod="openstack/nova-cell0-cdd3-account-create-update-bjk6c" Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.710468 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/164b8cc5-aab5-4e71-af01-d39679d4b3ba-operator-scripts\") pod \"nova-cell0-cdd3-account-create-update-bjk6c\" (UID: \"164b8cc5-aab5-4e71-af01-d39679d4b3ba\") " pod="openstack/nova-cell0-cdd3-account-create-update-bjk6c" Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.711154 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/164b8cc5-aab5-4e71-af01-d39679d4b3ba-operator-scripts\") pod \"nova-cell0-cdd3-account-create-update-bjk6c\" (UID: \"164b8cc5-aab5-4e71-af01-d39679d4b3ba\") " pod="openstack/nova-cell0-cdd3-account-create-update-bjk6c" Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.745067 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sk6ph\" (UniqueName: \"kubernetes.io/projected/164b8cc5-aab5-4e71-af01-d39679d4b3ba-kube-api-access-sk6ph\") pod \"nova-cell0-cdd3-account-create-update-bjk6c\" (UID: \"164b8cc5-aab5-4e71-af01-d39679d4b3ba\") " pod="openstack/nova-cell0-cdd3-account-create-update-bjk6c" Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.798171 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cdd3-account-create-update-bjk6c" Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.812226 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a5a43d4-c219-4090-acc3-252e127e48e7-operator-scripts\") pod \"nova-cell1-66f6-account-create-update-th68q\" (UID: \"8a5a43d4-c219-4090-acc3-252e127e48e7\") " pod="openstack/nova-cell1-66f6-account-create-update-th68q" Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.812617 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwbqw\" (UniqueName: \"kubernetes.io/projected/8a5a43d4-c219-4090-acc3-252e127e48e7-kube-api-access-kwbqw\") pod \"nova-cell1-66f6-account-create-update-th68q\" (UID: \"8a5a43d4-c219-4090-acc3-252e127e48e7\") " pod="openstack/nova-cell1-66f6-account-create-update-th68q" Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.921434 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwbqw\" (UniqueName: \"kubernetes.io/projected/8a5a43d4-c219-4090-acc3-252e127e48e7-kube-api-access-kwbqw\") pod \"nova-cell1-66f6-account-create-update-th68q\" (UID: \"8a5a43d4-c219-4090-acc3-252e127e48e7\") " pod="openstack/nova-cell1-66f6-account-create-update-th68q" Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.921688 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a5a43d4-c219-4090-acc3-252e127e48e7-operator-scripts\") pod \"nova-cell1-66f6-account-create-update-th68q\" (UID: \"8a5a43d4-c219-4090-acc3-252e127e48e7\") " pod="openstack/nova-cell1-66f6-account-create-update-th68q" Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.922392 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a5a43d4-c219-4090-acc3-252e127e48e7-operator-scripts\") pod \"nova-cell1-66f6-account-create-update-th68q\" (UID: \"8a5a43d4-c219-4090-acc3-252e127e48e7\") " pod="openstack/nova-cell1-66f6-account-create-update-th68q" Dec 06 04:00:26 crc kubenswrapper[4802]: I1206 04:00:26.949262 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwbqw\" (UniqueName: \"kubernetes.io/projected/8a5a43d4-c219-4090-acc3-252e127e48e7-kube-api-access-kwbqw\") pod \"nova-cell1-66f6-account-create-update-th68q\" (UID: \"8a5a43d4-c219-4090-acc3-252e127e48e7\") " pod="openstack/nova-cell1-66f6-account-create-update-th68q" Dec 06 04:00:27 crc kubenswrapper[4802]: I1206 04:00:27.098125 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-gv9nf"] Dec 06 04:00:27 crc kubenswrapper[4802]: I1206 04:00:27.122983 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-66f6-account-create-update-th68q" Dec 06 04:00:27 crc kubenswrapper[4802]: I1206 04:00:27.175604 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-1043-account-create-update-xzd9x"] Dec 06 04:00:27 crc kubenswrapper[4802]: I1206 04:00:27.353782 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-rhcrb"] Dec 06 04:00:27 crc kubenswrapper[4802]: W1206 04:00:27.363631 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod04a970e8_495a_491a_9f83_1b25159450b1.slice/crio-3075cd1dc6fc542a71d7fe4db60ccecfbf362f4a8e63e246a4956368ecfb00e0 WatchSource:0}: Error finding container 3075cd1dc6fc542a71d7fe4db60ccecfbf362f4a8e63e246a4956368ecfb00e0: Status 404 returned error can't find the container with id 3075cd1dc6fc542a71d7fe4db60ccecfbf362f4a8e63e246a4956368ecfb00e0 Dec 06 04:00:27 crc kubenswrapper[4802]: I1206 04:00:27.364546 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-b8sxp"] Dec 06 04:00:27 crc kubenswrapper[4802]: E1206 04:00:27.456363 4802 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="15461749ec2cccc9a0f2a06a22252592288f81b9b432dde2bc6f7f9917e9bb8a" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 06 04:00:27 crc kubenswrapper[4802]: E1206 04:00:27.458955 4802 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="15461749ec2cccc9a0f2a06a22252592288f81b9b432dde2bc6f7f9917e9bb8a" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 06 04:00:27 crc kubenswrapper[4802]: E1206 04:00:27.468120 4802 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="15461749ec2cccc9a0f2a06a22252592288f81b9b432dde2bc6f7f9917e9bb8a" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 06 04:00:27 crc kubenswrapper[4802]: E1206 04:00:27.468195 4802 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-5d96658447-rg8zs" podUID="d3da4180-e8fe-4785-8d4f-6cd172fb6399" containerName="heat-engine" Dec 06 04:00:27 crc kubenswrapper[4802]: I1206 04:00:27.471254 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11b0ceda-c31c-4bc3-ad63-849047b7f80c" path="/var/lib/kubelet/pods/11b0ceda-c31c-4bc3-ad63-849047b7f80c/volumes" Dec 06 04:00:27 crc kubenswrapper[4802]: I1206 04:00:27.517958 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cdd3-account-create-update-bjk6c"] Dec 06 04:00:27 crc kubenswrapper[4802]: I1206 04:00:27.652550 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-66f6-account-create-update-th68q"] Dec 06 04:00:27 crc kubenswrapper[4802]: I1206 04:00:27.752810 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cdd3-account-create-update-bjk6c" event={"ID":"164b8cc5-aab5-4e71-af01-d39679d4b3ba","Type":"ContainerStarted","Data":"ad5035d9c0a9713ee286fc5f357cf0ddf5d25d8808c617ef34049db28ecbec87"} Dec 06 04:00:27 crc kubenswrapper[4802]: I1206 04:00:27.769809 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-b8sxp" event={"ID":"04a970e8-495a-491a-9f83-1b25159450b1","Type":"ContainerStarted","Data":"27a342c130d871ab7897d1d85dbab34c5507f64a62e6aa01189f4c55aa35e1ad"} Dec 06 04:00:27 crc kubenswrapper[4802]: I1206 04:00:27.769866 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-b8sxp" event={"ID":"04a970e8-495a-491a-9f83-1b25159450b1","Type":"ContainerStarted","Data":"3075cd1dc6fc542a71d7fe4db60ccecfbf362f4a8e63e246a4956368ecfb00e0"} Dec 06 04:00:27 crc kubenswrapper[4802]: I1206 04:00:27.779537 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-gv9nf" event={"ID":"2f247bdd-ddf7-4872-a1c5-597fec152cde","Type":"ContainerStarted","Data":"5ab9087cf180f2a18ce6757d996a25ac4c3e22341ed3ae95f9624968cb5de9b2"} Dec 06 04:00:27 crc kubenswrapper[4802]: I1206 04:00:27.779592 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-gv9nf" event={"ID":"2f247bdd-ddf7-4872-a1c5-597fec152cde","Type":"ContainerStarted","Data":"0e7f9793d27a6fc8131ef36388f67af7f156baebed2e20f4e765c08af68fa4d5"} Dec 06 04:00:27 crc kubenswrapper[4802]: I1206 04:00:27.792023 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-rhcrb" event={"ID":"b2e9e18c-0730-4b74-8eaf-941cbeb08246","Type":"ContainerStarted","Data":"e760325fd8c5ac0b334091d72ec787801e5b3752b6a5b79af58b3464668f4397"} Dec 06 04:00:27 crc kubenswrapper[4802]: I1206 04:00:27.794776 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e13eaebb-eae6-4924-8940-9c899466b9ad","Type":"ContainerStarted","Data":"971d5774565cea6f18af2a6d15a3735192ad2bf168ed1e3ad329eee93cbe4282"} Dec 06 04:00:27 crc kubenswrapper[4802]: I1206 04:00:27.797999 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-1043-account-create-update-xzd9x" event={"ID":"59d551a5-c1ef-46ee-8ae2-f98526b2ab09","Type":"ContainerStarted","Data":"a6c5c28aa66df89c8e48ac36cf218e1d67940dcd025ae23684066eb35df96f0d"} Dec 06 04:00:27 crc kubenswrapper[4802]: I1206 04:00:27.798060 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-1043-account-create-update-xzd9x" event={"ID":"59d551a5-c1ef-46ee-8ae2-f98526b2ab09","Type":"ContainerStarted","Data":"1460f3c36934cefede236489181860f0e1245a09ad8a6c10c21284573a3e315c"} Dec 06 04:00:27 crc kubenswrapper[4802]: I1206 04:00:27.836067 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-b8sxp" podStartSLOduration=1.8360499460000002 podStartE2EDuration="1.836049946s" podCreationTimestamp="2025-12-06 04:00:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:00:27.831762129 +0000 UTC m=+1220.703671281" watchObservedRunningTime="2025-12-06 04:00:27.836049946 +0000 UTC m=+1220.707959098" Dec 06 04:00:27 crc kubenswrapper[4802]: I1206 04:00:27.875237 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-1043-account-create-update-xzd9x" podStartSLOduration=1.875211832 podStartE2EDuration="1.875211832s" podCreationTimestamp="2025-12-06 04:00:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:00:27.861775436 +0000 UTC m=+1220.733684598" watchObservedRunningTime="2025-12-06 04:00:27.875211832 +0000 UTC m=+1220.747120984" Dec 06 04:00:27 crc kubenswrapper[4802]: I1206 04:00:27.905633 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-gv9nf" podStartSLOduration=1.90561468 podStartE2EDuration="1.90561468s" podCreationTimestamp="2025-12-06 04:00:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:00:27.902663809 +0000 UTC m=+1220.774572961" watchObservedRunningTime="2025-12-06 04:00:27.90561468 +0000 UTC m=+1220.777523842" Dec 06 04:00:28 crc kubenswrapper[4802]: I1206 04:00:28.818004 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cdd3-account-create-update-bjk6c" event={"ID":"164b8cc5-aab5-4e71-af01-d39679d4b3ba","Type":"ContainerStarted","Data":"71f041b5f79ae84d3ca26838b981a0e01fda8163adf89de78ed9c1bdb15ccff3"} Dec 06 04:00:28 crc kubenswrapper[4802]: I1206 04:00:28.826130 4802 generic.go:334] "Generic (PLEG): container finished" podID="04a970e8-495a-491a-9f83-1b25159450b1" containerID="27a342c130d871ab7897d1d85dbab34c5507f64a62e6aa01189f4c55aa35e1ad" exitCode=0 Dec 06 04:00:28 crc kubenswrapper[4802]: I1206 04:00:28.826203 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-b8sxp" event={"ID":"04a970e8-495a-491a-9f83-1b25159450b1","Type":"ContainerDied","Data":"27a342c130d871ab7897d1d85dbab34c5507f64a62e6aa01189f4c55aa35e1ad"} Dec 06 04:00:28 crc kubenswrapper[4802]: I1206 04:00:28.844161 4802 generic.go:334] "Generic (PLEG): container finished" podID="2f247bdd-ddf7-4872-a1c5-597fec152cde" containerID="5ab9087cf180f2a18ce6757d996a25ac4c3e22341ed3ae95f9624968cb5de9b2" exitCode=0 Dec 06 04:00:28 crc kubenswrapper[4802]: I1206 04:00:28.844252 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-gv9nf" event={"ID":"2f247bdd-ddf7-4872-a1c5-597fec152cde","Type":"ContainerDied","Data":"5ab9087cf180f2a18ce6757d996a25ac4c3e22341ed3ae95f9624968cb5de9b2"} Dec 06 04:00:28 crc kubenswrapper[4802]: I1206 04:00:28.849126 4802 generic.go:334] "Generic (PLEG): container finished" podID="b2e9e18c-0730-4b74-8eaf-941cbeb08246" containerID="be5c2c6140d48e66a81cd2b5f33d8dd4541e5161488459f5109327ce75d6bccf" exitCode=0 Dec 06 04:00:28 crc kubenswrapper[4802]: I1206 04:00:28.849192 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-rhcrb" event={"ID":"b2e9e18c-0730-4b74-8eaf-941cbeb08246","Type":"ContainerDied","Data":"be5c2c6140d48e66a81cd2b5f33d8dd4541e5161488459f5109327ce75d6bccf"} Dec 06 04:00:28 crc kubenswrapper[4802]: I1206 04:00:28.851766 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-66f6-account-create-update-th68q" event={"ID":"8a5a43d4-c219-4090-acc3-252e127e48e7","Type":"ContainerStarted","Data":"8f9b17b6778156f4f51764b0b22842a3799ddd93aca19a6b91890d38395cbde7"} Dec 06 04:00:28 crc kubenswrapper[4802]: I1206 04:00:28.851797 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-66f6-account-create-update-th68q" event={"ID":"8a5a43d4-c219-4090-acc3-252e127e48e7","Type":"ContainerStarted","Data":"bd97e2dc55442562f71c6e00c58d021ea9686a6fdae52b6ee4688e9af003cbd0"} Dec 06 04:00:28 crc kubenswrapper[4802]: I1206 04:00:28.857833 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e13eaebb-eae6-4924-8940-9c899466b9ad","Type":"ContainerStarted","Data":"274c790f2eed55f67a6fccc26373e759bcd33643ca3200f9e001c6b44a32e349"} Dec 06 04:00:28 crc kubenswrapper[4802]: I1206 04:00:28.865798 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cdd3-account-create-update-bjk6c" podStartSLOduration=2.8657804650000003 podStartE2EDuration="2.865780465s" podCreationTimestamp="2025-12-06 04:00:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:00:28.855967829 +0000 UTC m=+1221.727876981" watchObservedRunningTime="2025-12-06 04:00:28.865780465 +0000 UTC m=+1221.737689617" Dec 06 04:00:28 crc kubenswrapper[4802]: I1206 04:00:28.892515 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-66f6-account-create-update-th68q" podStartSLOduration=2.892498213 podStartE2EDuration="2.892498213s" podCreationTimestamp="2025-12-06 04:00:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:00:28.886635953 +0000 UTC m=+1221.758545115" watchObservedRunningTime="2025-12-06 04:00:28.892498213 +0000 UTC m=+1221.764407355" Dec 06 04:00:29 crc kubenswrapper[4802]: I1206 04:00:29.869527 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e13eaebb-eae6-4924-8940-9c899466b9ad","Type":"ContainerStarted","Data":"50661514a1b6227caec3ab91719f2ccb04df0f94cd7d1c533d692cbcfdf293ce"} Dec 06 04:00:29 crc kubenswrapper[4802]: I1206 04:00:29.876146 4802 generic.go:334] "Generic (PLEG): container finished" podID="59d551a5-c1ef-46ee-8ae2-f98526b2ab09" containerID="a6c5c28aa66df89c8e48ac36cf218e1d67940dcd025ae23684066eb35df96f0d" exitCode=0 Dec 06 04:00:29 crc kubenswrapper[4802]: I1206 04:00:29.876550 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-1043-account-create-update-xzd9x" event={"ID":"59d551a5-c1ef-46ee-8ae2-f98526b2ab09","Type":"ContainerDied","Data":"a6c5c28aa66df89c8e48ac36cf218e1d67940dcd025ae23684066eb35df96f0d"} Dec 06 04:00:30 crc kubenswrapper[4802]: I1206 04:00:30.311117 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-rhcrb" Dec 06 04:00:30 crc kubenswrapper[4802]: I1206 04:00:30.398829 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b2e9e18c-0730-4b74-8eaf-941cbeb08246-operator-scripts\") pod \"b2e9e18c-0730-4b74-8eaf-941cbeb08246\" (UID: \"b2e9e18c-0730-4b74-8eaf-941cbeb08246\") " Dec 06 04:00:30 crc kubenswrapper[4802]: I1206 04:00:30.398871 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q4gsh\" (UniqueName: \"kubernetes.io/projected/b2e9e18c-0730-4b74-8eaf-941cbeb08246-kube-api-access-q4gsh\") pod \"b2e9e18c-0730-4b74-8eaf-941cbeb08246\" (UID: \"b2e9e18c-0730-4b74-8eaf-941cbeb08246\") " Dec 06 04:00:30 crc kubenswrapper[4802]: I1206 04:00:30.403250 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2e9e18c-0730-4b74-8eaf-941cbeb08246-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b2e9e18c-0730-4b74-8eaf-941cbeb08246" (UID: "b2e9e18c-0730-4b74-8eaf-941cbeb08246"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:00:30 crc kubenswrapper[4802]: I1206 04:00:30.407559 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2e9e18c-0730-4b74-8eaf-941cbeb08246-kube-api-access-q4gsh" (OuterVolumeSpecName: "kube-api-access-q4gsh") pod "b2e9e18c-0730-4b74-8eaf-941cbeb08246" (UID: "b2e9e18c-0730-4b74-8eaf-941cbeb08246"). InnerVolumeSpecName "kube-api-access-q4gsh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:00:30 crc kubenswrapper[4802]: I1206 04:00:30.491903 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-b8sxp" Dec 06 04:00:30 crc kubenswrapper[4802]: I1206 04:00:30.500498 4802 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b2e9e18c-0730-4b74-8eaf-941cbeb08246-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:30 crc kubenswrapper[4802]: I1206 04:00:30.500525 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q4gsh\" (UniqueName: \"kubernetes.io/projected/b2e9e18c-0730-4b74-8eaf-941cbeb08246-kube-api-access-q4gsh\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:30 crc kubenswrapper[4802]: I1206 04:00:30.502725 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-gv9nf" Dec 06 04:00:30 crc kubenswrapper[4802]: I1206 04:00:30.601280 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-87l9f\" (UniqueName: \"kubernetes.io/projected/2f247bdd-ddf7-4872-a1c5-597fec152cde-kube-api-access-87l9f\") pod \"2f247bdd-ddf7-4872-a1c5-597fec152cde\" (UID: \"2f247bdd-ddf7-4872-a1c5-597fec152cde\") " Dec 06 04:00:30 crc kubenswrapper[4802]: I1206 04:00:30.601408 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-czqc9\" (UniqueName: \"kubernetes.io/projected/04a970e8-495a-491a-9f83-1b25159450b1-kube-api-access-czqc9\") pod \"04a970e8-495a-491a-9f83-1b25159450b1\" (UID: \"04a970e8-495a-491a-9f83-1b25159450b1\") " Dec 06 04:00:30 crc kubenswrapper[4802]: I1206 04:00:30.601471 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2f247bdd-ddf7-4872-a1c5-597fec152cde-operator-scripts\") pod \"2f247bdd-ddf7-4872-a1c5-597fec152cde\" (UID: \"2f247bdd-ddf7-4872-a1c5-597fec152cde\") " Dec 06 04:00:30 crc kubenswrapper[4802]: I1206 04:00:30.601602 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/04a970e8-495a-491a-9f83-1b25159450b1-operator-scripts\") pod \"04a970e8-495a-491a-9f83-1b25159450b1\" (UID: \"04a970e8-495a-491a-9f83-1b25159450b1\") " Dec 06 04:00:30 crc kubenswrapper[4802]: I1206 04:00:30.602323 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f247bdd-ddf7-4872-a1c5-597fec152cde-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2f247bdd-ddf7-4872-a1c5-597fec152cde" (UID: "2f247bdd-ddf7-4872-a1c5-597fec152cde"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:00:30 crc kubenswrapper[4802]: I1206 04:00:30.602464 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04a970e8-495a-491a-9f83-1b25159450b1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "04a970e8-495a-491a-9f83-1b25159450b1" (UID: "04a970e8-495a-491a-9f83-1b25159450b1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:00:30 crc kubenswrapper[4802]: I1206 04:00:30.605385 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f247bdd-ddf7-4872-a1c5-597fec152cde-kube-api-access-87l9f" (OuterVolumeSpecName: "kube-api-access-87l9f") pod "2f247bdd-ddf7-4872-a1c5-597fec152cde" (UID: "2f247bdd-ddf7-4872-a1c5-597fec152cde"). InnerVolumeSpecName "kube-api-access-87l9f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:00:30 crc kubenswrapper[4802]: I1206 04:00:30.611608 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04a970e8-495a-491a-9f83-1b25159450b1-kube-api-access-czqc9" (OuterVolumeSpecName: "kube-api-access-czqc9") pod "04a970e8-495a-491a-9f83-1b25159450b1" (UID: "04a970e8-495a-491a-9f83-1b25159450b1"). InnerVolumeSpecName "kube-api-access-czqc9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:00:30 crc kubenswrapper[4802]: I1206 04:00:30.703807 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-czqc9\" (UniqueName: \"kubernetes.io/projected/04a970e8-495a-491a-9f83-1b25159450b1-kube-api-access-czqc9\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:30 crc kubenswrapper[4802]: I1206 04:00:30.704097 4802 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2f247bdd-ddf7-4872-a1c5-597fec152cde-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:30 crc kubenswrapper[4802]: I1206 04:00:30.704111 4802 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/04a970e8-495a-491a-9f83-1b25159450b1-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:30 crc kubenswrapper[4802]: I1206 04:00:30.704123 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-87l9f\" (UniqueName: \"kubernetes.io/projected/2f247bdd-ddf7-4872-a1c5-597fec152cde-kube-api-access-87l9f\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:30 crc kubenswrapper[4802]: I1206 04:00:30.888409 4802 generic.go:334] "Generic (PLEG): container finished" podID="8a5a43d4-c219-4090-acc3-252e127e48e7" containerID="8f9b17b6778156f4f51764b0b22842a3799ddd93aca19a6b91890d38395cbde7" exitCode=0 Dec 06 04:00:30 crc kubenswrapper[4802]: I1206 04:00:30.888431 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-66f6-account-create-update-th68q" event={"ID":"8a5a43d4-c219-4090-acc3-252e127e48e7","Type":"ContainerDied","Data":"8f9b17b6778156f4f51764b0b22842a3799ddd93aca19a6b91890d38395cbde7"} Dec 06 04:00:30 crc kubenswrapper[4802]: I1206 04:00:30.890987 4802 generic.go:334] "Generic (PLEG): container finished" podID="164b8cc5-aab5-4e71-af01-d39679d4b3ba" containerID="71f041b5f79ae84d3ca26838b981a0e01fda8163adf89de78ed9c1bdb15ccff3" exitCode=0 Dec 06 04:00:30 crc kubenswrapper[4802]: I1206 04:00:30.891020 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cdd3-account-create-update-bjk6c" event={"ID":"164b8cc5-aab5-4e71-af01-d39679d4b3ba","Type":"ContainerDied","Data":"71f041b5f79ae84d3ca26838b981a0e01fda8163adf89de78ed9c1bdb15ccff3"} Dec 06 04:00:30 crc kubenswrapper[4802]: I1206 04:00:30.893373 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-b8sxp" event={"ID":"04a970e8-495a-491a-9f83-1b25159450b1","Type":"ContainerDied","Data":"3075cd1dc6fc542a71d7fe4db60ccecfbf362f4a8e63e246a4956368ecfb00e0"} Dec 06 04:00:30 crc kubenswrapper[4802]: I1206 04:00:30.893409 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3075cd1dc6fc542a71d7fe4db60ccecfbf362f4a8e63e246a4956368ecfb00e0" Dec 06 04:00:30 crc kubenswrapper[4802]: I1206 04:00:30.894065 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-b8sxp" Dec 06 04:00:30 crc kubenswrapper[4802]: I1206 04:00:30.895211 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-gv9nf" event={"ID":"2f247bdd-ddf7-4872-a1c5-597fec152cde","Type":"ContainerDied","Data":"0e7f9793d27a6fc8131ef36388f67af7f156baebed2e20f4e765c08af68fa4d5"} Dec 06 04:00:30 crc kubenswrapper[4802]: I1206 04:00:30.895258 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0e7f9793d27a6fc8131ef36388f67af7f156baebed2e20f4e765c08af68fa4d5" Dec 06 04:00:30 crc kubenswrapper[4802]: I1206 04:00:30.895316 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-gv9nf" Dec 06 04:00:30 crc kubenswrapper[4802]: I1206 04:00:30.897298 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-rhcrb" event={"ID":"b2e9e18c-0730-4b74-8eaf-941cbeb08246","Type":"ContainerDied","Data":"e760325fd8c5ac0b334091d72ec787801e5b3752b6a5b79af58b3464668f4397"} Dec 06 04:00:30 crc kubenswrapper[4802]: I1206 04:00:30.897329 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e760325fd8c5ac0b334091d72ec787801e5b3752b6a5b79af58b3464668f4397" Dec 06 04:00:30 crc kubenswrapper[4802]: I1206 04:00:30.897373 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-rhcrb" Dec 06 04:00:30 crc kubenswrapper[4802]: I1206 04:00:30.913123 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e13eaebb-eae6-4924-8940-9c899466b9ad","Type":"ContainerStarted","Data":"51346b8336d0a5d2e1ba8efeca7ef377dc0eef4d543eb15a02c6bc246a78aad3"} Dec 06 04:00:31 crc kubenswrapper[4802]: I1206 04:00:31.198929 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-1043-account-create-update-xzd9x" Dec 06 04:00:31 crc kubenswrapper[4802]: I1206 04:00:31.238717 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-phn2m\" (UniqueName: \"kubernetes.io/projected/59d551a5-c1ef-46ee-8ae2-f98526b2ab09-kube-api-access-phn2m\") pod \"59d551a5-c1ef-46ee-8ae2-f98526b2ab09\" (UID: \"59d551a5-c1ef-46ee-8ae2-f98526b2ab09\") " Dec 06 04:00:31 crc kubenswrapper[4802]: I1206 04:00:31.238996 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/59d551a5-c1ef-46ee-8ae2-f98526b2ab09-operator-scripts\") pod \"59d551a5-c1ef-46ee-8ae2-f98526b2ab09\" (UID: \"59d551a5-c1ef-46ee-8ae2-f98526b2ab09\") " Dec 06 04:00:31 crc kubenswrapper[4802]: I1206 04:00:31.240182 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59d551a5-c1ef-46ee-8ae2-f98526b2ab09-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "59d551a5-c1ef-46ee-8ae2-f98526b2ab09" (UID: "59d551a5-c1ef-46ee-8ae2-f98526b2ab09"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:00:31 crc kubenswrapper[4802]: I1206 04:00:31.245453 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59d551a5-c1ef-46ee-8ae2-f98526b2ab09-kube-api-access-phn2m" (OuterVolumeSpecName: "kube-api-access-phn2m") pod "59d551a5-c1ef-46ee-8ae2-f98526b2ab09" (UID: "59d551a5-c1ef-46ee-8ae2-f98526b2ab09"). InnerVolumeSpecName "kube-api-access-phn2m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:00:31 crc kubenswrapper[4802]: I1206 04:00:31.341268 4802 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/59d551a5-c1ef-46ee-8ae2-f98526b2ab09-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:31 crc kubenswrapper[4802]: I1206 04:00:31.341620 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-phn2m\" (UniqueName: \"kubernetes.io/projected/59d551a5-c1ef-46ee-8ae2-f98526b2ab09-kube-api-access-phn2m\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:31 crc kubenswrapper[4802]: I1206 04:00:31.931956 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-1043-account-create-update-xzd9x" event={"ID":"59d551a5-c1ef-46ee-8ae2-f98526b2ab09","Type":"ContainerDied","Data":"1460f3c36934cefede236489181860f0e1245a09ad8a6c10c21284573a3e315c"} Dec 06 04:00:31 crc kubenswrapper[4802]: I1206 04:00:31.932022 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1460f3c36934cefede236489181860f0e1245a09ad8a6c10c21284573a3e315c" Dec 06 04:00:31 crc kubenswrapper[4802]: I1206 04:00:31.932110 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-1043-account-create-update-xzd9x" Dec 06 04:00:32 crc kubenswrapper[4802]: I1206 04:00:32.366080 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cdd3-account-create-update-bjk6c" Dec 06 04:00:32 crc kubenswrapper[4802]: I1206 04:00:32.372236 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-66f6-account-create-update-th68q" Dec 06 04:00:32 crc kubenswrapper[4802]: I1206 04:00:32.473776 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/164b8cc5-aab5-4e71-af01-d39679d4b3ba-operator-scripts\") pod \"164b8cc5-aab5-4e71-af01-d39679d4b3ba\" (UID: \"164b8cc5-aab5-4e71-af01-d39679d4b3ba\") " Dec 06 04:00:32 crc kubenswrapper[4802]: I1206 04:00:32.473860 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a5a43d4-c219-4090-acc3-252e127e48e7-operator-scripts\") pod \"8a5a43d4-c219-4090-acc3-252e127e48e7\" (UID: \"8a5a43d4-c219-4090-acc3-252e127e48e7\") " Dec 06 04:00:32 crc kubenswrapper[4802]: I1206 04:00:32.474038 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kwbqw\" (UniqueName: \"kubernetes.io/projected/8a5a43d4-c219-4090-acc3-252e127e48e7-kube-api-access-kwbqw\") pod \"8a5a43d4-c219-4090-acc3-252e127e48e7\" (UID: \"8a5a43d4-c219-4090-acc3-252e127e48e7\") " Dec 06 04:00:32 crc kubenswrapper[4802]: I1206 04:00:32.474095 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sk6ph\" (UniqueName: \"kubernetes.io/projected/164b8cc5-aab5-4e71-af01-d39679d4b3ba-kube-api-access-sk6ph\") pod \"164b8cc5-aab5-4e71-af01-d39679d4b3ba\" (UID: \"164b8cc5-aab5-4e71-af01-d39679d4b3ba\") " Dec 06 04:00:32 crc kubenswrapper[4802]: I1206 04:00:32.474764 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a5a43d4-c219-4090-acc3-252e127e48e7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8a5a43d4-c219-4090-acc3-252e127e48e7" (UID: "8a5a43d4-c219-4090-acc3-252e127e48e7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:00:32 crc kubenswrapper[4802]: I1206 04:00:32.475103 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/164b8cc5-aab5-4e71-af01-d39679d4b3ba-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "164b8cc5-aab5-4e71-af01-d39679d4b3ba" (UID: "164b8cc5-aab5-4e71-af01-d39679d4b3ba"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:00:32 crc kubenswrapper[4802]: I1206 04:00:32.485746 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a5a43d4-c219-4090-acc3-252e127e48e7-kube-api-access-kwbqw" (OuterVolumeSpecName: "kube-api-access-kwbqw") pod "8a5a43d4-c219-4090-acc3-252e127e48e7" (UID: "8a5a43d4-c219-4090-acc3-252e127e48e7"). InnerVolumeSpecName "kube-api-access-kwbqw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:00:32 crc kubenswrapper[4802]: I1206 04:00:32.487841 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/164b8cc5-aab5-4e71-af01-d39679d4b3ba-kube-api-access-sk6ph" (OuterVolumeSpecName: "kube-api-access-sk6ph") pod "164b8cc5-aab5-4e71-af01-d39679d4b3ba" (UID: "164b8cc5-aab5-4e71-af01-d39679d4b3ba"). InnerVolumeSpecName "kube-api-access-sk6ph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:00:32 crc kubenswrapper[4802]: I1206 04:00:32.577381 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kwbqw\" (UniqueName: \"kubernetes.io/projected/8a5a43d4-c219-4090-acc3-252e127e48e7-kube-api-access-kwbqw\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:32 crc kubenswrapper[4802]: I1206 04:00:32.577416 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sk6ph\" (UniqueName: \"kubernetes.io/projected/164b8cc5-aab5-4e71-af01-d39679d4b3ba-kube-api-access-sk6ph\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:32 crc kubenswrapper[4802]: I1206 04:00:32.577429 4802 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/164b8cc5-aab5-4e71-af01-d39679d4b3ba-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:32 crc kubenswrapper[4802]: I1206 04:00:32.577442 4802 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a5a43d4-c219-4090-acc3-252e127e48e7-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:32 crc kubenswrapper[4802]: I1206 04:00:32.959986 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cdd3-account-create-update-bjk6c" Dec 06 04:00:32 crc kubenswrapper[4802]: I1206 04:00:32.961459 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cdd3-account-create-update-bjk6c" event={"ID":"164b8cc5-aab5-4e71-af01-d39679d4b3ba","Type":"ContainerDied","Data":"ad5035d9c0a9713ee286fc5f357cf0ddf5d25d8808c617ef34049db28ecbec87"} Dec 06 04:00:32 crc kubenswrapper[4802]: I1206 04:00:32.961572 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ad5035d9c0a9713ee286fc5f357cf0ddf5d25d8808c617ef34049db28ecbec87" Dec 06 04:00:32 crc kubenswrapper[4802]: I1206 04:00:32.973101 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-66f6-account-create-update-th68q" event={"ID":"8a5a43d4-c219-4090-acc3-252e127e48e7","Type":"ContainerDied","Data":"bd97e2dc55442562f71c6e00c58d021ea9686a6fdae52b6ee4688e9af003cbd0"} Dec 06 04:00:32 crc kubenswrapper[4802]: I1206 04:00:32.973344 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bd97e2dc55442562f71c6e00c58d021ea9686a6fdae52b6ee4688e9af003cbd0" Dec 06 04:00:32 crc kubenswrapper[4802]: I1206 04:00:32.973395 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-66f6-account-create-update-th68q" Dec 06 04:00:33 crc kubenswrapper[4802]: I1206 04:00:33.982336 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e13eaebb-eae6-4924-8940-9c899466b9ad","Type":"ContainerStarted","Data":"5baeab8eb36579d18ecafdf8bb17f68bfb647d42a04756ae8b5a6a82f64a5610"} Dec 06 04:00:33 crc kubenswrapper[4802]: I1206 04:00:33.982847 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 06 04:00:34 crc kubenswrapper[4802]: I1206 04:00:34.016992 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.856422349 podStartE2EDuration="9.016976594s" podCreationTimestamp="2025-12-06 04:00:25 +0000 UTC" firstStartedPulling="2025-12-06 04:00:26.678484816 +0000 UTC m=+1219.550393968" lastFinishedPulling="2025-12-06 04:00:32.839039061 +0000 UTC m=+1225.710948213" observedRunningTime="2025-12-06 04:00:34.013526701 +0000 UTC m=+1226.885435883" watchObservedRunningTime="2025-12-06 04:00:34.016976594 +0000 UTC m=+1226.888885746" Dec 06 04:00:36 crc kubenswrapper[4802]: I1206 04:00:36.772891 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-xnnxk"] Dec 06 04:00:36 crc kubenswrapper[4802]: E1206 04:00:36.773849 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2e9e18c-0730-4b74-8eaf-941cbeb08246" containerName="mariadb-database-create" Dec 06 04:00:36 crc kubenswrapper[4802]: I1206 04:00:36.773865 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2e9e18c-0730-4b74-8eaf-941cbeb08246" containerName="mariadb-database-create" Dec 06 04:00:36 crc kubenswrapper[4802]: E1206 04:00:36.773889 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="164b8cc5-aab5-4e71-af01-d39679d4b3ba" containerName="mariadb-account-create-update" Dec 06 04:00:36 crc kubenswrapper[4802]: I1206 04:00:36.773900 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="164b8cc5-aab5-4e71-af01-d39679d4b3ba" containerName="mariadb-account-create-update" Dec 06 04:00:36 crc kubenswrapper[4802]: E1206 04:00:36.773927 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59d551a5-c1ef-46ee-8ae2-f98526b2ab09" containerName="mariadb-account-create-update" Dec 06 04:00:36 crc kubenswrapper[4802]: I1206 04:00:36.773936 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="59d551a5-c1ef-46ee-8ae2-f98526b2ab09" containerName="mariadb-account-create-update" Dec 06 04:00:36 crc kubenswrapper[4802]: E1206 04:00:36.773947 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a5a43d4-c219-4090-acc3-252e127e48e7" containerName="mariadb-account-create-update" Dec 06 04:00:36 crc kubenswrapper[4802]: I1206 04:00:36.773955 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a5a43d4-c219-4090-acc3-252e127e48e7" containerName="mariadb-account-create-update" Dec 06 04:00:36 crc kubenswrapper[4802]: E1206 04:00:36.773966 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04a970e8-495a-491a-9f83-1b25159450b1" containerName="mariadb-database-create" Dec 06 04:00:36 crc kubenswrapper[4802]: I1206 04:00:36.773974 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="04a970e8-495a-491a-9f83-1b25159450b1" containerName="mariadb-database-create" Dec 06 04:00:36 crc kubenswrapper[4802]: E1206 04:00:36.773986 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f247bdd-ddf7-4872-a1c5-597fec152cde" containerName="mariadb-database-create" Dec 06 04:00:36 crc kubenswrapper[4802]: I1206 04:00:36.773995 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f247bdd-ddf7-4872-a1c5-597fec152cde" containerName="mariadb-database-create" Dec 06 04:00:36 crc kubenswrapper[4802]: I1206 04:00:36.774213 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="164b8cc5-aab5-4e71-af01-d39679d4b3ba" containerName="mariadb-account-create-update" Dec 06 04:00:36 crc kubenswrapper[4802]: I1206 04:00:36.774234 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a5a43d4-c219-4090-acc3-252e127e48e7" containerName="mariadb-account-create-update" Dec 06 04:00:36 crc kubenswrapper[4802]: I1206 04:00:36.774251 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="04a970e8-495a-491a-9f83-1b25159450b1" containerName="mariadb-database-create" Dec 06 04:00:36 crc kubenswrapper[4802]: I1206 04:00:36.774263 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="59d551a5-c1ef-46ee-8ae2-f98526b2ab09" containerName="mariadb-account-create-update" Dec 06 04:00:36 crc kubenswrapper[4802]: I1206 04:00:36.774279 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2e9e18c-0730-4b74-8eaf-941cbeb08246" containerName="mariadb-database-create" Dec 06 04:00:36 crc kubenswrapper[4802]: I1206 04:00:36.774297 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f247bdd-ddf7-4872-a1c5-597fec152cde" containerName="mariadb-database-create" Dec 06 04:00:36 crc kubenswrapper[4802]: I1206 04:00:36.775034 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-xnnxk" Dec 06 04:00:36 crc kubenswrapper[4802]: I1206 04:00:36.780469 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-z765z" Dec 06 04:00:36 crc kubenswrapper[4802]: I1206 04:00:36.782166 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 06 04:00:36 crc kubenswrapper[4802]: I1206 04:00:36.782991 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 06 04:00:36 crc kubenswrapper[4802]: I1206 04:00:36.796504 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-xnnxk"] Dec 06 04:00:36 crc kubenswrapper[4802]: I1206 04:00:36.881030 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f09f7e4f-6e98-4461-bf3a-1a6aa9528c09-scripts\") pod \"nova-cell0-conductor-db-sync-xnnxk\" (UID: \"f09f7e4f-6e98-4461-bf3a-1a6aa9528c09\") " pod="openstack/nova-cell0-conductor-db-sync-xnnxk" Dec 06 04:00:36 crc kubenswrapper[4802]: I1206 04:00:36.881107 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f09f7e4f-6e98-4461-bf3a-1a6aa9528c09-config-data\") pod \"nova-cell0-conductor-db-sync-xnnxk\" (UID: \"f09f7e4f-6e98-4461-bf3a-1a6aa9528c09\") " pod="openstack/nova-cell0-conductor-db-sync-xnnxk" Dec 06 04:00:36 crc kubenswrapper[4802]: I1206 04:00:36.881218 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f09f7e4f-6e98-4461-bf3a-1a6aa9528c09-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-xnnxk\" (UID: \"f09f7e4f-6e98-4461-bf3a-1a6aa9528c09\") " pod="openstack/nova-cell0-conductor-db-sync-xnnxk" Dec 06 04:00:36 crc kubenswrapper[4802]: I1206 04:00:36.881607 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-csvhd\" (UniqueName: \"kubernetes.io/projected/f09f7e4f-6e98-4461-bf3a-1a6aa9528c09-kube-api-access-csvhd\") pod \"nova-cell0-conductor-db-sync-xnnxk\" (UID: \"f09f7e4f-6e98-4461-bf3a-1a6aa9528c09\") " pod="openstack/nova-cell0-conductor-db-sync-xnnxk" Dec 06 04:00:36 crc kubenswrapper[4802]: I1206 04:00:36.983393 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f09f7e4f-6e98-4461-bf3a-1a6aa9528c09-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-xnnxk\" (UID: \"f09f7e4f-6e98-4461-bf3a-1a6aa9528c09\") " pod="openstack/nova-cell0-conductor-db-sync-xnnxk" Dec 06 04:00:36 crc kubenswrapper[4802]: I1206 04:00:36.983462 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-csvhd\" (UniqueName: \"kubernetes.io/projected/f09f7e4f-6e98-4461-bf3a-1a6aa9528c09-kube-api-access-csvhd\") pod \"nova-cell0-conductor-db-sync-xnnxk\" (UID: \"f09f7e4f-6e98-4461-bf3a-1a6aa9528c09\") " pod="openstack/nova-cell0-conductor-db-sync-xnnxk" Dec 06 04:00:36 crc kubenswrapper[4802]: I1206 04:00:36.983529 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f09f7e4f-6e98-4461-bf3a-1a6aa9528c09-scripts\") pod \"nova-cell0-conductor-db-sync-xnnxk\" (UID: \"f09f7e4f-6e98-4461-bf3a-1a6aa9528c09\") " pod="openstack/nova-cell0-conductor-db-sync-xnnxk" Dec 06 04:00:36 crc kubenswrapper[4802]: I1206 04:00:36.983563 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f09f7e4f-6e98-4461-bf3a-1a6aa9528c09-config-data\") pod \"nova-cell0-conductor-db-sync-xnnxk\" (UID: \"f09f7e4f-6e98-4461-bf3a-1a6aa9528c09\") " pod="openstack/nova-cell0-conductor-db-sync-xnnxk" Dec 06 04:00:36 crc kubenswrapper[4802]: I1206 04:00:36.989706 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f09f7e4f-6e98-4461-bf3a-1a6aa9528c09-config-data\") pod \"nova-cell0-conductor-db-sync-xnnxk\" (UID: \"f09f7e4f-6e98-4461-bf3a-1a6aa9528c09\") " pod="openstack/nova-cell0-conductor-db-sync-xnnxk" Dec 06 04:00:36 crc kubenswrapper[4802]: I1206 04:00:36.995235 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f09f7e4f-6e98-4461-bf3a-1a6aa9528c09-scripts\") pod \"nova-cell0-conductor-db-sync-xnnxk\" (UID: \"f09f7e4f-6e98-4461-bf3a-1a6aa9528c09\") " pod="openstack/nova-cell0-conductor-db-sync-xnnxk" Dec 06 04:00:36 crc kubenswrapper[4802]: I1206 04:00:36.995818 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f09f7e4f-6e98-4461-bf3a-1a6aa9528c09-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-xnnxk\" (UID: \"f09f7e4f-6e98-4461-bf3a-1a6aa9528c09\") " pod="openstack/nova-cell0-conductor-db-sync-xnnxk" Dec 06 04:00:37 crc kubenswrapper[4802]: I1206 04:00:37.000724 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-csvhd\" (UniqueName: \"kubernetes.io/projected/f09f7e4f-6e98-4461-bf3a-1a6aa9528c09-kube-api-access-csvhd\") pod \"nova-cell0-conductor-db-sync-xnnxk\" (UID: \"f09f7e4f-6e98-4461-bf3a-1a6aa9528c09\") " pod="openstack/nova-cell0-conductor-db-sync-xnnxk" Dec 06 04:00:37 crc kubenswrapper[4802]: I1206 04:00:37.100406 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-xnnxk" Dec 06 04:00:37 crc kubenswrapper[4802]: E1206 04:00:37.459009 4802 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="15461749ec2cccc9a0f2a06a22252592288f81b9b432dde2bc6f7f9917e9bb8a" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 06 04:00:37 crc kubenswrapper[4802]: E1206 04:00:37.462140 4802 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="15461749ec2cccc9a0f2a06a22252592288f81b9b432dde2bc6f7f9917e9bb8a" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 06 04:00:37 crc kubenswrapper[4802]: E1206 04:00:37.463282 4802 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="15461749ec2cccc9a0f2a06a22252592288f81b9b432dde2bc6f7f9917e9bb8a" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 06 04:00:37 crc kubenswrapper[4802]: E1206 04:00:37.463334 4802 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-5d96658447-rg8zs" podUID="d3da4180-e8fe-4785-8d4f-6cd172fb6399" containerName="heat-engine" Dec 06 04:00:37 crc kubenswrapper[4802]: I1206 04:00:37.576267 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-xnnxk"] Dec 06 04:00:37 crc kubenswrapper[4802]: W1206 04:00:37.583278 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf09f7e4f_6e98_4461_bf3a_1a6aa9528c09.slice/crio-f6179c8e1032314aca8ad9f81dcfb870cb9717ebf0a2fdd6cc8068e3b5b3f2dd WatchSource:0}: Error finding container f6179c8e1032314aca8ad9f81dcfb870cb9717ebf0a2fdd6cc8068e3b5b3f2dd: Status 404 returned error can't find the container with id f6179c8e1032314aca8ad9f81dcfb870cb9717ebf0a2fdd6cc8068e3b5b3f2dd Dec 06 04:00:37 crc kubenswrapper[4802]: I1206 04:00:37.585993 4802 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 04:00:38 crc kubenswrapper[4802]: I1206 04:00:38.027490 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-xnnxk" event={"ID":"f09f7e4f-6e98-4461-bf3a-1a6aa9528c09","Type":"ContainerStarted","Data":"f6179c8e1032314aca8ad9f81dcfb870cb9717ebf0a2fdd6cc8068e3b5b3f2dd"} Dec 06 04:00:38 crc kubenswrapper[4802]: I1206 04:00:38.953885 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 04:00:38 crc kubenswrapper[4802]: I1206 04:00:38.954428 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e13eaebb-eae6-4924-8940-9c899466b9ad" containerName="ceilometer-central-agent" containerID="cri-o://274c790f2eed55f67a6fccc26373e759bcd33643ca3200f9e001c6b44a32e349" gracePeriod=30 Dec 06 04:00:38 crc kubenswrapper[4802]: I1206 04:00:38.954957 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e13eaebb-eae6-4924-8940-9c899466b9ad" containerName="proxy-httpd" containerID="cri-o://5baeab8eb36579d18ecafdf8bb17f68bfb647d42a04756ae8b5a6a82f64a5610" gracePeriod=30 Dec 06 04:00:38 crc kubenswrapper[4802]: I1206 04:00:38.955025 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e13eaebb-eae6-4924-8940-9c899466b9ad" containerName="sg-core" containerID="cri-o://51346b8336d0a5d2e1ba8efeca7ef377dc0eef4d543eb15a02c6bc246a78aad3" gracePeriod=30 Dec 06 04:00:38 crc kubenswrapper[4802]: I1206 04:00:38.955059 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e13eaebb-eae6-4924-8940-9c899466b9ad" containerName="ceilometer-notification-agent" containerID="cri-o://50661514a1b6227caec3ab91719f2ccb04df0f94cd7d1c533d692cbcfdf293ce" gracePeriod=30 Dec 06 04:00:40 crc kubenswrapper[4802]: I1206 04:00:40.048307 4802 generic.go:334] "Generic (PLEG): container finished" podID="e13eaebb-eae6-4924-8940-9c899466b9ad" containerID="5baeab8eb36579d18ecafdf8bb17f68bfb647d42a04756ae8b5a6a82f64a5610" exitCode=0 Dec 06 04:00:40 crc kubenswrapper[4802]: I1206 04:00:40.048898 4802 generic.go:334] "Generic (PLEG): container finished" podID="e13eaebb-eae6-4924-8940-9c899466b9ad" containerID="51346b8336d0a5d2e1ba8efeca7ef377dc0eef4d543eb15a02c6bc246a78aad3" exitCode=2 Dec 06 04:00:40 crc kubenswrapper[4802]: I1206 04:00:40.048914 4802 generic.go:334] "Generic (PLEG): container finished" podID="e13eaebb-eae6-4924-8940-9c899466b9ad" containerID="274c790f2eed55f67a6fccc26373e759bcd33643ca3200f9e001c6b44a32e349" exitCode=0 Dec 06 04:00:40 crc kubenswrapper[4802]: I1206 04:00:40.048534 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e13eaebb-eae6-4924-8940-9c899466b9ad","Type":"ContainerDied","Data":"5baeab8eb36579d18ecafdf8bb17f68bfb647d42a04756ae8b5a6a82f64a5610"} Dec 06 04:00:40 crc kubenswrapper[4802]: I1206 04:00:40.048990 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e13eaebb-eae6-4924-8940-9c899466b9ad","Type":"ContainerDied","Data":"51346b8336d0a5d2e1ba8efeca7ef377dc0eef4d543eb15a02c6bc246a78aad3"} Dec 06 04:00:40 crc kubenswrapper[4802]: I1206 04:00:40.049009 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e13eaebb-eae6-4924-8940-9c899466b9ad","Type":"ContainerDied","Data":"274c790f2eed55f67a6fccc26373e759bcd33643ca3200f9e001c6b44a32e349"} Dec 06 04:00:40 crc kubenswrapper[4802]: I1206 04:00:40.052083 4802 generic.go:334] "Generic (PLEG): container finished" podID="d3da4180-e8fe-4785-8d4f-6cd172fb6399" containerID="15461749ec2cccc9a0f2a06a22252592288f81b9b432dde2bc6f7f9917e9bb8a" exitCode=0 Dec 06 04:00:40 crc kubenswrapper[4802]: I1206 04:00:40.052121 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-5d96658447-rg8zs" event={"ID":"d3da4180-e8fe-4785-8d4f-6cd172fb6399","Type":"ContainerDied","Data":"15461749ec2cccc9a0f2a06a22252592288f81b9b432dde2bc6f7f9917e9bb8a"} Dec 06 04:00:40 crc kubenswrapper[4802]: I1206 04:00:40.401439 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-5d96658447-rg8zs" Dec 06 04:00:40 crc kubenswrapper[4802]: I1206 04:00:40.444043 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3da4180-e8fe-4785-8d4f-6cd172fb6399-config-data\") pod \"d3da4180-e8fe-4785-8d4f-6cd172fb6399\" (UID: \"d3da4180-e8fe-4785-8d4f-6cd172fb6399\") " Dec 06 04:00:40 crc kubenswrapper[4802]: I1206 04:00:40.444107 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sv9rl\" (UniqueName: \"kubernetes.io/projected/d3da4180-e8fe-4785-8d4f-6cd172fb6399-kube-api-access-sv9rl\") pod \"d3da4180-e8fe-4785-8d4f-6cd172fb6399\" (UID: \"d3da4180-e8fe-4785-8d4f-6cd172fb6399\") " Dec 06 04:00:40 crc kubenswrapper[4802]: I1206 04:00:40.444339 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d3da4180-e8fe-4785-8d4f-6cd172fb6399-config-data-custom\") pod \"d3da4180-e8fe-4785-8d4f-6cd172fb6399\" (UID: \"d3da4180-e8fe-4785-8d4f-6cd172fb6399\") " Dec 06 04:00:40 crc kubenswrapper[4802]: I1206 04:00:40.444463 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3da4180-e8fe-4785-8d4f-6cd172fb6399-combined-ca-bundle\") pod \"d3da4180-e8fe-4785-8d4f-6cd172fb6399\" (UID: \"d3da4180-e8fe-4785-8d4f-6cd172fb6399\") " Dec 06 04:00:40 crc kubenswrapper[4802]: I1206 04:00:40.450802 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3da4180-e8fe-4785-8d4f-6cd172fb6399-kube-api-access-sv9rl" (OuterVolumeSpecName: "kube-api-access-sv9rl") pod "d3da4180-e8fe-4785-8d4f-6cd172fb6399" (UID: "d3da4180-e8fe-4785-8d4f-6cd172fb6399"). InnerVolumeSpecName "kube-api-access-sv9rl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:00:40 crc kubenswrapper[4802]: I1206 04:00:40.454869 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3da4180-e8fe-4785-8d4f-6cd172fb6399-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "d3da4180-e8fe-4785-8d4f-6cd172fb6399" (UID: "d3da4180-e8fe-4785-8d4f-6cd172fb6399"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:00:40 crc kubenswrapper[4802]: I1206 04:00:40.480866 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3da4180-e8fe-4785-8d4f-6cd172fb6399-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d3da4180-e8fe-4785-8d4f-6cd172fb6399" (UID: "d3da4180-e8fe-4785-8d4f-6cd172fb6399"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:00:40 crc kubenswrapper[4802]: I1206 04:00:40.518539 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3da4180-e8fe-4785-8d4f-6cd172fb6399-config-data" (OuterVolumeSpecName: "config-data") pod "d3da4180-e8fe-4785-8d4f-6cd172fb6399" (UID: "d3da4180-e8fe-4785-8d4f-6cd172fb6399"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:00:40 crc kubenswrapper[4802]: I1206 04:00:40.535296 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 04:00:40 crc kubenswrapper[4802]: I1206 04:00:40.547015 4802 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d3da4180-e8fe-4785-8d4f-6cd172fb6399-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:40 crc kubenswrapper[4802]: I1206 04:00:40.547045 4802 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3da4180-e8fe-4785-8d4f-6cd172fb6399-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:40 crc kubenswrapper[4802]: I1206 04:00:40.547054 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3da4180-e8fe-4785-8d4f-6cd172fb6399-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:40 crc kubenswrapper[4802]: I1206 04:00:40.547064 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sv9rl\" (UniqueName: \"kubernetes.io/projected/d3da4180-e8fe-4785-8d4f-6cd172fb6399-kube-api-access-sv9rl\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:40 crc kubenswrapper[4802]: I1206 04:00:40.648437 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e13eaebb-eae6-4924-8940-9c899466b9ad-combined-ca-bundle\") pod \"e13eaebb-eae6-4924-8940-9c899466b9ad\" (UID: \"e13eaebb-eae6-4924-8940-9c899466b9ad\") " Dec 06 04:00:40 crc kubenswrapper[4802]: I1206 04:00:40.649290 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e13eaebb-eae6-4924-8940-9c899466b9ad-scripts\") pod \"e13eaebb-eae6-4924-8940-9c899466b9ad\" (UID: \"e13eaebb-eae6-4924-8940-9c899466b9ad\") " Dec 06 04:00:40 crc kubenswrapper[4802]: I1206 04:00:40.649446 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e13eaebb-eae6-4924-8940-9c899466b9ad-config-data\") pod \"e13eaebb-eae6-4924-8940-9c899466b9ad\" (UID: \"e13eaebb-eae6-4924-8940-9c899466b9ad\") " Dec 06 04:00:40 crc kubenswrapper[4802]: I1206 04:00:40.649541 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e13eaebb-eae6-4924-8940-9c899466b9ad-log-httpd\") pod \"e13eaebb-eae6-4924-8940-9c899466b9ad\" (UID: \"e13eaebb-eae6-4924-8940-9c899466b9ad\") " Dec 06 04:00:40 crc kubenswrapper[4802]: I1206 04:00:40.649674 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e13eaebb-eae6-4924-8940-9c899466b9ad-run-httpd\") pod \"e13eaebb-eae6-4924-8940-9c899466b9ad\" (UID: \"e13eaebb-eae6-4924-8940-9c899466b9ad\") " Dec 06 04:00:40 crc kubenswrapper[4802]: I1206 04:00:40.649883 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q9nzw\" (UniqueName: \"kubernetes.io/projected/e13eaebb-eae6-4924-8940-9c899466b9ad-kube-api-access-q9nzw\") pod \"e13eaebb-eae6-4924-8940-9c899466b9ad\" (UID: \"e13eaebb-eae6-4924-8940-9c899466b9ad\") " Dec 06 04:00:40 crc kubenswrapper[4802]: I1206 04:00:40.650115 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e13eaebb-eae6-4924-8940-9c899466b9ad-sg-core-conf-yaml\") pod \"e13eaebb-eae6-4924-8940-9c899466b9ad\" (UID: \"e13eaebb-eae6-4924-8940-9c899466b9ad\") " Dec 06 04:00:40 crc kubenswrapper[4802]: I1206 04:00:40.650898 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e13eaebb-eae6-4924-8940-9c899466b9ad-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "e13eaebb-eae6-4924-8940-9c899466b9ad" (UID: "e13eaebb-eae6-4924-8940-9c899466b9ad"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:00:40 crc kubenswrapper[4802]: I1206 04:00:40.651276 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e13eaebb-eae6-4924-8940-9c899466b9ad-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "e13eaebb-eae6-4924-8940-9c899466b9ad" (UID: "e13eaebb-eae6-4924-8940-9c899466b9ad"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:00:40 crc kubenswrapper[4802]: I1206 04:00:40.653529 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e13eaebb-eae6-4924-8940-9c899466b9ad-scripts" (OuterVolumeSpecName: "scripts") pod "e13eaebb-eae6-4924-8940-9c899466b9ad" (UID: "e13eaebb-eae6-4924-8940-9c899466b9ad"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:00:40 crc kubenswrapper[4802]: I1206 04:00:40.654727 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e13eaebb-eae6-4924-8940-9c899466b9ad-kube-api-access-q9nzw" (OuterVolumeSpecName: "kube-api-access-q9nzw") pod "e13eaebb-eae6-4924-8940-9c899466b9ad" (UID: "e13eaebb-eae6-4924-8940-9c899466b9ad"). InnerVolumeSpecName "kube-api-access-q9nzw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:00:40 crc kubenswrapper[4802]: I1206 04:00:40.680825 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e13eaebb-eae6-4924-8940-9c899466b9ad-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "e13eaebb-eae6-4924-8940-9c899466b9ad" (UID: "e13eaebb-eae6-4924-8940-9c899466b9ad"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:00:40 crc kubenswrapper[4802]: I1206 04:00:40.731632 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e13eaebb-eae6-4924-8940-9c899466b9ad-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e13eaebb-eae6-4924-8940-9c899466b9ad" (UID: "e13eaebb-eae6-4924-8940-9c899466b9ad"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:00:40 crc kubenswrapper[4802]: I1206 04:00:40.753056 4802 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e13eaebb-eae6-4924-8940-9c899466b9ad-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:40 crc kubenswrapper[4802]: I1206 04:00:40.753317 4802 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e13eaebb-eae6-4924-8940-9c899466b9ad-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:40 crc kubenswrapper[4802]: I1206 04:00:40.753403 4802 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e13eaebb-eae6-4924-8940-9c899466b9ad-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:40 crc kubenswrapper[4802]: I1206 04:00:40.753479 4802 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e13eaebb-eae6-4924-8940-9c899466b9ad-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:40 crc kubenswrapper[4802]: I1206 04:00:40.753560 4802 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e13eaebb-eae6-4924-8940-9c899466b9ad-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:40 crc kubenswrapper[4802]: I1206 04:00:40.753631 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q9nzw\" (UniqueName: \"kubernetes.io/projected/e13eaebb-eae6-4924-8940-9c899466b9ad-kube-api-access-q9nzw\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:40 crc kubenswrapper[4802]: I1206 04:00:40.772446 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e13eaebb-eae6-4924-8940-9c899466b9ad-config-data" (OuterVolumeSpecName: "config-data") pod "e13eaebb-eae6-4924-8940-9c899466b9ad" (UID: "e13eaebb-eae6-4924-8940-9c899466b9ad"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:00:40 crc kubenswrapper[4802]: I1206 04:00:40.855046 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e13eaebb-eae6-4924-8940-9c899466b9ad-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.063086 4802 generic.go:334] "Generic (PLEG): container finished" podID="e13eaebb-eae6-4924-8940-9c899466b9ad" containerID="50661514a1b6227caec3ab91719f2ccb04df0f94cd7d1c533d692cbcfdf293ce" exitCode=0 Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.063144 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e13eaebb-eae6-4924-8940-9c899466b9ad","Type":"ContainerDied","Data":"50661514a1b6227caec3ab91719f2ccb04df0f94cd7d1c533d692cbcfdf293ce"} Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.063170 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e13eaebb-eae6-4924-8940-9c899466b9ad","Type":"ContainerDied","Data":"971d5774565cea6f18af2a6d15a3735192ad2bf168ed1e3ad329eee93cbe4282"} Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.063185 4802 scope.go:117] "RemoveContainer" containerID="5baeab8eb36579d18ecafdf8bb17f68bfb647d42a04756ae8b5a6a82f64a5610" Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.063294 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.072778 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-5d96658447-rg8zs" event={"ID":"d3da4180-e8fe-4785-8d4f-6cd172fb6399","Type":"ContainerDied","Data":"663395a8db84a3f93d8c284a83195cda768663bccaaf7e4a1ab8ce35a11fcb70"} Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.072879 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-5d96658447-rg8zs" Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.102956 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.108092 4802 scope.go:117] "RemoveContainer" containerID="51346b8336d0a5d2e1ba8efeca7ef377dc0eef4d543eb15a02c6bc246a78aad3" Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.120627 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.138291 4802 scope.go:117] "RemoveContainer" containerID="50661514a1b6227caec3ab91719f2ccb04df0f94cd7d1c533d692cbcfdf293ce" Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.165611 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-5d96658447-rg8zs"] Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.180186 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-engine-5d96658447-rg8zs"] Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.184214 4802 scope.go:117] "RemoveContainer" containerID="274c790f2eed55f67a6fccc26373e759bcd33643ca3200f9e001c6b44a32e349" Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.191198 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 06 04:00:41 crc kubenswrapper[4802]: E1206 04:00:41.191695 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e13eaebb-eae6-4924-8940-9c899466b9ad" containerName="proxy-httpd" Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.191787 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="e13eaebb-eae6-4924-8940-9c899466b9ad" containerName="proxy-httpd" Dec 06 04:00:41 crc kubenswrapper[4802]: E1206 04:00:41.191856 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e13eaebb-eae6-4924-8940-9c899466b9ad" containerName="sg-core" Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.191915 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="e13eaebb-eae6-4924-8940-9c899466b9ad" containerName="sg-core" Dec 06 04:00:41 crc kubenswrapper[4802]: E1206 04:00:41.191981 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e13eaebb-eae6-4924-8940-9c899466b9ad" containerName="ceilometer-central-agent" Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.192044 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="e13eaebb-eae6-4924-8940-9c899466b9ad" containerName="ceilometer-central-agent" Dec 06 04:00:41 crc kubenswrapper[4802]: E1206 04:00:41.192105 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e13eaebb-eae6-4924-8940-9c899466b9ad" containerName="ceilometer-notification-agent" Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.192161 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="e13eaebb-eae6-4924-8940-9c899466b9ad" containerName="ceilometer-notification-agent" Dec 06 04:00:41 crc kubenswrapper[4802]: E1206 04:00:41.192219 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3da4180-e8fe-4785-8d4f-6cd172fb6399" containerName="heat-engine" Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.192269 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3da4180-e8fe-4785-8d4f-6cd172fb6399" containerName="heat-engine" Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.192496 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="e13eaebb-eae6-4924-8940-9c899466b9ad" containerName="sg-core" Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.192561 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="e13eaebb-eae6-4924-8940-9c899466b9ad" containerName="ceilometer-central-agent" Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.192629 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3da4180-e8fe-4785-8d4f-6cd172fb6399" containerName="heat-engine" Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.192688 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="e13eaebb-eae6-4924-8940-9c899466b9ad" containerName="ceilometer-notification-agent" Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.192781 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="e13eaebb-eae6-4924-8940-9c899466b9ad" containerName="proxy-httpd" Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.194618 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.198315 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.198667 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.214477 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.225955 4802 scope.go:117] "RemoveContainer" containerID="5baeab8eb36579d18ecafdf8bb17f68bfb647d42a04756ae8b5a6a82f64a5610" Dec 06 04:00:41 crc kubenswrapper[4802]: E1206 04:00:41.226395 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5baeab8eb36579d18ecafdf8bb17f68bfb647d42a04756ae8b5a6a82f64a5610\": container with ID starting with 5baeab8eb36579d18ecafdf8bb17f68bfb647d42a04756ae8b5a6a82f64a5610 not found: ID does not exist" containerID="5baeab8eb36579d18ecafdf8bb17f68bfb647d42a04756ae8b5a6a82f64a5610" Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.226439 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5baeab8eb36579d18ecafdf8bb17f68bfb647d42a04756ae8b5a6a82f64a5610"} err="failed to get container status \"5baeab8eb36579d18ecafdf8bb17f68bfb647d42a04756ae8b5a6a82f64a5610\": rpc error: code = NotFound desc = could not find container \"5baeab8eb36579d18ecafdf8bb17f68bfb647d42a04756ae8b5a6a82f64a5610\": container with ID starting with 5baeab8eb36579d18ecafdf8bb17f68bfb647d42a04756ae8b5a6a82f64a5610 not found: ID does not exist" Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.226464 4802 scope.go:117] "RemoveContainer" containerID="51346b8336d0a5d2e1ba8efeca7ef377dc0eef4d543eb15a02c6bc246a78aad3" Dec 06 04:00:41 crc kubenswrapper[4802]: E1206 04:00:41.227124 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51346b8336d0a5d2e1ba8efeca7ef377dc0eef4d543eb15a02c6bc246a78aad3\": container with ID starting with 51346b8336d0a5d2e1ba8efeca7ef377dc0eef4d543eb15a02c6bc246a78aad3 not found: ID does not exist" containerID="51346b8336d0a5d2e1ba8efeca7ef377dc0eef4d543eb15a02c6bc246a78aad3" Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.227229 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51346b8336d0a5d2e1ba8efeca7ef377dc0eef4d543eb15a02c6bc246a78aad3"} err="failed to get container status \"51346b8336d0a5d2e1ba8efeca7ef377dc0eef4d543eb15a02c6bc246a78aad3\": rpc error: code = NotFound desc = could not find container \"51346b8336d0a5d2e1ba8efeca7ef377dc0eef4d543eb15a02c6bc246a78aad3\": container with ID starting with 51346b8336d0a5d2e1ba8efeca7ef377dc0eef4d543eb15a02c6bc246a78aad3 not found: ID does not exist" Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.227317 4802 scope.go:117] "RemoveContainer" containerID="50661514a1b6227caec3ab91719f2ccb04df0f94cd7d1c533d692cbcfdf293ce" Dec 06 04:00:41 crc kubenswrapper[4802]: E1206 04:00:41.228812 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50661514a1b6227caec3ab91719f2ccb04df0f94cd7d1c533d692cbcfdf293ce\": container with ID starting with 50661514a1b6227caec3ab91719f2ccb04df0f94cd7d1c533d692cbcfdf293ce not found: ID does not exist" containerID="50661514a1b6227caec3ab91719f2ccb04df0f94cd7d1c533d692cbcfdf293ce" Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.228842 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50661514a1b6227caec3ab91719f2ccb04df0f94cd7d1c533d692cbcfdf293ce"} err="failed to get container status \"50661514a1b6227caec3ab91719f2ccb04df0f94cd7d1c533d692cbcfdf293ce\": rpc error: code = NotFound desc = could not find container \"50661514a1b6227caec3ab91719f2ccb04df0f94cd7d1c533d692cbcfdf293ce\": container with ID starting with 50661514a1b6227caec3ab91719f2ccb04df0f94cd7d1c533d692cbcfdf293ce not found: ID does not exist" Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.228861 4802 scope.go:117] "RemoveContainer" containerID="274c790f2eed55f67a6fccc26373e759bcd33643ca3200f9e001c6b44a32e349" Dec 06 04:00:41 crc kubenswrapper[4802]: E1206 04:00:41.229214 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"274c790f2eed55f67a6fccc26373e759bcd33643ca3200f9e001c6b44a32e349\": container with ID starting with 274c790f2eed55f67a6fccc26373e759bcd33643ca3200f9e001c6b44a32e349 not found: ID does not exist" containerID="274c790f2eed55f67a6fccc26373e759bcd33643ca3200f9e001c6b44a32e349" Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.229247 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"274c790f2eed55f67a6fccc26373e759bcd33643ca3200f9e001c6b44a32e349"} err="failed to get container status \"274c790f2eed55f67a6fccc26373e759bcd33643ca3200f9e001c6b44a32e349\": rpc error: code = NotFound desc = could not find container \"274c790f2eed55f67a6fccc26373e759bcd33643ca3200f9e001c6b44a32e349\": container with ID starting with 274c790f2eed55f67a6fccc26373e759bcd33643ca3200f9e001c6b44a32e349 not found: ID does not exist" Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.229269 4802 scope.go:117] "RemoveContainer" containerID="15461749ec2cccc9a0f2a06a22252592288f81b9b432dde2bc6f7f9917e9bb8a" Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.263611 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a128d3aa-1f44-4a6d-92ca-df51bc6cd097-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a128d3aa-1f44-4a6d-92ca-df51bc6cd097\") " pod="openstack/ceilometer-0" Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.263699 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zs9dr\" (UniqueName: \"kubernetes.io/projected/a128d3aa-1f44-4a6d-92ca-df51bc6cd097-kube-api-access-zs9dr\") pod \"ceilometer-0\" (UID: \"a128d3aa-1f44-4a6d-92ca-df51bc6cd097\") " pod="openstack/ceilometer-0" Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.263731 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a128d3aa-1f44-4a6d-92ca-df51bc6cd097-scripts\") pod \"ceilometer-0\" (UID: \"a128d3aa-1f44-4a6d-92ca-df51bc6cd097\") " pod="openstack/ceilometer-0" Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.263794 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a128d3aa-1f44-4a6d-92ca-df51bc6cd097-log-httpd\") pod \"ceilometer-0\" (UID: \"a128d3aa-1f44-4a6d-92ca-df51bc6cd097\") " pod="openstack/ceilometer-0" Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.263813 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a128d3aa-1f44-4a6d-92ca-df51bc6cd097-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a128d3aa-1f44-4a6d-92ca-df51bc6cd097\") " pod="openstack/ceilometer-0" Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.263850 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a128d3aa-1f44-4a6d-92ca-df51bc6cd097-run-httpd\") pod \"ceilometer-0\" (UID: \"a128d3aa-1f44-4a6d-92ca-df51bc6cd097\") " pod="openstack/ceilometer-0" Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.263898 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a128d3aa-1f44-4a6d-92ca-df51bc6cd097-config-data\") pod \"ceilometer-0\" (UID: \"a128d3aa-1f44-4a6d-92ca-df51bc6cd097\") " pod="openstack/ceilometer-0" Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.365206 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a128d3aa-1f44-4a6d-92ca-df51bc6cd097-config-data\") pod \"ceilometer-0\" (UID: \"a128d3aa-1f44-4a6d-92ca-df51bc6cd097\") " pod="openstack/ceilometer-0" Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.365311 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a128d3aa-1f44-4a6d-92ca-df51bc6cd097-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a128d3aa-1f44-4a6d-92ca-df51bc6cd097\") " pod="openstack/ceilometer-0" Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.365365 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zs9dr\" (UniqueName: \"kubernetes.io/projected/a128d3aa-1f44-4a6d-92ca-df51bc6cd097-kube-api-access-zs9dr\") pod \"ceilometer-0\" (UID: \"a128d3aa-1f44-4a6d-92ca-df51bc6cd097\") " pod="openstack/ceilometer-0" Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.365399 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a128d3aa-1f44-4a6d-92ca-df51bc6cd097-scripts\") pod \"ceilometer-0\" (UID: \"a128d3aa-1f44-4a6d-92ca-df51bc6cd097\") " pod="openstack/ceilometer-0" Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.365443 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a128d3aa-1f44-4a6d-92ca-df51bc6cd097-log-httpd\") pod \"ceilometer-0\" (UID: \"a128d3aa-1f44-4a6d-92ca-df51bc6cd097\") " pod="openstack/ceilometer-0" Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.365464 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a128d3aa-1f44-4a6d-92ca-df51bc6cd097-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a128d3aa-1f44-4a6d-92ca-df51bc6cd097\") " pod="openstack/ceilometer-0" Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.365487 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a128d3aa-1f44-4a6d-92ca-df51bc6cd097-run-httpd\") pod \"ceilometer-0\" (UID: \"a128d3aa-1f44-4a6d-92ca-df51bc6cd097\") " pod="openstack/ceilometer-0" Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.365983 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a128d3aa-1f44-4a6d-92ca-df51bc6cd097-run-httpd\") pod \"ceilometer-0\" (UID: \"a128d3aa-1f44-4a6d-92ca-df51bc6cd097\") " pod="openstack/ceilometer-0" Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.366887 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a128d3aa-1f44-4a6d-92ca-df51bc6cd097-log-httpd\") pod \"ceilometer-0\" (UID: \"a128d3aa-1f44-4a6d-92ca-df51bc6cd097\") " pod="openstack/ceilometer-0" Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.370391 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a128d3aa-1f44-4a6d-92ca-df51bc6cd097-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a128d3aa-1f44-4a6d-92ca-df51bc6cd097\") " pod="openstack/ceilometer-0" Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.370669 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a128d3aa-1f44-4a6d-92ca-df51bc6cd097-config-data\") pod \"ceilometer-0\" (UID: \"a128d3aa-1f44-4a6d-92ca-df51bc6cd097\") " pod="openstack/ceilometer-0" Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.371285 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a128d3aa-1f44-4a6d-92ca-df51bc6cd097-scripts\") pod \"ceilometer-0\" (UID: \"a128d3aa-1f44-4a6d-92ca-df51bc6cd097\") " pod="openstack/ceilometer-0" Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.377269 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a128d3aa-1f44-4a6d-92ca-df51bc6cd097-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a128d3aa-1f44-4a6d-92ca-df51bc6cd097\") " pod="openstack/ceilometer-0" Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.390609 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zs9dr\" (UniqueName: \"kubernetes.io/projected/a128d3aa-1f44-4a6d-92ca-df51bc6cd097-kube-api-access-zs9dr\") pod \"ceilometer-0\" (UID: \"a128d3aa-1f44-4a6d-92ca-df51bc6cd097\") " pod="openstack/ceilometer-0" Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.470340 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3da4180-e8fe-4785-8d4f-6cd172fb6399" path="/var/lib/kubelet/pods/d3da4180-e8fe-4785-8d4f-6cd172fb6399/volumes" Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.473457 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e13eaebb-eae6-4924-8940-9c899466b9ad" path="/var/lib/kubelet/pods/e13eaebb-eae6-4924-8940-9c899466b9ad/volumes" Dec 06 04:00:41 crc kubenswrapper[4802]: I1206 04:00:41.517282 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 04:00:47 crc kubenswrapper[4802]: W1206 04:00:47.363928 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda128d3aa_1f44_4a6d_92ca_df51bc6cd097.slice/crio-08bc58a7edb9579326afd1934b352067c3c90faaa613cea99c0319fab87f568a WatchSource:0}: Error finding container 08bc58a7edb9579326afd1934b352067c3c90faaa613cea99c0319fab87f568a: Status 404 returned error can't find the container with id 08bc58a7edb9579326afd1934b352067c3c90faaa613cea99c0319fab87f568a Dec 06 04:00:47 crc kubenswrapper[4802]: I1206 04:00:47.365562 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 04:00:47 crc kubenswrapper[4802]: I1206 04:00:47.698254 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 04:00:48 crc kubenswrapper[4802]: I1206 04:00:48.137983 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a128d3aa-1f44-4a6d-92ca-df51bc6cd097","Type":"ContainerStarted","Data":"cb11fba1b171ed3e1a85133de81039e2e67c9cff821abaecbe190a62b89fa7c2"} Dec 06 04:00:48 crc kubenswrapper[4802]: I1206 04:00:48.138472 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a128d3aa-1f44-4a6d-92ca-df51bc6cd097","Type":"ContainerStarted","Data":"08bc58a7edb9579326afd1934b352067c3c90faaa613cea99c0319fab87f568a"} Dec 06 04:00:48 crc kubenswrapper[4802]: I1206 04:00:48.140228 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-xnnxk" event={"ID":"f09f7e4f-6e98-4461-bf3a-1a6aa9528c09","Type":"ContainerStarted","Data":"e646cfb755995e7f7666c31c64b92161aece638ad58007ab7195f0fdb2a45316"} Dec 06 04:00:48 crc kubenswrapper[4802]: I1206 04:00:48.163505 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-xnnxk" podStartSLOduration=2.8079529389999998 podStartE2EDuration="12.163485219s" podCreationTimestamp="2025-12-06 04:00:36 +0000 UTC" firstStartedPulling="2025-12-06 04:00:37.585743909 +0000 UTC m=+1230.457653061" lastFinishedPulling="2025-12-06 04:00:46.941276189 +0000 UTC m=+1239.813185341" observedRunningTime="2025-12-06 04:00:48.158970115 +0000 UTC m=+1241.030879277" watchObservedRunningTime="2025-12-06 04:00:48.163485219 +0000 UTC m=+1241.035394371" Dec 06 04:00:49 crc kubenswrapper[4802]: I1206 04:00:49.152633 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a128d3aa-1f44-4a6d-92ca-df51bc6cd097","Type":"ContainerStarted","Data":"3ab1775540cb21c804ef34451616e8edc0e7a1f035e391f33df05e801c645e4d"} Dec 06 04:00:50 crc kubenswrapper[4802]: I1206 04:00:50.161262 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a128d3aa-1f44-4a6d-92ca-df51bc6cd097","Type":"ContainerStarted","Data":"75b98ba7c36b2ecb2ceea7d9979de123c001191248f80acbc51471e2e6cc98d8"} Dec 06 04:00:51 crc kubenswrapper[4802]: I1206 04:00:51.172813 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a128d3aa-1f44-4a6d-92ca-df51bc6cd097","Type":"ContainerStarted","Data":"d9b46c170bb4e57da13c1765cb60284b4284b5215be0ba5b64dcc3e4557574eb"} Dec 06 04:00:51 crc kubenswrapper[4802]: I1206 04:00:51.173077 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 06 04:00:51 crc kubenswrapper[4802]: I1206 04:00:51.172978 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a128d3aa-1f44-4a6d-92ca-df51bc6cd097" containerName="sg-core" containerID="cri-o://75b98ba7c36b2ecb2ceea7d9979de123c001191248f80acbc51471e2e6cc98d8" gracePeriod=30 Dec 06 04:00:51 crc kubenswrapper[4802]: I1206 04:00:51.172954 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a128d3aa-1f44-4a6d-92ca-df51bc6cd097" containerName="ceilometer-central-agent" containerID="cri-o://cb11fba1b171ed3e1a85133de81039e2e67c9cff821abaecbe190a62b89fa7c2" gracePeriod=30 Dec 06 04:00:51 crc kubenswrapper[4802]: I1206 04:00:51.172993 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a128d3aa-1f44-4a6d-92ca-df51bc6cd097" containerName="ceilometer-notification-agent" containerID="cri-o://3ab1775540cb21c804ef34451616e8edc0e7a1f035e391f33df05e801c645e4d" gracePeriod=30 Dec 06 04:00:51 crc kubenswrapper[4802]: I1206 04:00:51.172997 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a128d3aa-1f44-4a6d-92ca-df51bc6cd097" containerName="proxy-httpd" containerID="cri-o://d9b46c170bb4e57da13c1765cb60284b4284b5215be0ba5b64dcc3e4557574eb" gracePeriod=30 Dec 06 04:00:51 crc kubenswrapper[4802]: I1206 04:00:51.204407 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=6.926603565 podStartE2EDuration="10.204391619s" podCreationTimestamp="2025-12-06 04:00:41 +0000 UTC" firstStartedPulling="2025-12-06 04:00:47.373685867 +0000 UTC m=+1240.245595009" lastFinishedPulling="2025-12-06 04:00:50.651473911 +0000 UTC m=+1243.523383063" observedRunningTime="2025-12-06 04:00:51.1949628 +0000 UTC m=+1244.066871952" watchObservedRunningTime="2025-12-06 04:00:51.204391619 +0000 UTC m=+1244.076300771" Dec 06 04:00:52 crc kubenswrapper[4802]: I1206 04:00:52.183592 4802 generic.go:334] "Generic (PLEG): container finished" podID="a128d3aa-1f44-4a6d-92ca-df51bc6cd097" containerID="d9b46c170bb4e57da13c1765cb60284b4284b5215be0ba5b64dcc3e4557574eb" exitCode=0 Dec 06 04:00:52 crc kubenswrapper[4802]: I1206 04:00:52.183862 4802 generic.go:334] "Generic (PLEG): container finished" podID="a128d3aa-1f44-4a6d-92ca-df51bc6cd097" containerID="75b98ba7c36b2ecb2ceea7d9979de123c001191248f80acbc51471e2e6cc98d8" exitCode=2 Dec 06 04:00:52 crc kubenswrapper[4802]: I1206 04:00:52.183871 4802 generic.go:334] "Generic (PLEG): container finished" podID="a128d3aa-1f44-4a6d-92ca-df51bc6cd097" containerID="3ab1775540cb21c804ef34451616e8edc0e7a1f035e391f33df05e801c645e4d" exitCode=0 Dec 06 04:00:52 crc kubenswrapper[4802]: I1206 04:00:52.183682 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a128d3aa-1f44-4a6d-92ca-df51bc6cd097","Type":"ContainerDied","Data":"d9b46c170bb4e57da13c1765cb60284b4284b5215be0ba5b64dcc3e4557574eb"} Dec 06 04:00:52 crc kubenswrapper[4802]: I1206 04:00:52.183913 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a128d3aa-1f44-4a6d-92ca-df51bc6cd097","Type":"ContainerDied","Data":"75b98ba7c36b2ecb2ceea7d9979de123c001191248f80acbc51471e2e6cc98d8"} Dec 06 04:00:52 crc kubenswrapper[4802]: I1206 04:00:52.183932 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a128d3aa-1f44-4a6d-92ca-df51bc6cd097","Type":"ContainerDied","Data":"3ab1775540cb21c804ef34451616e8edc0e7a1f035e391f33df05e801c645e4d"} Dec 06 04:00:59 crc kubenswrapper[4802]: I1206 04:00:59.275933 4802 generic.go:334] "Generic (PLEG): container finished" podID="a128d3aa-1f44-4a6d-92ca-df51bc6cd097" containerID="cb11fba1b171ed3e1a85133de81039e2e67c9cff821abaecbe190a62b89fa7c2" exitCode=0 Dec 06 04:00:59 crc kubenswrapper[4802]: I1206 04:00:59.275997 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a128d3aa-1f44-4a6d-92ca-df51bc6cd097","Type":"ContainerDied","Data":"cb11fba1b171ed3e1a85133de81039e2e67c9cff821abaecbe190a62b89fa7c2"} Dec 06 04:00:59 crc kubenswrapper[4802]: I1206 04:00:59.278163 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a128d3aa-1f44-4a6d-92ca-df51bc6cd097","Type":"ContainerDied","Data":"08bc58a7edb9579326afd1934b352067c3c90faaa613cea99c0319fab87f568a"} Dec 06 04:00:59 crc kubenswrapper[4802]: I1206 04:00:59.278287 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="08bc58a7edb9579326afd1934b352067c3c90faaa613cea99c0319fab87f568a" Dec 06 04:00:59 crc kubenswrapper[4802]: I1206 04:00:59.314177 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 04:00:59 crc kubenswrapper[4802]: I1206 04:00:59.362337 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zs9dr\" (UniqueName: \"kubernetes.io/projected/a128d3aa-1f44-4a6d-92ca-df51bc6cd097-kube-api-access-zs9dr\") pod \"a128d3aa-1f44-4a6d-92ca-df51bc6cd097\" (UID: \"a128d3aa-1f44-4a6d-92ca-df51bc6cd097\") " Dec 06 04:00:59 crc kubenswrapper[4802]: I1206 04:00:59.362430 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a128d3aa-1f44-4a6d-92ca-df51bc6cd097-log-httpd\") pod \"a128d3aa-1f44-4a6d-92ca-df51bc6cd097\" (UID: \"a128d3aa-1f44-4a6d-92ca-df51bc6cd097\") " Dec 06 04:00:59 crc kubenswrapper[4802]: I1206 04:00:59.362452 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a128d3aa-1f44-4a6d-92ca-df51bc6cd097-scripts\") pod \"a128d3aa-1f44-4a6d-92ca-df51bc6cd097\" (UID: \"a128d3aa-1f44-4a6d-92ca-df51bc6cd097\") " Dec 06 04:00:59 crc kubenswrapper[4802]: I1206 04:00:59.362470 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a128d3aa-1f44-4a6d-92ca-df51bc6cd097-run-httpd\") pod \"a128d3aa-1f44-4a6d-92ca-df51bc6cd097\" (UID: \"a128d3aa-1f44-4a6d-92ca-df51bc6cd097\") " Dec 06 04:00:59 crc kubenswrapper[4802]: I1206 04:00:59.362488 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a128d3aa-1f44-4a6d-92ca-df51bc6cd097-sg-core-conf-yaml\") pod \"a128d3aa-1f44-4a6d-92ca-df51bc6cd097\" (UID: \"a128d3aa-1f44-4a6d-92ca-df51bc6cd097\") " Dec 06 04:00:59 crc kubenswrapper[4802]: I1206 04:00:59.362523 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a128d3aa-1f44-4a6d-92ca-df51bc6cd097-config-data\") pod \"a128d3aa-1f44-4a6d-92ca-df51bc6cd097\" (UID: \"a128d3aa-1f44-4a6d-92ca-df51bc6cd097\") " Dec 06 04:00:59 crc kubenswrapper[4802]: I1206 04:00:59.362548 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a128d3aa-1f44-4a6d-92ca-df51bc6cd097-combined-ca-bundle\") pod \"a128d3aa-1f44-4a6d-92ca-df51bc6cd097\" (UID: \"a128d3aa-1f44-4a6d-92ca-df51bc6cd097\") " Dec 06 04:00:59 crc kubenswrapper[4802]: I1206 04:00:59.363042 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a128d3aa-1f44-4a6d-92ca-df51bc6cd097-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a128d3aa-1f44-4a6d-92ca-df51bc6cd097" (UID: "a128d3aa-1f44-4a6d-92ca-df51bc6cd097"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:00:59 crc kubenswrapper[4802]: I1206 04:00:59.363168 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a128d3aa-1f44-4a6d-92ca-df51bc6cd097-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a128d3aa-1f44-4a6d-92ca-df51bc6cd097" (UID: "a128d3aa-1f44-4a6d-92ca-df51bc6cd097"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:00:59 crc kubenswrapper[4802]: I1206 04:00:59.368859 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a128d3aa-1f44-4a6d-92ca-df51bc6cd097-kube-api-access-zs9dr" (OuterVolumeSpecName: "kube-api-access-zs9dr") pod "a128d3aa-1f44-4a6d-92ca-df51bc6cd097" (UID: "a128d3aa-1f44-4a6d-92ca-df51bc6cd097"). InnerVolumeSpecName "kube-api-access-zs9dr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:00:59 crc kubenswrapper[4802]: I1206 04:00:59.384003 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a128d3aa-1f44-4a6d-92ca-df51bc6cd097-scripts" (OuterVolumeSpecName: "scripts") pod "a128d3aa-1f44-4a6d-92ca-df51bc6cd097" (UID: "a128d3aa-1f44-4a6d-92ca-df51bc6cd097"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:00:59 crc kubenswrapper[4802]: I1206 04:00:59.409173 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a128d3aa-1f44-4a6d-92ca-df51bc6cd097-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a128d3aa-1f44-4a6d-92ca-df51bc6cd097" (UID: "a128d3aa-1f44-4a6d-92ca-df51bc6cd097"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:00:59 crc kubenswrapper[4802]: I1206 04:00:59.450155 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a128d3aa-1f44-4a6d-92ca-df51bc6cd097-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a128d3aa-1f44-4a6d-92ca-df51bc6cd097" (UID: "a128d3aa-1f44-4a6d-92ca-df51bc6cd097"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:00:59 crc kubenswrapper[4802]: I1206 04:00:59.464232 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zs9dr\" (UniqueName: \"kubernetes.io/projected/a128d3aa-1f44-4a6d-92ca-df51bc6cd097-kube-api-access-zs9dr\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:59 crc kubenswrapper[4802]: I1206 04:00:59.464696 4802 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a128d3aa-1f44-4a6d-92ca-df51bc6cd097-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:59 crc kubenswrapper[4802]: I1206 04:00:59.464743 4802 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a128d3aa-1f44-4a6d-92ca-df51bc6cd097-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:59 crc kubenswrapper[4802]: I1206 04:00:59.464781 4802 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a128d3aa-1f44-4a6d-92ca-df51bc6cd097-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:59 crc kubenswrapper[4802]: I1206 04:00:59.464797 4802 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a128d3aa-1f44-4a6d-92ca-df51bc6cd097-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:59 crc kubenswrapper[4802]: I1206 04:00:59.464817 4802 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a128d3aa-1f44-4a6d-92ca-df51bc6cd097-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:59 crc kubenswrapper[4802]: I1206 04:00:59.494715 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a128d3aa-1f44-4a6d-92ca-df51bc6cd097-config-data" (OuterVolumeSpecName: "config-data") pod "a128d3aa-1f44-4a6d-92ca-df51bc6cd097" (UID: "a128d3aa-1f44-4a6d-92ca-df51bc6cd097"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:00:59 crc kubenswrapper[4802]: I1206 04:00:59.566032 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a128d3aa-1f44-4a6d-92ca-df51bc6cd097-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:00 crc kubenswrapper[4802]: I1206 04:01:00.169740 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29416561-rnfxm"] Dec 06 04:01:00 crc kubenswrapper[4802]: E1206 04:01:00.170946 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a128d3aa-1f44-4a6d-92ca-df51bc6cd097" containerName="ceilometer-central-agent" Dec 06 04:01:00 crc kubenswrapper[4802]: I1206 04:01:00.170981 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="a128d3aa-1f44-4a6d-92ca-df51bc6cd097" containerName="ceilometer-central-agent" Dec 06 04:01:00 crc kubenswrapper[4802]: E1206 04:01:00.171036 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a128d3aa-1f44-4a6d-92ca-df51bc6cd097" containerName="sg-core" Dec 06 04:01:00 crc kubenswrapper[4802]: I1206 04:01:00.171056 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="a128d3aa-1f44-4a6d-92ca-df51bc6cd097" containerName="sg-core" Dec 06 04:01:00 crc kubenswrapper[4802]: E1206 04:01:00.171091 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a128d3aa-1f44-4a6d-92ca-df51bc6cd097" containerName="ceilometer-notification-agent" Dec 06 04:01:00 crc kubenswrapper[4802]: I1206 04:01:00.171108 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="a128d3aa-1f44-4a6d-92ca-df51bc6cd097" containerName="ceilometer-notification-agent" Dec 06 04:01:00 crc kubenswrapper[4802]: E1206 04:01:00.171159 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a128d3aa-1f44-4a6d-92ca-df51bc6cd097" containerName="proxy-httpd" Dec 06 04:01:00 crc kubenswrapper[4802]: I1206 04:01:00.171173 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="a128d3aa-1f44-4a6d-92ca-df51bc6cd097" containerName="proxy-httpd" Dec 06 04:01:00 crc kubenswrapper[4802]: I1206 04:01:00.171571 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="a128d3aa-1f44-4a6d-92ca-df51bc6cd097" containerName="ceilometer-central-agent" Dec 06 04:01:00 crc kubenswrapper[4802]: I1206 04:01:00.171620 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="a128d3aa-1f44-4a6d-92ca-df51bc6cd097" containerName="ceilometer-notification-agent" Dec 06 04:01:00 crc kubenswrapper[4802]: I1206 04:01:00.171645 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="a128d3aa-1f44-4a6d-92ca-df51bc6cd097" containerName="proxy-httpd" Dec 06 04:01:00 crc kubenswrapper[4802]: I1206 04:01:00.171665 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="a128d3aa-1f44-4a6d-92ca-df51bc6cd097" containerName="sg-core" Dec 06 04:01:00 crc kubenswrapper[4802]: I1206 04:01:00.172797 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29416561-rnfxm" Dec 06 04:01:00 crc kubenswrapper[4802]: I1206 04:01:00.176299 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/613478db-c10d-476b-b66d-4ae3572067ff-fernet-keys\") pod \"keystone-cron-29416561-rnfxm\" (UID: \"613478db-c10d-476b-b66d-4ae3572067ff\") " pod="openstack/keystone-cron-29416561-rnfxm" Dec 06 04:01:00 crc kubenswrapper[4802]: I1206 04:01:00.176499 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/613478db-c10d-476b-b66d-4ae3572067ff-combined-ca-bundle\") pod \"keystone-cron-29416561-rnfxm\" (UID: \"613478db-c10d-476b-b66d-4ae3572067ff\") " pod="openstack/keystone-cron-29416561-rnfxm" Dec 06 04:01:00 crc kubenswrapper[4802]: I1206 04:01:00.176650 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/613478db-c10d-476b-b66d-4ae3572067ff-config-data\") pod \"keystone-cron-29416561-rnfxm\" (UID: \"613478db-c10d-476b-b66d-4ae3572067ff\") " pod="openstack/keystone-cron-29416561-rnfxm" Dec 06 04:01:00 crc kubenswrapper[4802]: I1206 04:01:00.176808 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74skm\" (UniqueName: \"kubernetes.io/projected/613478db-c10d-476b-b66d-4ae3572067ff-kube-api-access-74skm\") pod \"keystone-cron-29416561-rnfxm\" (UID: \"613478db-c10d-476b-b66d-4ae3572067ff\") " pod="openstack/keystone-cron-29416561-rnfxm" Dec 06 04:01:00 crc kubenswrapper[4802]: I1206 04:01:00.183980 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29416561-rnfxm"] Dec 06 04:01:00 crc kubenswrapper[4802]: I1206 04:01:00.277530 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/613478db-c10d-476b-b66d-4ae3572067ff-combined-ca-bundle\") pod \"keystone-cron-29416561-rnfxm\" (UID: \"613478db-c10d-476b-b66d-4ae3572067ff\") " pod="openstack/keystone-cron-29416561-rnfxm" Dec 06 04:01:00 crc kubenswrapper[4802]: I1206 04:01:00.277577 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/613478db-c10d-476b-b66d-4ae3572067ff-config-data\") pod \"keystone-cron-29416561-rnfxm\" (UID: \"613478db-c10d-476b-b66d-4ae3572067ff\") " pod="openstack/keystone-cron-29416561-rnfxm" Dec 06 04:01:00 crc kubenswrapper[4802]: I1206 04:01:00.277607 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74skm\" (UniqueName: \"kubernetes.io/projected/613478db-c10d-476b-b66d-4ae3572067ff-kube-api-access-74skm\") pod \"keystone-cron-29416561-rnfxm\" (UID: \"613478db-c10d-476b-b66d-4ae3572067ff\") " pod="openstack/keystone-cron-29416561-rnfxm" Dec 06 04:01:00 crc kubenswrapper[4802]: I1206 04:01:00.277701 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/613478db-c10d-476b-b66d-4ae3572067ff-fernet-keys\") pod \"keystone-cron-29416561-rnfxm\" (UID: \"613478db-c10d-476b-b66d-4ae3572067ff\") " pod="openstack/keystone-cron-29416561-rnfxm" Dec 06 04:01:00 crc kubenswrapper[4802]: I1206 04:01:00.284937 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/613478db-c10d-476b-b66d-4ae3572067ff-combined-ca-bundle\") pod \"keystone-cron-29416561-rnfxm\" (UID: \"613478db-c10d-476b-b66d-4ae3572067ff\") " pod="openstack/keystone-cron-29416561-rnfxm" Dec 06 04:01:00 crc kubenswrapper[4802]: I1206 04:01:00.284978 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/613478db-c10d-476b-b66d-4ae3572067ff-fernet-keys\") pod \"keystone-cron-29416561-rnfxm\" (UID: \"613478db-c10d-476b-b66d-4ae3572067ff\") " pod="openstack/keystone-cron-29416561-rnfxm" Dec 06 04:01:00 crc kubenswrapper[4802]: I1206 04:01:00.286368 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/613478db-c10d-476b-b66d-4ae3572067ff-config-data\") pod \"keystone-cron-29416561-rnfxm\" (UID: \"613478db-c10d-476b-b66d-4ae3572067ff\") " pod="openstack/keystone-cron-29416561-rnfxm" Dec 06 04:01:00 crc kubenswrapper[4802]: I1206 04:01:00.286450 4802 generic.go:334] "Generic (PLEG): container finished" podID="f09f7e4f-6e98-4461-bf3a-1a6aa9528c09" containerID="e646cfb755995e7f7666c31c64b92161aece638ad58007ab7195f0fdb2a45316" exitCode=0 Dec 06 04:01:00 crc kubenswrapper[4802]: I1206 04:01:00.286514 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-xnnxk" event={"ID":"f09f7e4f-6e98-4461-bf3a-1a6aa9528c09","Type":"ContainerDied","Data":"e646cfb755995e7f7666c31c64b92161aece638ad58007ab7195f0fdb2a45316"} Dec 06 04:01:00 crc kubenswrapper[4802]: I1206 04:01:00.286528 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 04:01:00 crc kubenswrapper[4802]: I1206 04:01:00.301975 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74skm\" (UniqueName: \"kubernetes.io/projected/613478db-c10d-476b-b66d-4ae3572067ff-kube-api-access-74skm\") pod \"keystone-cron-29416561-rnfxm\" (UID: \"613478db-c10d-476b-b66d-4ae3572067ff\") " pod="openstack/keystone-cron-29416561-rnfxm" Dec 06 04:01:00 crc kubenswrapper[4802]: I1206 04:01:00.402984 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 04:01:00 crc kubenswrapper[4802]: I1206 04:01:00.421517 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 06 04:01:00 crc kubenswrapper[4802]: I1206 04:01:00.432275 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 06 04:01:00 crc kubenswrapper[4802]: I1206 04:01:00.435127 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 04:01:00 crc kubenswrapper[4802]: I1206 04:01:00.438316 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 06 04:01:00 crc kubenswrapper[4802]: I1206 04:01:00.438549 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 06 04:01:00 crc kubenswrapper[4802]: I1206 04:01:00.461823 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 04:01:00 crc kubenswrapper[4802]: I1206 04:01:00.483602 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30152cc8-8968-443f-923a-0ddca8c016ee-scripts\") pod \"ceilometer-0\" (UID: \"30152cc8-8968-443f-923a-0ddca8c016ee\") " pod="openstack/ceilometer-0" Dec 06 04:01:00 crc kubenswrapper[4802]: I1206 04:01:00.483728 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/30152cc8-8968-443f-923a-0ddca8c016ee-run-httpd\") pod \"ceilometer-0\" (UID: \"30152cc8-8968-443f-923a-0ddca8c016ee\") " pod="openstack/ceilometer-0" Dec 06 04:01:00 crc kubenswrapper[4802]: I1206 04:01:00.483794 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/30152cc8-8968-443f-923a-0ddca8c016ee-log-httpd\") pod \"ceilometer-0\" (UID: \"30152cc8-8968-443f-923a-0ddca8c016ee\") " pod="openstack/ceilometer-0" Dec 06 04:01:00 crc kubenswrapper[4802]: I1206 04:01:00.484032 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30152cc8-8968-443f-923a-0ddca8c016ee-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"30152cc8-8968-443f-923a-0ddca8c016ee\") " pod="openstack/ceilometer-0" Dec 06 04:01:00 crc kubenswrapper[4802]: I1206 04:01:00.484135 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/30152cc8-8968-443f-923a-0ddca8c016ee-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"30152cc8-8968-443f-923a-0ddca8c016ee\") " pod="openstack/ceilometer-0" Dec 06 04:01:00 crc kubenswrapper[4802]: I1206 04:01:00.484214 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29vhd\" (UniqueName: \"kubernetes.io/projected/30152cc8-8968-443f-923a-0ddca8c016ee-kube-api-access-29vhd\") pod \"ceilometer-0\" (UID: \"30152cc8-8968-443f-923a-0ddca8c016ee\") " pod="openstack/ceilometer-0" Dec 06 04:01:00 crc kubenswrapper[4802]: I1206 04:01:00.484304 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30152cc8-8968-443f-923a-0ddca8c016ee-config-data\") pod \"ceilometer-0\" (UID: \"30152cc8-8968-443f-923a-0ddca8c016ee\") " pod="openstack/ceilometer-0" Dec 06 04:01:00 crc kubenswrapper[4802]: I1206 04:01:00.512052 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29416561-rnfxm" Dec 06 04:01:00 crc kubenswrapper[4802]: I1206 04:01:00.586303 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29vhd\" (UniqueName: \"kubernetes.io/projected/30152cc8-8968-443f-923a-0ddca8c016ee-kube-api-access-29vhd\") pod \"ceilometer-0\" (UID: \"30152cc8-8968-443f-923a-0ddca8c016ee\") " pod="openstack/ceilometer-0" Dec 06 04:01:00 crc kubenswrapper[4802]: I1206 04:01:00.586947 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30152cc8-8968-443f-923a-0ddca8c016ee-config-data\") pod \"ceilometer-0\" (UID: \"30152cc8-8968-443f-923a-0ddca8c016ee\") " pod="openstack/ceilometer-0" Dec 06 04:01:00 crc kubenswrapper[4802]: I1206 04:01:00.587107 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30152cc8-8968-443f-923a-0ddca8c016ee-scripts\") pod \"ceilometer-0\" (UID: \"30152cc8-8968-443f-923a-0ddca8c016ee\") " pod="openstack/ceilometer-0" Dec 06 04:01:00 crc kubenswrapper[4802]: I1206 04:01:00.587283 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/30152cc8-8968-443f-923a-0ddca8c016ee-run-httpd\") pod \"ceilometer-0\" (UID: \"30152cc8-8968-443f-923a-0ddca8c016ee\") " pod="openstack/ceilometer-0" Dec 06 04:01:00 crc kubenswrapper[4802]: I1206 04:01:00.587431 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/30152cc8-8968-443f-923a-0ddca8c016ee-log-httpd\") pod \"ceilometer-0\" (UID: \"30152cc8-8968-443f-923a-0ddca8c016ee\") " pod="openstack/ceilometer-0" Dec 06 04:01:00 crc kubenswrapper[4802]: I1206 04:01:00.587625 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30152cc8-8968-443f-923a-0ddca8c016ee-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"30152cc8-8968-443f-923a-0ddca8c016ee\") " pod="openstack/ceilometer-0" Dec 06 04:01:00 crc kubenswrapper[4802]: I1206 04:01:00.587786 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/30152cc8-8968-443f-923a-0ddca8c016ee-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"30152cc8-8968-443f-923a-0ddca8c016ee\") " pod="openstack/ceilometer-0" Dec 06 04:01:00 crc kubenswrapper[4802]: I1206 04:01:00.587901 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/30152cc8-8968-443f-923a-0ddca8c016ee-run-httpd\") pod \"ceilometer-0\" (UID: \"30152cc8-8968-443f-923a-0ddca8c016ee\") " pod="openstack/ceilometer-0" Dec 06 04:01:00 crc kubenswrapper[4802]: I1206 04:01:00.587906 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/30152cc8-8968-443f-923a-0ddca8c016ee-log-httpd\") pod \"ceilometer-0\" (UID: \"30152cc8-8968-443f-923a-0ddca8c016ee\") " pod="openstack/ceilometer-0" Dec 06 04:01:00 crc kubenswrapper[4802]: I1206 04:01:00.591622 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30152cc8-8968-443f-923a-0ddca8c016ee-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"30152cc8-8968-443f-923a-0ddca8c016ee\") " pod="openstack/ceilometer-0" Dec 06 04:01:00 crc kubenswrapper[4802]: I1206 04:01:00.592171 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30152cc8-8968-443f-923a-0ddca8c016ee-config-data\") pod \"ceilometer-0\" (UID: \"30152cc8-8968-443f-923a-0ddca8c016ee\") " pod="openstack/ceilometer-0" Dec 06 04:01:00 crc kubenswrapper[4802]: I1206 04:01:00.592479 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/30152cc8-8968-443f-923a-0ddca8c016ee-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"30152cc8-8968-443f-923a-0ddca8c016ee\") " pod="openstack/ceilometer-0" Dec 06 04:01:00 crc kubenswrapper[4802]: I1206 04:01:00.597773 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30152cc8-8968-443f-923a-0ddca8c016ee-scripts\") pod \"ceilometer-0\" (UID: \"30152cc8-8968-443f-923a-0ddca8c016ee\") " pod="openstack/ceilometer-0" Dec 06 04:01:00 crc kubenswrapper[4802]: I1206 04:01:00.602464 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29vhd\" (UniqueName: \"kubernetes.io/projected/30152cc8-8968-443f-923a-0ddca8c016ee-kube-api-access-29vhd\") pod \"ceilometer-0\" (UID: \"30152cc8-8968-443f-923a-0ddca8c016ee\") " pod="openstack/ceilometer-0" Dec 06 04:01:00 crc kubenswrapper[4802]: I1206 04:01:00.758620 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 04:01:00 crc kubenswrapper[4802]: I1206 04:01:00.984713 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29416561-rnfxm"] Dec 06 04:01:01 crc kubenswrapper[4802]: I1206 04:01:01.241393 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 04:01:01 crc kubenswrapper[4802]: I1206 04:01:01.295732 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"30152cc8-8968-443f-923a-0ddca8c016ee","Type":"ContainerStarted","Data":"2b782ebf701bf7af80fdaa69877b316efde31731a2bf48eee1e4c33c5d1a1917"} Dec 06 04:01:01 crc kubenswrapper[4802]: I1206 04:01:01.297303 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29416561-rnfxm" event={"ID":"613478db-c10d-476b-b66d-4ae3572067ff","Type":"ContainerStarted","Data":"c58f5148ad0e68a7754ea51674c34cf83f6dcdc2f9b32cb93434b60414c0d5eb"} Dec 06 04:01:01 crc kubenswrapper[4802]: I1206 04:01:01.297379 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29416561-rnfxm" event={"ID":"613478db-c10d-476b-b66d-4ae3572067ff","Type":"ContainerStarted","Data":"79714aff7c8ee9d9ea09bdfda43fcba819a5f2b0f22e9b2cc547c36818bbbcf2"} Dec 06 04:01:01 crc kubenswrapper[4802]: I1206 04:01:01.330365 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29416561-rnfxm" podStartSLOduration=1.330344342 podStartE2EDuration="1.330344342s" podCreationTimestamp="2025-12-06 04:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:01:01.325526749 +0000 UTC m=+1254.197435921" watchObservedRunningTime="2025-12-06 04:01:01.330344342 +0000 UTC m=+1254.202253494" Dec 06 04:01:01 crc kubenswrapper[4802]: I1206 04:01:01.464275 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a128d3aa-1f44-4a6d-92ca-df51bc6cd097" path="/var/lib/kubelet/pods/a128d3aa-1f44-4a6d-92ca-df51bc6cd097/volumes" Dec 06 04:01:01 crc kubenswrapper[4802]: I1206 04:01:01.579378 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-xnnxk" Dec 06 04:01:01 crc kubenswrapper[4802]: I1206 04:01:01.612977 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f09f7e4f-6e98-4461-bf3a-1a6aa9528c09-config-data\") pod \"f09f7e4f-6e98-4461-bf3a-1a6aa9528c09\" (UID: \"f09f7e4f-6e98-4461-bf3a-1a6aa9528c09\") " Dec 06 04:01:01 crc kubenswrapper[4802]: I1206 04:01:01.613096 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f09f7e4f-6e98-4461-bf3a-1a6aa9528c09-combined-ca-bundle\") pod \"f09f7e4f-6e98-4461-bf3a-1a6aa9528c09\" (UID: \"f09f7e4f-6e98-4461-bf3a-1a6aa9528c09\") " Dec 06 04:01:01 crc kubenswrapper[4802]: I1206 04:01:01.613154 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f09f7e4f-6e98-4461-bf3a-1a6aa9528c09-scripts\") pod \"f09f7e4f-6e98-4461-bf3a-1a6aa9528c09\" (UID: \"f09f7e4f-6e98-4461-bf3a-1a6aa9528c09\") " Dec 06 04:01:01 crc kubenswrapper[4802]: I1206 04:01:01.613287 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-csvhd\" (UniqueName: \"kubernetes.io/projected/f09f7e4f-6e98-4461-bf3a-1a6aa9528c09-kube-api-access-csvhd\") pod \"f09f7e4f-6e98-4461-bf3a-1a6aa9528c09\" (UID: \"f09f7e4f-6e98-4461-bf3a-1a6aa9528c09\") " Dec 06 04:01:01 crc kubenswrapper[4802]: I1206 04:01:01.617547 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f09f7e4f-6e98-4461-bf3a-1a6aa9528c09-scripts" (OuterVolumeSpecName: "scripts") pod "f09f7e4f-6e98-4461-bf3a-1a6aa9528c09" (UID: "f09f7e4f-6e98-4461-bf3a-1a6aa9528c09"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:01:01 crc kubenswrapper[4802]: I1206 04:01:01.620040 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f09f7e4f-6e98-4461-bf3a-1a6aa9528c09-kube-api-access-csvhd" (OuterVolumeSpecName: "kube-api-access-csvhd") pod "f09f7e4f-6e98-4461-bf3a-1a6aa9528c09" (UID: "f09f7e4f-6e98-4461-bf3a-1a6aa9528c09"). InnerVolumeSpecName "kube-api-access-csvhd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:01:01 crc kubenswrapper[4802]: I1206 04:01:01.641475 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f09f7e4f-6e98-4461-bf3a-1a6aa9528c09-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f09f7e4f-6e98-4461-bf3a-1a6aa9528c09" (UID: "f09f7e4f-6e98-4461-bf3a-1a6aa9528c09"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:01:01 crc kubenswrapper[4802]: I1206 04:01:01.650516 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f09f7e4f-6e98-4461-bf3a-1a6aa9528c09-config-data" (OuterVolumeSpecName: "config-data") pod "f09f7e4f-6e98-4461-bf3a-1a6aa9528c09" (UID: "f09f7e4f-6e98-4461-bf3a-1a6aa9528c09"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:01:01 crc kubenswrapper[4802]: I1206 04:01:01.720320 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f09f7e4f-6e98-4461-bf3a-1a6aa9528c09-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:01 crc kubenswrapper[4802]: I1206 04:01:01.720352 4802 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f09f7e4f-6e98-4461-bf3a-1a6aa9528c09-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:01 crc kubenswrapper[4802]: I1206 04:01:01.720418 4802 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f09f7e4f-6e98-4461-bf3a-1a6aa9528c09-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:01 crc kubenswrapper[4802]: I1206 04:01:01.720429 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-csvhd\" (UniqueName: \"kubernetes.io/projected/f09f7e4f-6e98-4461-bf3a-1a6aa9528c09-kube-api-access-csvhd\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:02 crc kubenswrapper[4802]: I1206 04:01:02.307196 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"30152cc8-8968-443f-923a-0ddca8c016ee","Type":"ContainerStarted","Data":"9419cab08a0d0f6b39eefd268cd4b3759dab87c76047fa89e3000f6949febb58"} Dec 06 04:01:02 crc kubenswrapper[4802]: I1206 04:01:02.309594 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-xnnxk" Dec 06 04:01:02 crc kubenswrapper[4802]: I1206 04:01:02.310336 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-xnnxk" event={"ID":"f09f7e4f-6e98-4461-bf3a-1a6aa9528c09","Type":"ContainerDied","Data":"f6179c8e1032314aca8ad9f81dcfb870cb9717ebf0a2fdd6cc8068e3b5b3f2dd"} Dec 06 04:01:02 crc kubenswrapper[4802]: I1206 04:01:02.310396 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f6179c8e1032314aca8ad9f81dcfb870cb9717ebf0a2fdd6cc8068e3b5b3f2dd" Dec 06 04:01:02 crc kubenswrapper[4802]: I1206 04:01:02.387124 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 06 04:01:02 crc kubenswrapper[4802]: E1206 04:01:02.387619 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f09f7e4f-6e98-4461-bf3a-1a6aa9528c09" containerName="nova-cell0-conductor-db-sync" Dec 06 04:01:02 crc kubenswrapper[4802]: I1206 04:01:02.387640 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="f09f7e4f-6e98-4461-bf3a-1a6aa9528c09" containerName="nova-cell0-conductor-db-sync" Dec 06 04:01:02 crc kubenswrapper[4802]: I1206 04:01:02.387885 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="f09f7e4f-6e98-4461-bf3a-1a6aa9528c09" containerName="nova-cell0-conductor-db-sync" Dec 06 04:01:02 crc kubenswrapper[4802]: I1206 04:01:02.388621 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 06 04:01:02 crc kubenswrapper[4802]: I1206 04:01:02.393571 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 06 04:01:02 crc kubenswrapper[4802]: I1206 04:01:02.393782 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-z765z" Dec 06 04:01:02 crc kubenswrapper[4802]: I1206 04:01:02.407193 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 06 04:01:02 crc kubenswrapper[4802]: I1206 04:01:02.431458 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c1b64f2-04f9-4fa9-ba9b-852f1383f6ed-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"2c1b64f2-04f9-4fa9-ba9b-852f1383f6ed\") " pod="openstack/nova-cell0-conductor-0" Dec 06 04:01:02 crc kubenswrapper[4802]: I1206 04:01:02.431574 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c1b64f2-04f9-4fa9-ba9b-852f1383f6ed-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"2c1b64f2-04f9-4fa9-ba9b-852f1383f6ed\") " pod="openstack/nova-cell0-conductor-0" Dec 06 04:01:02 crc kubenswrapper[4802]: I1206 04:01:02.431676 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kx24p\" (UniqueName: \"kubernetes.io/projected/2c1b64f2-04f9-4fa9-ba9b-852f1383f6ed-kube-api-access-kx24p\") pod \"nova-cell0-conductor-0\" (UID: \"2c1b64f2-04f9-4fa9-ba9b-852f1383f6ed\") " pod="openstack/nova-cell0-conductor-0" Dec 06 04:01:02 crc kubenswrapper[4802]: I1206 04:01:02.533476 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kx24p\" (UniqueName: \"kubernetes.io/projected/2c1b64f2-04f9-4fa9-ba9b-852f1383f6ed-kube-api-access-kx24p\") pod \"nova-cell0-conductor-0\" (UID: \"2c1b64f2-04f9-4fa9-ba9b-852f1383f6ed\") " pod="openstack/nova-cell0-conductor-0" Dec 06 04:01:02 crc kubenswrapper[4802]: I1206 04:01:02.533576 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c1b64f2-04f9-4fa9-ba9b-852f1383f6ed-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"2c1b64f2-04f9-4fa9-ba9b-852f1383f6ed\") " pod="openstack/nova-cell0-conductor-0" Dec 06 04:01:02 crc kubenswrapper[4802]: I1206 04:01:02.533719 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c1b64f2-04f9-4fa9-ba9b-852f1383f6ed-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"2c1b64f2-04f9-4fa9-ba9b-852f1383f6ed\") " pod="openstack/nova-cell0-conductor-0" Dec 06 04:01:02 crc kubenswrapper[4802]: I1206 04:01:02.539359 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c1b64f2-04f9-4fa9-ba9b-852f1383f6ed-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"2c1b64f2-04f9-4fa9-ba9b-852f1383f6ed\") " pod="openstack/nova-cell0-conductor-0" Dec 06 04:01:02 crc kubenswrapper[4802]: I1206 04:01:02.555232 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kx24p\" (UniqueName: \"kubernetes.io/projected/2c1b64f2-04f9-4fa9-ba9b-852f1383f6ed-kube-api-access-kx24p\") pod \"nova-cell0-conductor-0\" (UID: \"2c1b64f2-04f9-4fa9-ba9b-852f1383f6ed\") " pod="openstack/nova-cell0-conductor-0" Dec 06 04:01:02 crc kubenswrapper[4802]: I1206 04:01:02.560288 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c1b64f2-04f9-4fa9-ba9b-852f1383f6ed-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"2c1b64f2-04f9-4fa9-ba9b-852f1383f6ed\") " pod="openstack/nova-cell0-conductor-0" Dec 06 04:01:02 crc kubenswrapper[4802]: I1206 04:01:02.740824 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 06 04:01:03 crc kubenswrapper[4802]: I1206 04:01:03.259281 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 06 04:01:03 crc kubenswrapper[4802]: W1206 04:01:03.262275 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2c1b64f2_04f9_4fa9_ba9b_852f1383f6ed.slice/crio-a0d86e36c9d8d9b15d353e433bfe24bfea2c9ec21a5a04dbec5f668a9d5bba25 WatchSource:0}: Error finding container a0d86e36c9d8d9b15d353e433bfe24bfea2c9ec21a5a04dbec5f668a9d5bba25: Status 404 returned error can't find the container with id a0d86e36c9d8d9b15d353e433bfe24bfea2c9ec21a5a04dbec5f668a9d5bba25 Dec 06 04:01:03 crc kubenswrapper[4802]: I1206 04:01:03.335330 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"30152cc8-8968-443f-923a-0ddca8c016ee","Type":"ContainerStarted","Data":"4e67921587bb5beca9f50d6d82a1e653bdf472c8cbd66272021421e0474dba82"} Dec 06 04:01:03 crc kubenswrapper[4802]: I1206 04:01:03.335383 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"30152cc8-8968-443f-923a-0ddca8c016ee","Type":"ContainerStarted","Data":"4b2f1959e7f989b0dc68503aec47cb31044c3a075f0bc892a91e9336697a3c15"} Dec 06 04:01:03 crc kubenswrapper[4802]: I1206 04:01:03.340479 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"2c1b64f2-04f9-4fa9-ba9b-852f1383f6ed","Type":"ContainerStarted","Data":"a0d86e36c9d8d9b15d353e433bfe24bfea2c9ec21a5a04dbec5f668a9d5bba25"} Dec 06 04:01:04 crc kubenswrapper[4802]: I1206 04:01:04.352186 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"2c1b64f2-04f9-4fa9-ba9b-852f1383f6ed","Type":"ContainerStarted","Data":"a6aebbf55e0758061bc13068c965403948e93f17c5c886f2db90c89c42450d0b"} Dec 06 04:01:04 crc kubenswrapper[4802]: I1206 04:01:04.352527 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 06 04:01:04 crc kubenswrapper[4802]: I1206 04:01:04.354878 4802 generic.go:334] "Generic (PLEG): container finished" podID="613478db-c10d-476b-b66d-4ae3572067ff" containerID="c58f5148ad0e68a7754ea51674c34cf83f6dcdc2f9b32cb93434b60414c0d5eb" exitCode=0 Dec 06 04:01:04 crc kubenswrapper[4802]: I1206 04:01:04.354908 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29416561-rnfxm" event={"ID":"613478db-c10d-476b-b66d-4ae3572067ff","Type":"ContainerDied","Data":"c58f5148ad0e68a7754ea51674c34cf83f6dcdc2f9b32cb93434b60414c0d5eb"} Dec 06 04:01:04 crc kubenswrapper[4802]: I1206 04:01:04.371796 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.371781756 podStartE2EDuration="2.371781756s" podCreationTimestamp="2025-12-06 04:01:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:01:04.369812692 +0000 UTC m=+1257.241721844" watchObservedRunningTime="2025-12-06 04:01:04.371781756 +0000 UTC m=+1257.243690908" Dec 06 04:01:05 crc kubenswrapper[4802]: I1206 04:01:05.366611 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"30152cc8-8968-443f-923a-0ddca8c016ee","Type":"ContainerStarted","Data":"7b9a8b8ec82130774b4701a4abc56aeb457b4266e4d179334ca9825fb673f20b"} Dec 06 04:01:05 crc kubenswrapper[4802]: I1206 04:01:05.391972 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.9537038070000001 podStartE2EDuration="5.391950024s" podCreationTimestamp="2025-12-06 04:01:00 +0000 UTC" firstStartedPulling="2025-12-06 04:01:01.258551467 +0000 UTC m=+1254.130460629" lastFinishedPulling="2025-12-06 04:01:04.696797664 +0000 UTC m=+1257.568706846" observedRunningTime="2025-12-06 04:01:05.388439829 +0000 UTC m=+1258.260349021" watchObservedRunningTime="2025-12-06 04:01:05.391950024 +0000 UTC m=+1258.263859196" Dec 06 04:01:05 crc kubenswrapper[4802]: I1206 04:01:05.784969 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29416561-rnfxm" Dec 06 04:01:05 crc kubenswrapper[4802]: I1206 04:01:05.796378 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-74skm\" (UniqueName: \"kubernetes.io/projected/613478db-c10d-476b-b66d-4ae3572067ff-kube-api-access-74skm\") pod \"613478db-c10d-476b-b66d-4ae3572067ff\" (UID: \"613478db-c10d-476b-b66d-4ae3572067ff\") " Dec 06 04:01:05 crc kubenswrapper[4802]: I1206 04:01:05.796423 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/613478db-c10d-476b-b66d-4ae3572067ff-config-data\") pod \"613478db-c10d-476b-b66d-4ae3572067ff\" (UID: \"613478db-c10d-476b-b66d-4ae3572067ff\") " Dec 06 04:01:05 crc kubenswrapper[4802]: I1206 04:01:05.796549 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/613478db-c10d-476b-b66d-4ae3572067ff-fernet-keys\") pod \"613478db-c10d-476b-b66d-4ae3572067ff\" (UID: \"613478db-c10d-476b-b66d-4ae3572067ff\") " Dec 06 04:01:05 crc kubenswrapper[4802]: I1206 04:01:05.796592 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/613478db-c10d-476b-b66d-4ae3572067ff-combined-ca-bundle\") pod \"613478db-c10d-476b-b66d-4ae3572067ff\" (UID: \"613478db-c10d-476b-b66d-4ae3572067ff\") " Dec 06 04:01:05 crc kubenswrapper[4802]: I1206 04:01:05.819351 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/613478db-c10d-476b-b66d-4ae3572067ff-kube-api-access-74skm" (OuterVolumeSpecName: "kube-api-access-74skm") pod "613478db-c10d-476b-b66d-4ae3572067ff" (UID: "613478db-c10d-476b-b66d-4ae3572067ff"). InnerVolumeSpecName "kube-api-access-74skm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:01:05 crc kubenswrapper[4802]: I1206 04:01:05.845927 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/613478db-c10d-476b-b66d-4ae3572067ff-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "613478db-c10d-476b-b66d-4ae3572067ff" (UID: "613478db-c10d-476b-b66d-4ae3572067ff"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:01:05 crc kubenswrapper[4802]: I1206 04:01:05.858264 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/613478db-c10d-476b-b66d-4ae3572067ff-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "613478db-c10d-476b-b66d-4ae3572067ff" (UID: "613478db-c10d-476b-b66d-4ae3572067ff"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:01:05 crc kubenswrapper[4802]: I1206 04:01:05.869385 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/613478db-c10d-476b-b66d-4ae3572067ff-config-data" (OuterVolumeSpecName: "config-data") pod "613478db-c10d-476b-b66d-4ae3572067ff" (UID: "613478db-c10d-476b-b66d-4ae3572067ff"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:01:05 crc kubenswrapper[4802]: I1206 04:01:05.898327 4802 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/613478db-c10d-476b-b66d-4ae3572067ff-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:05 crc kubenswrapper[4802]: I1206 04:01:05.898366 4802 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/613478db-c10d-476b-b66d-4ae3572067ff-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:05 crc kubenswrapper[4802]: I1206 04:01:05.898382 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-74skm\" (UniqueName: \"kubernetes.io/projected/613478db-c10d-476b-b66d-4ae3572067ff-kube-api-access-74skm\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:05 crc kubenswrapper[4802]: I1206 04:01:05.898396 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/613478db-c10d-476b-b66d-4ae3572067ff-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:06 crc kubenswrapper[4802]: I1206 04:01:06.379822 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29416561-rnfxm" event={"ID":"613478db-c10d-476b-b66d-4ae3572067ff","Type":"ContainerDied","Data":"79714aff7c8ee9d9ea09bdfda43fcba819a5f2b0f22e9b2cc547c36818bbbcf2"} Dec 06 04:01:06 crc kubenswrapper[4802]: I1206 04:01:06.379881 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="79714aff7c8ee9d9ea09bdfda43fcba819a5f2b0f22e9b2cc547c36818bbbcf2" Dec 06 04:01:06 crc kubenswrapper[4802]: I1206 04:01:06.379842 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29416561-rnfxm" Dec 06 04:01:06 crc kubenswrapper[4802]: I1206 04:01:06.380072 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 06 04:01:12 crc kubenswrapper[4802]: I1206 04:01:12.771555 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.239221 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-kx676"] Dec 06 04:01:13 crc kubenswrapper[4802]: E1206 04:01:13.239793 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="613478db-c10d-476b-b66d-4ae3572067ff" containerName="keystone-cron" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.239816 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="613478db-c10d-476b-b66d-4ae3572067ff" containerName="keystone-cron" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.240070 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="613478db-c10d-476b-b66d-4ae3572067ff" containerName="keystone-cron" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.240900 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-kx676" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.247639 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-kx676"] Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.250221 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.250310 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.346975 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb7d4cc2-ea97-4675-bc65-8803fa1e3708-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-kx676\" (UID: \"bb7d4cc2-ea97-4675-bc65-8803fa1e3708\") " pod="openstack/nova-cell0-cell-mapping-kx676" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.347185 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb7d4cc2-ea97-4675-bc65-8803fa1e3708-scripts\") pod \"nova-cell0-cell-mapping-kx676\" (UID: \"bb7d4cc2-ea97-4675-bc65-8803fa1e3708\") " pod="openstack/nova-cell0-cell-mapping-kx676" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.347433 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbpx8\" (UniqueName: \"kubernetes.io/projected/bb7d4cc2-ea97-4675-bc65-8803fa1e3708-kube-api-access-bbpx8\") pod \"nova-cell0-cell-mapping-kx676\" (UID: \"bb7d4cc2-ea97-4675-bc65-8803fa1e3708\") " pod="openstack/nova-cell0-cell-mapping-kx676" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.347553 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb7d4cc2-ea97-4675-bc65-8803fa1e3708-config-data\") pod \"nova-cell0-cell-mapping-kx676\" (UID: \"bb7d4cc2-ea97-4675-bc65-8803fa1e3708\") " pod="openstack/nova-cell0-cell-mapping-kx676" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.444777 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.447074 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.454307 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.464246 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb7d4cc2-ea97-4675-bc65-8803fa1e3708-config-data\") pod \"nova-cell0-cell-mapping-kx676\" (UID: \"bb7d4cc2-ea97-4675-bc65-8803fa1e3708\") " pod="openstack/nova-cell0-cell-mapping-kx676" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.464399 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb7d4cc2-ea97-4675-bc65-8803fa1e3708-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-kx676\" (UID: \"bb7d4cc2-ea97-4675-bc65-8803fa1e3708\") " pod="openstack/nova-cell0-cell-mapping-kx676" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.464494 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb7d4cc2-ea97-4675-bc65-8803fa1e3708-scripts\") pod \"nova-cell0-cell-mapping-kx676\" (UID: \"bb7d4cc2-ea97-4675-bc65-8803fa1e3708\") " pod="openstack/nova-cell0-cell-mapping-kx676" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.464649 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbpx8\" (UniqueName: \"kubernetes.io/projected/bb7d4cc2-ea97-4675-bc65-8803fa1e3708-kube-api-access-bbpx8\") pod \"nova-cell0-cell-mapping-kx676\" (UID: \"bb7d4cc2-ea97-4675-bc65-8803fa1e3708\") " pod="openstack/nova-cell0-cell-mapping-kx676" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.465107 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.476666 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb7d4cc2-ea97-4675-bc65-8803fa1e3708-config-data\") pod \"nova-cell0-cell-mapping-kx676\" (UID: \"bb7d4cc2-ea97-4675-bc65-8803fa1e3708\") " pod="openstack/nova-cell0-cell-mapping-kx676" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.479920 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb7d4cc2-ea97-4675-bc65-8803fa1e3708-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-kx676\" (UID: \"bb7d4cc2-ea97-4675-bc65-8803fa1e3708\") " pod="openstack/nova-cell0-cell-mapping-kx676" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.492530 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb7d4cc2-ea97-4675-bc65-8803fa1e3708-scripts\") pod \"nova-cell0-cell-mapping-kx676\" (UID: \"bb7d4cc2-ea97-4675-bc65-8803fa1e3708\") " pod="openstack/nova-cell0-cell-mapping-kx676" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.517143 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbpx8\" (UniqueName: \"kubernetes.io/projected/bb7d4cc2-ea97-4675-bc65-8803fa1e3708-kube-api-access-bbpx8\") pod \"nova-cell0-cell-mapping-kx676\" (UID: \"bb7d4cc2-ea97-4675-bc65-8803fa1e3708\") " pod="openstack/nova-cell0-cell-mapping-kx676" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.563152 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-kx676" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.567846 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.569446 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.574394 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.574868 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fs2n\" (UniqueName: \"kubernetes.io/projected/05ec15f4-9b81-4585-adfd-c749c7b12c38-kube-api-access-8fs2n\") pod \"nova-api-0\" (UID: \"05ec15f4-9b81-4585-adfd-c749c7b12c38\") " pod="openstack/nova-api-0" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.575005 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/05ec15f4-9b81-4585-adfd-c749c7b12c38-logs\") pod \"nova-api-0\" (UID: \"05ec15f4-9b81-4585-adfd-c749c7b12c38\") " pod="openstack/nova-api-0" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.575040 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05ec15f4-9b81-4585-adfd-c749c7b12c38-config-data\") pod \"nova-api-0\" (UID: \"05ec15f4-9b81-4585-adfd-c749c7b12c38\") " pod="openstack/nova-api-0" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.575139 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05ec15f4-9b81-4585-adfd-c749c7b12c38-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"05ec15f4-9b81-4585-adfd-c749c7b12c38\") " pod="openstack/nova-api-0" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.588074 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.598898 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.602189 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.620053 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.650812 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.652086 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.657100 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.673294 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.676696 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8fs2n\" (UniqueName: \"kubernetes.io/projected/05ec15f4-9b81-4585-adfd-c749c7b12c38-kube-api-access-8fs2n\") pod \"nova-api-0\" (UID: \"05ec15f4-9b81-4585-adfd-c749c7b12c38\") " pod="openstack/nova-api-0" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.676800 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab782538-46b6-4426-9bb4-e6aba3007da6-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ab782538-46b6-4426-9bb4-e6aba3007da6\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.676859 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/05ec15f4-9b81-4585-adfd-c749c7b12c38-logs\") pod \"nova-api-0\" (UID: \"05ec15f4-9b81-4585-adfd-c749c7b12c38\") " pod="openstack/nova-api-0" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.676880 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab782538-46b6-4426-9bb4-e6aba3007da6-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ab782538-46b6-4426-9bb4-e6aba3007da6\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.676902 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05ec15f4-9b81-4585-adfd-c749c7b12c38-config-data\") pod \"nova-api-0\" (UID: \"05ec15f4-9b81-4585-adfd-c749c7b12c38\") " pod="openstack/nova-api-0" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.676958 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhsd4\" (UniqueName: \"kubernetes.io/projected/ab782538-46b6-4426-9bb4-e6aba3007da6-kube-api-access-qhsd4\") pod \"nova-cell1-novncproxy-0\" (UID: \"ab782538-46b6-4426-9bb4-e6aba3007da6\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.676990 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05ec15f4-9b81-4585-adfd-c749c7b12c38-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"05ec15f4-9b81-4585-adfd-c749c7b12c38\") " pod="openstack/nova-api-0" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.677619 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/05ec15f4-9b81-4585-adfd-c749c7b12c38-logs\") pod \"nova-api-0\" (UID: \"05ec15f4-9b81-4585-adfd-c749c7b12c38\") " pod="openstack/nova-api-0" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.681308 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05ec15f4-9b81-4585-adfd-c749c7b12c38-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"05ec15f4-9b81-4585-adfd-c749c7b12c38\") " pod="openstack/nova-api-0" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.684416 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05ec15f4-9b81-4585-adfd-c749c7b12c38-config-data\") pod \"nova-api-0\" (UID: \"05ec15f4-9b81-4585-adfd-c749c7b12c38\") " pod="openstack/nova-api-0" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.714974 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8fs2n\" (UniqueName: \"kubernetes.io/projected/05ec15f4-9b81-4585-adfd-c749c7b12c38-kube-api-access-8fs2n\") pod \"nova-api-0\" (UID: \"05ec15f4-9b81-4585-adfd-c749c7b12c38\") " pod="openstack/nova-api-0" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.715070 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.762060 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-568d7fd7cf-246lw"] Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.763778 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.764236 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-568d7fd7cf-246lw" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.780370 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4590771f-ac87-4cbc-86c3-a069ca74cf43-logs\") pod \"nova-metadata-0\" (UID: \"4590771f-ac87-4cbc-86c3-a069ca74cf43\") " pod="openstack/nova-metadata-0" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.780440 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab782538-46b6-4426-9bb4-e6aba3007da6-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ab782538-46b6-4426-9bb4-e6aba3007da6\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.780458 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71c5fc79-7c68-4406-b02a-6cf325075a7a-config-data\") pod \"nova-scheduler-0\" (UID: \"71c5fc79-7c68-4406-b02a-6cf325075a7a\") " pod="openstack/nova-scheduler-0" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.780487 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71c5fc79-7c68-4406-b02a-6cf325075a7a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"71c5fc79-7c68-4406-b02a-6cf325075a7a\") " pod="openstack/nova-scheduler-0" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.780517 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab782538-46b6-4426-9bb4-e6aba3007da6-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ab782538-46b6-4426-9bb4-e6aba3007da6\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.780551 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4590771f-ac87-4cbc-86c3-a069ca74cf43-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4590771f-ac87-4cbc-86c3-a069ca74cf43\") " pod="openstack/nova-metadata-0" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.780570 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nsc65\" (UniqueName: \"kubernetes.io/projected/4590771f-ac87-4cbc-86c3-a069ca74cf43-kube-api-access-nsc65\") pod \"nova-metadata-0\" (UID: \"4590771f-ac87-4cbc-86c3-a069ca74cf43\") " pod="openstack/nova-metadata-0" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.780588 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4590771f-ac87-4cbc-86c3-a069ca74cf43-config-data\") pod \"nova-metadata-0\" (UID: \"4590771f-ac87-4cbc-86c3-a069ca74cf43\") " pod="openstack/nova-metadata-0" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.780632 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhsd4\" (UniqueName: \"kubernetes.io/projected/ab782538-46b6-4426-9bb4-e6aba3007da6-kube-api-access-qhsd4\") pod \"nova-cell1-novncproxy-0\" (UID: \"ab782538-46b6-4426-9bb4-e6aba3007da6\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.780669 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n87f5\" (UniqueName: \"kubernetes.io/projected/71c5fc79-7c68-4406-b02a-6cf325075a7a-kube-api-access-n87f5\") pod \"nova-scheduler-0\" (UID: \"71c5fc79-7c68-4406-b02a-6cf325075a7a\") " pod="openstack/nova-scheduler-0" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.794669 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab782538-46b6-4426-9bb4-e6aba3007da6-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ab782538-46b6-4426-9bb4-e6aba3007da6\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.795352 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab782538-46b6-4426-9bb4-e6aba3007da6-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ab782538-46b6-4426-9bb4-e6aba3007da6\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.824495 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhsd4\" (UniqueName: \"kubernetes.io/projected/ab782538-46b6-4426-9bb4-e6aba3007da6-kube-api-access-qhsd4\") pod \"nova-cell1-novncproxy-0\" (UID: \"ab782538-46b6-4426-9bb4-e6aba3007da6\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.840426 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-568d7fd7cf-246lw"] Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.891330 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71c5fc79-7c68-4406-b02a-6cf325075a7a-config-data\") pod \"nova-scheduler-0\" (UID: \"71c5fc79-7c68-4406-b02a-6cf325075a7a\") " pod="openstack/nova-scheduler-0" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.891391 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dd12e103-8229-466d-b8e2-5bf5eaa140bc-dns-svc\") pod \"dnsmasq-dns-568d7fd7cf-246lw\" (UID: \"dd12e103-8229-466d-b8e2-5bf5eaa140bc\") " pod="openstack/dnsmasq-dns-568d7fd7cf-246lw" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.891430 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71c5fc79-7c68-4406-b02a-6cf325075a7a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"71c5fc79-7c68-4406-b02a-6cf325075a7a\") " pod="openstack/nova-scheduler-0" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.891478 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dd12e103-8229-466d-b8e2-5bf5eaa140bc-dns-swift-storage-0\") pod \"dnsmasq-dns-568d7fd7cf-246lw\" (UID: \"dd12e103-8229-466d-b8e2-5bf5eaa140bc\") " pod="openstack/dnsmasq-dns-568d7fd7cf-246lw" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.891522 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4590771f-ac87-4cbc-86c3-a069ca74cf43-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4590771f-ac87-4cbc-86c3-a069ca74cf43\") " pod="openstack/nova-metadata-0" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.891550 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nsc65\" (UniqueName: \"kubernetes.io/projected/4590771f-ac87-4cbc-86c3-a069ca74cf43-kube-api-access-nsc65\") pod \"nova-metadata-0\" (UID: \"4590771f-ac87-4cbc-86c3-a069ca74cf43\") " pod="openstack/nova-metadata-0" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.891573 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4590771f-ac87-4cbc-86c3-a069ca74cf43-config-data\") pod \"nova-metadata-0\" (UID: \"4590771f-ac87-4cbc-86c3-a069ca74cf43\") " pod="openstack/nova-metadata-0" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.891603 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd12e103-8229-466d-b8e2-5bf5eaa140bc-config\") pod \"dnsmasq-dns-568d7fd7cf-246lw\" (UID: \"dd12e103-8229-466d-b8e2-5bf5eaa140bc\") " pod="openstack/dnsmasq-dns-568d7fd7cf-246lw" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.891649 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dd12e103-8229-466d-b8e2-5bf5eaa140bc-ovsdbserver-nb\") pod \"dnsmasq-dns-568d7fd7cf-246lw\" (UID: \"dd12e103-8229-466d-b8e2-5bf5eaa140bc\") " pod="openstack/dnsmasq-dns-568d7fd7cf-246lw" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.891679 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dd12e103-8229-466d-b8e2-5bf5eaa140bc-ovsdbserver-sb\") pod \"dnsmasq-dns-568d7fd7cf-246lw\" (UID: \"dd12e103-8229-466d-b8e2-5bf5eaa140bc\") " pod="openstack/dnsmasq-dns-568d7fd7cf-246lw" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.891711 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wb66l\" (UniqueName: \"kubernetes.io/projected/dd12e103-8229-466d-b8e2-5bf5eaa140bc-kube-api-access-wb66l\") pod \"dnsmasq-dns-568d7fd7cf-246lw\" (UID: \"dd12e103-8229-466d-b8e2-5bf5eaa140bc\") " pod="openstack/dnsmasq-dns-568d7fd7cf-246lw" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.891739 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n87f5\" (UniqueName: \"kubernetes.io/projected/71c5fc79-7c68-4406-b02a-6cf325075a7a-kube-api-access-n87f5\") pod \"nova-scheduler-0\" (UID: \"71c5fc79-7c68-4406-b02a-6cf325075a7a\") " pod="openstack/nova-scheduler-0" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.891807 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4590771f-ac87-4cbc-86c3-a069ca74cf43-logs\") pod \"nova-metadata-0\" (UID: \"4590771f-ac87-4cbc-86c3-a069ca74cf43\") " pod="openstack/nova-metadata-0" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.892289 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4590771f-ac87-4cbc-86c3-a069ca74cf43-logs\") pod \"nova-metadata-0\" (UID: \"4590771f-ac87-4cbc-86c3-a069ca74cf43\") " pod="openstack/nova-metadata-0" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.903399 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4590771f-ac87-4cbc-86c3-a069ca74cf43-config-data\") pod \"nova-metadata-0\" (UID: \"4590771f-ac87-4cbc-86c3-a069ca74cf43\") " pod="openstack/nova-metadata-0" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.909239 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.910077 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71c5fc79-7c68-4406-b02a-6cf325075a7a-config-data\") pod \"nova-scheduler-0\" (UID: \"71c5fc79-7c68-4406-b02a-6cf325075a7a\") " pod="openstack/nova-scheduler-0" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.917470 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71c5fc79-7c68-4406-b02a-6cf325075a7a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"71c5fc79-7c68-4406-b02a-6cf325075a7a\") " pod="openstack/nova-scheduler-0" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.925314 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4590771f-ac87-4cbc-86c3-a069ca74cf43-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4590771f-ac87-4cbc-86c3-a069ca74cf43\") " pod="openstack/nova-metadata-0" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.981270 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n87f5\" (UniqueName: \"kubernetes.io/projected/71c5fc79-7c68-4406-b02a-6cf325075a7a-kube-api-access-n87f5\") pod \"nova-scheduler-0\" (UID: \"71c5fc79-7c68-4406-b02a-6cf325075a7a\") " pod="openstack/nova-scheduler-0" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.988541 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nsc65\" (UniqueName: \"kubernetes.io/projected/4590771f-ac87-4cbc-86c3-a069ca74cf43-kube-api-access-nsc65\") pod \"nova-metadata-0\" (UID: \"4590771f-ac87-4cbc-86c3-a069ca74cf43\") " pod="openstack/nova-metadata-0" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.993321 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dd12e103-8229-466d-b8e2-5bf5eaa140bc-dns-svc\") pod \"dnsmasq-dns-568d7fd7cf-246lw\" (UID: \"dd12e103-8229-466d-b8e2-5bf5eaa140bc\") " pod="openstack/dnsmasq-dns-568d7fd7cf-246lw" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.993390 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dd12e103-8229-466d-b8e2-5bf5eaa140bc-dns-swift-storage-0\") pod \"dnsmasq-dns-568d7fd7cf-246lw\" (UID: \"dd12e103-8229-466d-b8e2-5bf5eaa140bc\") " pod="openstack/dnsmasq-dns-568d7fd7cf-246lw" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.993430 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd12e103-8229-466d-b8e2-5bf5eaa140bc-config\") pod \"dnsmasq-dns-568d7fd7cf-246lw\" (UID: \"dd12e103-8229-466d-b8e2-5bf5eaa140bc\") " pod="openstack/dnsmasq-dns-568d7fd7cf-246lw" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.993464 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dd12e103-8229-466d-b8e2-5bf5eaa140bc-ovsdbserver-nb\") pod \"dnsmasq-dns-568d7fd7cf-246lw\" (UID: \"dd12e103-8229-466d-b8e2-5bf5eaa140bc\") " pod="openstack/dnsmasq-dns-568d7fd7cf-246lw" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.993485 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dd12e103-8229-466d-b8e2-5bf5eaa140bc-ovsdbserver-sb\") pod \"dnsmasq-dns-568d7fd7cf-246lw\" (UID: \"dd12e103-8229-466d-b8e2-5bf5eaa140bc\") " pod="openstack/dnsmasq-dns-568d7fd7cf-246lw" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.993511 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wb66l\" (UniqueName: \"kubernetes.io/projected/dd12e103-8229-466d-b8e2-5bf5eaa140bc-kube-api-access-wb66l\") pod \"dnsmasq-dns-568d7fd7cf-246lw\" (UID: \"dd12e103-8229-466d-b8e2-5bf5eaa140bc\") " pod="openstack/dnsmasq-dns-568d7fd7cf-246lw" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.994544 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dd12e103-8229-466d-b8e2-5bf5eaa140bc-dns-swift-storage-0\") pod \"dnsmasq-dns-568d7fd7cf-246lw\" (UID: \"dd12e103-8229-466d-b8e2-5bf5eaa140bc\") " pod="openstack/dnsmasq-dns-568d7fd7cf-246lw" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.994605 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dd12e103-8229-466d-b8e2-5bf5eaa140bc-dns-svc\") pod \"dnsmasq-dns-568d7fd7cf-246lw\" (UID: \"dd12e103-8229-466d-b8e2-5bf5eaa140bc\") " pod="openstack/dnsmasq-dns-568d7fd7cf-246lw" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.994718 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd12e103-8229-466d-b8e2-5bf5eaa140bc-config\") pod \"dnsmasq-dns-568d7fd7cf-246lw\" (UID: \"dd12e103-8229-466d-b8e2-5bf5eaa140bc\") " pod="openstack/dnsmasq-dns-568d7fd7cf-246lw" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.995085 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dd12e103-8229-466d-b8e2-5bf5eaa140bc-ovsdbserver-sb\") pod \"dnsmasq-dns-568d7fd7cf-246lw\" (UID: \"dd12e103-8229-466d-b8e2-5bf5eaa140bc\") " pod="openstack/dnsmasq-dns-568d7fd7cf-246lw" Dec 06 04:01:13 crc kubenswrapper[4802]: I1206 04:01:13.995159 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dd12e103-8229-466d-b8e2-5bf5eaa140bc-ovsdbserver-nb\") pod \"dnsmasq-dns-568d7fd7cf-246lw\" (UID: \"dd12e103-8229-466d-b8e2-5bf5eaa140bc\") " pod="openstack/dnsmasq-dns-568d7fd7cf-246lw" Dec 06 04:01:14 crc kubenswrapper[4802]: I1206 04:01:14.055391 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wb66l\" (UniqueName: \"kubernetes.io/projected/dd12e103-8229-466d-b8e2-5bf5eaa140bc-kube-api-access-wb66l\") pod \"dnsmasq-dns-568d7fd7cf-246lw\" (UID: \"dd12e103-8229-466d-b8e2-5bf5eaa140bc\") " pod="openstack/dnsmasq-dns-568d7fd7cf-246lw" Dec 06 04:01:14 crc kubenswrapper[4802]: I1206 04:01:14.065948 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 04:01:14 crc kubenswrapper[4802]: I1206 04:01:14.118526 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 04:01:14 crc kubenswrapper[4802]: I1206 04:01:14.129700 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-568d7fd7cf-246lw" Dec 06 04:01:14 crc kubenswrapper[4802]: I1206 04:01:14.344372 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-kx676"] Dec 06 04:01:14 crc kubenswrapper[4802]: I1206 04:01:14.470872 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-kx676" event={"ID":"bb7d4cc2-ea97-4675-bc65-8803fa1e3708","Type":"ContainerStarted","Data":"66747415eeef492e8b0853aca16ef0e036a53c19c47b149425105d835e5e55d3"} Dec 06 04:01:14 crc kubenswrapper[4802]: I1206 04:01:14.540437 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 06 04:01:14 crc kubenswrapper[4802]: I1206 04:01:14.615873 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-5sl8p"] Dec 06 04:01:14 crc kubenswrapper[4802]: I1206 04:01:14.617078 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-5sl8p" Dec 06 04:01:14 crc kubenswrapper[4802]: I1206 04:01:14.621207 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 06 04:01:14 crc kubenswrapper[4802]: I1206 04:01:14.621458 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 06 04:01:14 crc kubenswrapper[4802]: I1206 04:01:14.628802 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-5sl8p"] Dec 06 04:01:14 crc kubenswrapper[4802]: I1206 04:01:14.667597 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 04:01:14 crc kubenswrapper[4802]: I1206 04:01:14.725883 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 04:01:14 crc kubenswrapper[4802]: I1206 04:01:14.727409 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/735dec51-8659-44bf-bd0b-afd1a1d78d86-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-5sl8p\" (UID: \"735dec51-8659-44bf-bd0b-afd1a1d78d86\") " pod="openstack/nova-cell1-conductor-db-sync-5sl8p" Dec 06 04:01:14 crc kubenswrapper[4802]: I1206 04:01:14.727448 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxkwc\" (UniqueName: \"kubernetes.io/projected/735dec51-8659-44bf-bd0b-afd1a1d78d86-kube-api-access-bxkwc\") pod \"nova-cell1-conductor-db-sync-5sl8p\" (UID: \"735dec51-8659-44bf-bd0b-afd1a1d78d86\") " pod="openstack/nova-cell1-conductor-db-sync-5sl8p" Dec 06 04:01:14 crc kubenswrapper[4802]: I1206 04:01:14.727492 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/735dec51-8659-44bf-bd0b-afd1a1d78d86-config-data\") pod \"nova-cell1-conductor-db-sync-5sl8p\" (UID: \"735dec51-8659-44bf-bd0b-afd1a1d78d86\") " pod="openstack/nova-cell1-conductor-db-sync-5sl8p" Dec 06 04:01:14 crc kubenswrapper[4802]: I1206 04:01:14.727520 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/735dec51-8659-44bf-bd0b-afd1a1d78d86-scripts\") pod \"nova-cell1-conductor-db-sync-5sl8p\" (UID: \"735dec51-8659-44bf-bd0b-afd1a1d78d86\") " pod="openstack/nova-cell1-conductor-db-sync-5sl8p" Dec 06 04:01:14 crc kubenswrapper[4802]: I1206 04:01:14.829268 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/735dec51-8659-44bf-bd0b-afd1a1d78d86-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-5sl8p\" (UID: \"735dec51-8659-44bf-bd0b-afd1a1d78d86\") " pod="openstack/nova-cell1-conductor-db-sync-5sl8p" Dec 06 04:01:14 crc kubenswrapper[4802]: I1206 04:01:14.829322 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxkwc\" (UniqueName: \"kubernetes.io/projected/735dec51-8659-44bf-bd0b-afd1a1d78d86-kube-api-access-bxkwc\") pod \"nova-cell1-conductor-db-sync-5sl8p\" (UID: \"735dec51-8659-44bf-bd0b-afd1a1d78d86\") " pod="openstack/nova-cell1-conductor-db-sync-5sl8p" Dec 06 04:01:14 crc kubenswrapper[4802]: I1206 04:01:14.829383 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/735dec51-8659-44bf-bd0b-afd1a1d78d86-config-data\") pod \"nova-cell1-conductor-db-sync-5sl8p\" (UID: \"735dec51-8659-44bf-bd0b-afd1a1d78d86\") " pod="openstack/nova-cell1-conductor-db-sync-5sl8p" Dec 06 04:01:14 crc kubenswrapper[4802]: I1206 04:01:14.829423 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/735dec51-8659-44bf-bd0b-afd1a1d78d86-scripts\") pod \"nova-cell1-conductor-db-sync-5sl8p\" (UID: \"735dec51-8659-44bf-bd0b-afd1a1d78d86\") " pod="openstack/nova-cell1-conductor-db-sync-5sl8p" Dec 06 04:01:14 crc kubenswrapper[4802]: I1206 04:01:14.844307 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/735dec51-8659-44bf-bd0b-afd1a1d78d86-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-5sl8p\" (UID: \"735dec51-8659-44bf-bd0b-afd1a1d78d86\") " pod="openstack/nova-cell1-conductor-db-sync-5sl8p" Dec 06 04:01:14 crc kubenswrapper[4802]: I1206 04:01:14.844830 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/735dec51-8659-44bf-bd0b-afd1a1d78d86-scripts\") pod \"nova-cell1-conductor-db-sync-5sl8p\" (UID: \"735dec51-8659-44bf-bd0b-afd1a1d78d86\") " pod="openstack/nova-cell1-conductor-db-sync-5sl8p" Dec 06 04:01:14 crc kubenswrapper[4802]: I1206 04:01:14.847571 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/735dec51-8659-44bf-bd0b-afd1a1d78d86-config-data\") pod \"nova-cell1-conductor-db-sync-5sl8p\" (UID: \"735dec51-8659-44bf-bd0b-afd1a1d78d86\") " pod="openstack/nova-cell1-conductor-db-sync-5sl8p" Dec 06 04:01:14 crc kubenswrapper[4802]: I1206 04:01:14.849302 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxkwc\" (UniqueName: \"kubernetes.io/projected/735dec51-8659-44bf-bd0b-afd1a1d78d86-kube-api-access-bxkwc\") pod \"nova-cell1-conductor-db-sync-5sl8p\" (UID: \"735dec51-8659-44bf-bd0b-afd1a1d78d86\") " pod="openstack/nova-cell1-conductor-db-sync-5sl8p" Dec 06 04:01:14 crc kubenswrapper[4802]: I1206 04:01:14.879845 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 04:01:14 crc kubenswrapper[4802]: I1206 04:01:14.889654 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-568d7fd7cf-246lw"] Dec 06 04:01:14 crc kubenswrapper[4802]: I1206 04:01:14.943354 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-5sl8p" Dec 06 04:01:15 crc kubenswrapper[4802]: I1206 04:01:15.446917 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-5sl8p"] Dec 06 04:01:15 crc kubenswrapper[4802]: W1206 04:01:15.475933 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod735dec51_8659_44bf_bd0b_afd1a1d78d86.slice/crio-73abe00dafcadcca6e927805a277f8eef49528e5ee0867642b61611f1961be67 WatchSource:0}: Error finding container 73abe00dafcadcca6e927805a277f8eef49528e5ee0867642b61611f1961be67: Status 404 returned error can't find the container with id 73abe00dafcadcca6e927805a277f8eef49528e5ee0867642b61611f1961be67 Dec 06 04:01:15 crc kubenswrapper[4802]: I1206 04:01:15.510357 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-kx676" event={"ID":"bb7d4cc2-ea97-4675-bc65-8803fa1e3708","Type":"ContainerStarted","Data":"8e649e34b5ec7777e026185ed91acf6eb44aeaded3f44a8818c9d17502eb366d"} Dec 06 04:01:15 crc kubenswrapper[4802]: I1206 04:01:15.512716 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4590771f-ac87-4cbc-86c3-a069ca74cf43","Type":"ContainerStarted","Data":"5927928cc8d25aa9ea5f16989358ac1e435761cce9ebb659dfdfa4f9556973c4"} Dec 06 04:01:15 crc kubenswrapper[4802]: I1206 04:01:15.518803 4802 generic.go:334] "Generic (PLEG): container finished" podID="dd12e103-8229-466d-b8e2-5bf5eaa140bc" containerID="1c03aba4c50865b2bfc9d14891dbc7d7401c22bbdf3ca315d2b896283c9c70a0" exitCode=0 Dec 06 04:01:15 crc kubenswrapper[4802]: I1206 04:01:15.518902 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-568d7fd7cf-246lw" event={"ID":"dd12e103-8229-466d-b8e2-5bf5eaa140bc","Type":"ContainerDied","Data":"1c03aba4c50865b2bfc9d14891dbc7d7401c22bbdf3ca315d2b896283c9c70a0"} Dec 06 04:01:15 crc kubenswrapper[4802]: I1206 04:01:15.518938 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-568d7fd7cf-246lw" event={"ID":"dd12e103-8229-466d-b8e2-5bf5eaa140bc","Type":"ContainerStarted","Data":"6b9478bd0f2ad56fbb108028b8f264ba782adc418a18046814f3d26cff9b412b"} Dec 06 04:01:15 crc kubenswrapper[4802]: I1206 04:01:15.521523 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"05ec15f4-9b81-4585-adfd-c749c7b12c38","Type":"ContainerStarted","Data":"e977cdc34bb1a7fb53f49652f283594fa977f41520a14d51526cb52935405b1f"} Dec 06 04:01:15 crc kubenswrapper[4802]: I1206 04:01:15.532300 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"71c5fc79-7c68-4406-b02a-6cf325075a7a","Type":"ContainerStarted","Data":"475402fe59aa0ad464150ed550eefa2db54290355afdce7f5fc3f71dc2f5873b"} Dec 06 04:01:15 crc kubenswrapper[4802]: I1206 04:01:15.538376 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ab782538-46b6-4426-9bb4-e6aba3007da6","Type":"ContainerStarted","Data":"4ff9e1520cc1aca2a35c58fafa7add7422a74d0b6dbebcb5b02df899ec2306c0"} Dec 06 04:01:15 crc kubenswrapper[4802]: I1206 04:01:15.553961 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-kx676" podStartSLOduration=2.553939735 podStartE2EDuration="2.553939735s" podCreationTimestamp="2025-12-06 04:01:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:01:15.54793633 +0000 UTC m=+1268.419845492" watchObservedRunningTime="2025-12-06 04:01:15.553939735 +0000 UTC m=+1268.425848887" Dec 06 04:01:16 crc kubenswrapper[4802]: I1206 04:01:16.552639 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-568d7fd7cf-246lw" event={"ID":"dd12e103-8229-466d-b8e2-5bf5eaa140bc","Type":"ContainerStarted","Data":"55175fe1098f34cbf3f20d2c9e4fa9d2a30be8eba4ff9596438e223f98d62e83"} Dec 06 04:01:16 crc kubenswrapper[4802]: I1206 04:01:16.555984 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-568d7fd7cf-246lw" Dec 06 04:01:16 crc kubenswrapper[4802]: I1206 04:01:16.566920 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-5sl8p" event={"ID":"735dec51-8659-44bf-bd0b-afd1a1d78d86","Type":"ContainerStarted","Data":"3a7525fcbc73d3441910081d90724b15d3f5451db8ed85974e99baea473d1247"} Dec 06 04:01:16 crc kubenswrapper[4802]: I1206 04:01:16.566969 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-5sl8p" event={"ID":"735dec51-8659-44bf-bd0b-afd1a1d78d86","Type":"ContainerStarted","Data":"73abe00dafcadcca6e927805a277f8eef49528e5ee0867642b61611f1961be67"} Dec 06 04:01:16 crc kubenswrapper[4802]: I1206 04:01:16.588701 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-568d7fd7cf-246lw" podStartSLOduration=3.588683192 podStartE2EDuration="3.588683192s" podCreationTimestamp="2025-12-06 04:01:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:01:16.574963476 +0000 UTC m=+1269.446872648" watchObservedRunningTime="2025-12-06 04:01:16.588683192 +0000 UTC m=+1269.460592344" Dec 06 04:01:16 crc kubenswrapper[4802]: I1206 04:01:16.602453 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-5sl8p" podStartSLOduration=2.602437249 podStartE2EDuration="2.602437249s" podCreationTimestamp="2025-12-06 04:01:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:01:16.600412973 +0000 UTC m=+1269.472322125" watchObservedRunningTime="2025-12-06 04:01:16.602437249 +0000 UTC m=+1269.474346401" Dec 06 04:01:17 crc kubenswrapper[4802]: I1206 04:01:17.607409 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 04:01:17 crc kubenswrapper[4802]: I1206 04:01:17.621010 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 04:01:19 crc kubenswrapper[4802]: I1206 04:01:19.614390 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"71c5fc79-7c68-4406-b02a-6cf325075a7a","Type":"ContainerStarted","Data":"24648326d0ed4b1db9306783eeae2a39a0b46c7d4973ed6d5ad68bfc74434b67"} Dec 06 04:01:19 crc kubenswrapper[4802]: I1206 04:01:19.617148 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ab782538-46b6-4426-9bb4-e6aba3007da6","Type":"ContainerStarted","Data":"72d35a154ed99c981e13031caa104f30cd26dbfbe75fcce0172353ebb84806eb"} Dec 06 04:01:19 crc kubenswrapper[4802]: I1206 04:01:19.617231 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="ab782538-46b6-4426-9bb4-e6aba3007da6" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://72d35a154ed99c981e13031caa104f30cd26dbfbe75fcce0172353ebb84806eb" gracePeriod=30 Dec 06 04:01:19 crc kubenswrapper[4802]: I1206 04:01:19.619639 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4590771f-ac87-4cbc-86c3-a069ca74cf43","Type":"ContainerStarted","Data":"19b6365b8d16b8d17dbc5fbb16ed7a0235a0eaf366568c90d1010f400bd60d1c"} Dec 06 04:01:19 crc kubenswrapper[4802]: I1206 04:01:19.619688 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4590771f-ac87-4cbc-86c3-a069ca74cf43","Type":"ContainerStarted","Data":"c1d7f19deb9c8d205fc968f284ac68e9e1785fbaa425bc4a824ef166badbf78b"} Dec 06 04:01:19 crc kubenswrapper[4802]: I1206 04:01:19.619818 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="4590771f-ac87-4cbc-86c3-a069ca74cf43" containerName="nova-metadata-log" containerID="cri-o://c1d7f19deb9c8d205fc968f284ac68e9e1785fbaa425bc4a824ef166badbf78b" gracePeriod=30 Dec 06 04:01:19 crc kubenswrapper[4802]: I1206 04:01:19.619883 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="4590771f-ac87-4cbc-86c3-a069ca74cf43" containerName="nova-metadata-metadata" containerID="cri-o://19b6365b8d16b8d17dbc5fbb16ed7a0235a0eaf366568c90d1010f400bd60d1c" gracePeriod=30 Dec 06 04:01:19 crc kubenswrapper[4802]: I1206 04:01:19.627810 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"05ec15f4-9b81-4585-adfd-c749c7b12c38","Type":"ContainerStarted","Data":"c8fb2250bcb32700a9bcaa59793b36842ad24f7af491161c04f30f0dcb9dbd1f"} Dec 06 04:01:19 crc kubenswrapper[4802]: I1206 04:01:19.627862 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"05ec15f4-9b81-4585-adfd-c749c7b12c38","Type":"ContainerStarted","Data":"2dc54ae8aecbd3584f1f88ec065879db6466e3d5f5eae5ca7bfc561117d67fa2"} Dec 06 04:01:19 crc kubenswrapper[4802]: I1206 04:01:19.651851 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.215547867 podStartE2EDuration="6.651832929s" podCreationTimestamp="2025-12-06 04:01:13 +0000 UTC" firstStartedPulling="2025-12-06 04:01:14.91306222 +0000 UTC m=+1267.784971372" lastFinishedPulling="2025-12-06 04:01:18.349347282 +0000 UTC m=+1271.221256434" observedRunningTime="2025-12-06 04:01:19.643762178 +0000 UTC m=+1272.515671330" watchObservedRunningTime="2025-12-06 04:01:19.651832929 +0000 UTC m=+1272.523742081" Dec 06 04:01:19 crc kubenswrapper[4802]: I1206 04:01:19.703374 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.092898629 podStartE2EDuration="6.70335117s" podCreationTimestamp="2025-12-06 04:01:13 +0000 UTC" firstStartedPulling="2025-12-06 04:01:14.72972543 +0000 UTC m=+1267.601634582" lastFinishedPulling="2025-12-06 04:01:18.340177951 +0000 UTC m=+1271.212087123" observedRunningTime="2025-12-06 04:01:19.697352676 +0000 UTC m=+1272.569261828" watchObservedRunningTime="2025-12-06 04:01:19.70335117 +0000 UTC m=+1272.575260322" Dec 06 04:01:19 crc kubenswrapper[4802]: I1206 04:01:19.703982 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.021541355 podStartE2EDuration="6.703974207s" podCreationTimestamp="2025-12-06 04:01:13 +0000 UTC" firstStartedPulling="2025-12-06 04:01:14.659826326 +0000 UTC m=+1267.531735478" lastFinishedPulling="2025-12-06 04:01:18.342259168 +0000 UTC m=+1271.214168330" observedRunningTime="2025-12-06 04:01:19.679326212 +0000 UTC m=+1272.551235374" watchObservedRunningTime="2025-12-06 04:01:19.703974207 +0000 UTC m=+1272.575883359" Dec 06 04:01:19 crc kubenswrapper[4802]: I1206 04:01:19.716505 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.917336321 podStartE2EDuration="6.716487619s" podCreationTimestamp="2025-12-06 04:01:13 +0000 UTC" firstStartedPulling="2025-12-06 04:01:14.542396281 +0000 UTC m=+1267.414305433" lastFinishedPulling="2025-12-06 04:01:18.341547559 +0000 UTC m=+1271.213456731" observedRunningTime="2025-12-06 04:01:19.714192437 +0000 UTC m=+1272.586101589" watchObservedRunningTime="2025-12-06 04:01:19.716487619 +0000 UTC m=+1272.588396771" Dec 06 04:01:20 crc kubenswrapper[4802]: I1206 04:01:20.329436 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 04:01:20 crc kubenswrapper[4802]: I1206 04:01:20.495589 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4590771f-ac87-4cbc-86c3-a069ca74cf43-combined-ca-bundle\") pod \"4590771f-ac87-4cbc-86c3-a069ca74cf43\" (UID: \"4590771f-ac87-4cbc-86c3-a069ca74cf43\") " Dec 06 04:01:20 crc kubenswrapper[4802]: I1206 04:01:20.496167 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4590771f-ac87-4cbc-86c3-a069ca74cf43-logs\") pod \"4590771f-ac87-4cbc-86c3-a069ca74cf43\" (UID: \"4590771f-ac87-4cbc-86c3-a069ca74cf43\") " Dec 06 04:01:20 crc kubenswrapper[4802]: I1206 04:01:20.496241 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4590771f-ac87-4cbc-86c3-a069ca74cf43-config-data\") pod \"4590771f-ac87-4cbc-86c3-a069ca74cf43\" (UID: \"4590771f-ac87-4cbc-86c3-a069ca74cf43\") " Dec 06 04:01:20 crc kubenswrapper[4802]: I1206 04:01:20.496313 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nsc65\" (UniqueName: \"kubernetes.io/projected/4590771f-ac87-4cbc-86c3-a069ca74cf43-kube-api-access-nsc65\") pod \"4590771f-ac87-4cbc-86c3-a069ca74cf43\" (UID: \"4590771f-ac87-4cbc-86c3-a069ca74cf43\") " Dec 06 04:01:20 crc kubenswrapper[4802]: I1206 04:01:20.496680 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4590771f-ac87-4cbc-86c3-a069ca74cf43-logs" (OuterVolumeSpecName: "logs") pod "4590771f-ac87-4cbc-86c3-a069ca74cf43" (UID: "4590771f-ac87-4cbc-86c3-a069ca74cf43"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:01:20 crc kubenswrapper[4802]: I1206 04:01:20.501571 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4590771f-ac87-4cbc-86c3-a069ca74cf43-kube-api-access-nsc65" (OuterVolumeSpecName: "kube-api-access-nsc65") pod "4590771f-ac87-4cbc-86c3-a069ca74cf43" (UID: "4590771f-ac87-4cbc-86c3-a069ca74cf43"). InnerVolumeSpecName "kube-api-access-nsc65". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:01:20 crc kubenswrapper[4802]: I1206 04:01:20.549146 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4590771f-ac87-4cbc-86c3-a069ca74cf43-config-data" (OuterVolumeSpecName: "config-data") pod "4590771f-ac87-4cbc-86c3-a069ca74cf43" (UID: "4590771f-ac87-4cbc-86c3-a069ca74cf43"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:01:20 crc kubenswrapper[4802]: I1206 04:01:20.550600 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4590771f-ac87-4cbc-86c3-a069ca74cf43-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4590771f-ac87-4cbc-86c3-a069ca74cf43" (UID: "4590771f-ac87-4cbc-86c3-a069ca74cf43"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:01:20 crc kubenswrapper[4802]: I1206 04:01:20.598483 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nsc65\" (UniqueName: \"kubernetes.io/projected/4590771f-ac87-4cbc-86c3-a069ca74cf43-kube-api-access-nsc65\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:20 crc kubenswrapper[4802]: I1206 04:01:20.598522 4802 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4590771f-ac87-4cbc-86c3-a069ca74cf43-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:20 crc kubenswrapper[4802]: I1206 04:01:20.598536 4802 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4590771f-ac87-4cbc-86c3-a069ca74cf43-logs\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:20 crc kubenswrapper[4802]: I1206 04:01:20.598549 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4590771f-ac87-4cbc-86c3-a069ca74cf43-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:20 crc kubenswrapper[4802]: I1206 04:01:20.639448 4802 generic.go:334] "Generic (PLEG): container finished" podID="4590771f-ac87-4cbc-86c3-a069ca74cf43" containerID="19b6365b8d16b8d17dbc5fbb16ed7a0235a0eaf366568c90d1010f400bd60d1c" exitCode=0 Dec 06 04:01:20 crc kubenswrapper[4802]: I1206 04:01:20.639477 4802 generic.go:334] "Generic (PLEG): container finished" podID="4590771f-ac87-4cbc-86c3-a069ca74cf43" containerID="c1d7f19deb9c8d205fc968f284ac68e9e1785fbaa425bc4a824ef166badbf78b" exitCode=143 Dec 06 04:01:20 crc kubenswrapper[4802]: I1206 04:01:20.639493 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 04:01:20 crc kubenswrapper[4802]: I1206 04:01:20.639537 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4590771f-ac87-4cbc-86c3-a069ca74cf43","Type":"ContainerDied","Data":"19b6365b8d16b8d17dbc5fbb16ed7a0235a0eaf366568c90d1010f400bd60d1c"} Dec 06 04:01:20 crc kubenswrapper[4802]: I1206 04:01:20.639600 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4590771f-ac87-4cbc-86c3-a069ca74cf43","Type":"ContainerDied","Data":"c1d7f19deb9c8d205fc968f284ac68e9e1785fbaa425bc4a824ef166badbf78b"} Dec 06 04:01:20 crc kubenswrapper[4802]: I1206 04:01:20.639619 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4590771f-ac87-4cbc-86c3-a069ca74cf43","Type":"ContainerDied","Data":"5927928cc8d25aa9ea5f16989358ac1e435761cce9ebb659dfdfa4f9556973c4"} Dec 06 04:01:20 crc kubenswrapper[4802]: I1206 04:01:20.639640 4802 scope.go:117] "RemoveContainer" containerID="19b6365b8d16b8d17dbc5fbb16ed7a0235a0eaf366568c90d1010f400bd60d1c" Dec 06 04:01:20 crc kubenswrapper[4802]: I1206 04:01:20.676884 4802 scope.go:117] "RemoveContainer" containerID="c1d7f19deb9c8d205fc968f284ac68e9e1785fbaa425bc4a824ef166badbf78b" Dec 06 04:01:20 crc kubenswrapper[4802]: I1206 04:01:20.682946 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 04:01:20 crc kubenswrapper[4802]: I1206 04:01:20.693862 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 04:01:20 crc kubenswrapper[4802]: I1206 04:01:20.713615 4802 scope.go:117] "RemoveContainer" containerID="19b6365b8d16b8d17dbc5fbb16ed7a0235a0eaf366568c90d1010f400bd60d1c" Dec 06 04:01:20 crc kubenswrapper[4802]: E1206 04:01:20.714480 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19b6365b8d16b8d17dbc5fbb16ed7a0235a0eaf366568c90d1010f400bd60d1c\": container with ID starting with 19b6365b8d16b8d17dbc5fbb16ed7a0235a0eaf366568c90d1010f400bd60d1c not found: ID does not exist" containerID="19b6365b8d16b8d17dbc5fbb16ed7a0235a0eaf366568c90d1010f400bd60d1c" Dec 06 04:01:20 crc kubenswrapper[4802]: I1206 04:01:20.714551 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19b6365b8d16b8d17dbc5fbb16ed7a0235a0eaf366568c90d1010f400bd60d1c"} err="failed to get container status \"19b6365b8d16b8d17dbc5fbb16ed7a0235a0eaf366568c90d1010f400bd60d1c\": rpc error: code = NotFound desc = could not find container \"19b6365b8d16b8d17dbc5fbb16ed7a0235a0eaf366568c90d1010f400bd60d1c\": container with ID starting with 19b6365b8d16b8d17dbc5fbb16ed7a0235a0eaf366568c90d1010f400bd60d1c not found: ID does not exist" Dec 06 04:01:20 crc kubenswrapper[4802]: I1206 04:01:20.714576 4802 scope.go:117] "RemoveContainer" containerID="c1d7f19deb9c8d205fc968f284ac68e9e1785fbaa425bc4a824ef166badbf78b" Dec 06 04:01:20 crc kubenswrapper[4802]: E1206 04:01:20.715089 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1d7f19deb9c8d205fc968f284ac68e9e1785fbaa425bc4a824ef166badbf78b\": container with ID starting with c1d7f19deb9c8d205fc968f284ac68e9e1785fbaa425bc4a824ef166badbf78b not found: ID does not exist" containerID="c1d7f19deb9c8d205fc968f284ac68e9e1785fbaa425bc4a824ef166badbf78b" Dec 06 04:01:20 crc kubenswrapper[4802]: I1206 04:01:20.715149 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1d7f19deb9c8d205fc968f284ac68e9e1785fbaa425bc4a824ef166badbf78b"} err="failed to get container status \"c1d7f19deb9c8d205fc968f284ac68e9e1785fbaa425bc4a824ef166badbf78b\": rpc error: code = NotFound desc = could not find container \"c1d7f19deb9c8d205fc968f284ac68e9e1785fbaa425bc4a824ef166badbf78b\": container with ID starting with c1d7f19deb9c8d205fc968f284ac68e9e1785fbaa425bc4a824ef166badbf78b not found: ID does not exist" Dec 06 04:01:20 crc kubenswrapper[4802]: I1206 04:01:20.715185 4802 scope.go:117] "RemoveContainer" containerID="19b6365b8d16b8d17dbc5fbb16ed7a0235a0eaf366568c90d1010f400bd60d1c" Dec 06 04:01:20 crc kubenswrapper[4802]: I1206 04:01:20.715561 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19b6365b8d16b8d17dbc5fbb16ed7a0235a0eaf366568c90d1010f400bd60d1c"} err="failed to get container status \"19b6365b8d16b8d17dbc5fbb16ed7a0235a0eaf366568c90d1010f400bd60d1c\": rpc error: code = NotFound desc = could not find container \"19b6365b8d16b8d17dbc5fbb16ed7a0235a0eaf366568c90d1010f400bd60d1c\": container with ID starting with 19b6365b8d16b8d17dbc5fbb16ed7a0235a0eaf366568c90d1010f400bd60d1c not found: ID does not exist" Dec 06 04:01:20 crc kubenswrapper[4802]: I1206 04:01:20.715594 4802 scope.go:117] "RemoveContainer" containerID="c1d7f19deb9c8d205fc968f284ac68e9e1785fbaa425bc4a824ef166badbf78b" Dec 06 04:01:20 crc kubenswrapper[4802]: I1206 04:01:20.715821 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1d7f19deb9c8d205fc968f284ac68e9e1785fbaa425bc4a824ef166badbf78b"} err="failed to get container status \"c1d7f19deb9c8d205fc968f284ac68e9e1785fbaa425bc4a824ef166badbf78b\": rpc error: code = NotFound desc = could not find container \"c1d7f19deb9c8d205fc968f284ac68e9e1785fbaa425bc4a824ef166badbf78b\": container with ID starting with c1d7f19deb9c8d205fc968f284ac68e9e1785fbaa425bc4a824ef166badbf78b not found: ID does not exist" Dec 06 04:01:20 crc kubenswrapper[4802]: I1206 04:01:20.724472 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 06 04:01:20 crc kubenswrapper[4802]: E1206 04:01:20.725099 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4590771f-ac87-4cbc-86c3-a069ca74cf43" containerName="nova-metadata-metadata" Dec 06 04:01:20 crc kubenswrapper[4802]: I1206 04:01:20.725198 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="4590771f-ac87-4cbc-86c3-a069ca74cf43" containerName="nova-metadata-metadata" Dec 06 04:01:20 crc kubenswrapper[4802]: E1206 04:01:20.725296 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4590771f-ac87-4cbc-86c3-a069ca74cf43" containerName="nova-metadata-log" Dec 06 04:01:20 crc kubenswrapper[4802]: I1206 04:01:20.725366 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="4590771f-ac87-4cbc-86c3-a069ca74cf43" containerName="nova-metadata-log" Dec 06 04:01:20 crc kubenswrapper[4802]: I1206 04:01:20.725590 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="4590771f-ac87-4cbc-86c3-a069ca74cf43" containerName="nova-metadata-log" Dec 06 04:01:20 crc kubenswrapper[4802]: I1206 04:01:20.725663 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="4590771f-ac87-4cbc-86c3-a069ca74cf43" containerName="nova-metadata-metadata" Dec 06 04:01:20 crc kubenswrapper[4802]: I1206 04:01:20.727022 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 04:01:20 crc kubenswrapper[4802]: I1206 04:01:20.730874 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 04:01:20 crc kubenswrapper[4802]: I1206 04:01:20.732682 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 06 04:01:20 crc kubenswrapper[4802]: I1206 04:01:20.733391 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 06 04:01:20 crc kubenswrapper[4802]: I1206 04:01:20.904365 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2\") " pod="openstack/nova-metadata-0" Dec 06 04:01:20 crc kubenswrapper[4802]: I1206 04:01:20.904412 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2\") " pod="openstack/nova-metadata-0" Dec 06 04:01:20 crc kubenswrapper[4802]: I1206 04:01:20.904458 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gls7m\" (UniqueName: \"kubernetes.io/projected/82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2-kube-api-access-gls7m\") pod \"nova-metadata-0\" (UID: \"82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2\") " pod="openstack/nova-metadata-0" Dec 06 04:01:20 crc kubenswrapper[4802]: I1206 04:01:20.904500 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2-logs\") pod \"nova-metadata-0\" (UID: \"82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2\") " pod="openstack/nova-metadata-0" Dec 06 04:01:20 crc kubenswrapper[4802]: I1206 04:01:20.904521 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2-config-data\") pod \"nova-metadata-0\" (UID: \"82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2\") " pod="openstack/nova-metadata-0" Dec 06 04:01:21 crc kubenswrapper[4802]: I1206 04:01:21.006947 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2\") " pod="openstack/nova-metadata-0" Dec 06 04:01:21 crc kubenswrapper[4802]: I1206 04:01:21.006998 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2\") " pod="openstack/nova-metadata-0" Dec 06 04:01:21 crc kubenswrapper[4802]: I1206 04:01:21.007045 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gls7m\" (UniqueName: \"kubernetes.io/projected/82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2-kube-api-access-gls7m\") pod \"nova-metadata-0\" (UID: \"82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2\") " pod="openstack/nova-metadata-0" Dec 06 04:01:21 crc kubenswrapper[4802]: I1206 04:01:21.007084 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2-logs\") pod \"nova-metadata-0\" (UID: \"82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2\") " pod="openstack/nova-metadata-0" Dec 06 04:01:21 crc kubenswrapper[4802]: I1206 04:01:21.007105 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2-config-data\") pod \"nova-metadata-0\" (UID: \"82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2\") " pod="openstack/nova-metadata-0" Dec 06 04:01:21 crc kubenswrapper[4802]: I1206 04:01:21.008287 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2-logs\") pod \"nova-metadata-0\" (UID: \"82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2\") " pod="openstack/nova-metadata-0" Dec 06 04:01:21 crc kubenswrapper[4802]: I1206 04:01:21.011147 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2\") " pod="openstack/nova-metadata-0" Dec 06 04:01:21 crc kubenswrapper[4802]: I1206 04:01:21.011400 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2-config-data\") pod \"nova-metadata-0\" (UID: \"82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2\") " pod="openstack/nova-metadata-0" Dec 06 04:01:21 crc kubenswrapper[4802]: I1206 04:01:21.022256 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2\") " pod="openstack/nova-metadata-0" Dec 06 04:01:21 crc kubenswrapper[4802]: I1206 04:01:21.024859 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gls7m\" (UniqueName: \"kubernetes.io/projected/82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2-kube-api-access-gls7m\") pod \"nova-metadata-0\" (UID: \"82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2\") " pod="openstack/nova-metadata-0" Dec 06 04:01:21 crc kubenswrapper[4802]: I1206 04:01:21.056620 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 04:01:21 crc kubenswrapper[4802]: I1206 04:01:21.462541 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4590771f-ac87-4cbc-86c3-a069ca74cf43" path="/var/lib/kubelet/pods/4590771f-ac87-4cbc-86c3-a069ca74cf43/volumes" Dec 06 04:01:21 crc kubenswrapper[4802]: I1206 04:01:21.529369 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 04:01:21 crc kubenswrapper[4802]: I1206 04:01:21.653466 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2","Type":"ContainerStarted","Data":"5d3e4e4a075f7d6eecf3bde4c5aa9801bb895ca436d2b35d2e5b9bb85972b5cd"} Dec 06 04:01:22 crc kubenswrapper[4802]: I1206 04:01:22.682272 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2","Type":"ContainerStarted","Data":"79fc6c27474d5f7851d99ee13c14be384c15c465af9da4f56b45151fc52221a5"} Dec 06 04:01:22 crc kubenswrapper[4802]: I1206 04:01:22.682641 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2","Type":"ContainerStarted","Data":"db23039eb210af8755e0d9c3eead68049e65c14daae0069faf5d95ca6b676b98"} Dec 06 04:01:22 crc kubenswrapper[4802]: I1206 04:01:22.684637 4802 generic.go:334] "Generic (PLEG): container finished" podID="bb7d4cc2-ea97-4675-bc65-8803fa1e3708" containerID="8e649e34b5ec7777e026185ed91acf6eb44aeaded3f44a8818c9d17502eb366d" exitCode=0 Dec 06 04:01:22 crc kubenswrapper[4802]: I1206 04:01:22.684783 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-kx676" event={"ID":"bb7d4cc2-ea97-4675-bc65-8803fa1e3708","Type":"ContainerDied","Data":"8e649e34b5ec7777e026185ed91acf6eb44aeaded3f44a8818c9d17502eb366d"} Dec 06 04:01:22 crc kubenswrapper[4802]: I1206 04:01:22.707121 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.707091692 podStartE2EDuration="2.707091692s" podCreationTimestamp="2025-12-06 04:01:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:01:22.702110065 +0000 UTC m=+1275.574019217" watchObservedRunningTime="2025-12-06 04:01:22.707091692 +0000 UTC m=+1275.579000864" Dec 06 04:01:23 crc kubenswrapper[4802]: I1206 04:01:23.694816 4802 generic.go:334] "Generic (PLEG): container finished" podID="735dec51-8659-44bf-bd0b-afd1a1d78d86" containerID="3a7525fcbc73d3441910081d90724b15d3f5451db8ed85974e99baea473d1247" exitCode=0 Dec 06 04:01:23 crc kubenswrapper[4802]: I1206 04:01:23.694875 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-5sl8p" event={"ID":"735dec51-8659-44bf-bd0b-afd1a1d78d86","Type":"ContainerDied","Data":"3a7525fcbc73d3441910081d90724b15d3f5451db8ed85974e99baea473d1247"} Dec 06 04:01:23 crc kubenswrapper[4802]: I1206 04:01:23.765202 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 06 04:01:23 crc kubenswrapper[4802]: I1206 04:01:23.765268 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 06 04:01:23 crc kubenswrapper[4802]: I1206 04:01:23.910987 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 06 04:01:24 crc kubenswrapper[4802]: I1206 04:01:24.119776 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 06 04:01:24 crc kubenswrapper[4802]: I1206 04:01:24.121292 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 06 04:01:24 crc kubenswrapper[4802]: I1206 04:01:24.133172 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-568d7fd7cf-246lw" Dec 06 04:01:24 crc kubenswrapper[4802]: I1206 04:01:24.138399 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-kx676" Dec 06 04:01:24 crc kubenswrapper[4802]: I1206 04:01:24.166069 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 06 04:01:24 crc kubenswrapper[4802]: I1206 04:01:24.230673 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-688b9f5b49-vxlxf"] Dec 06 04:01:24 crc kubenswrapper[4802]: I1206 04:01:24.231029 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-688b9f5b49-vxlxf" podUID="aa11f460-8357-4771-9941-0d1682ed0b74" containerName="dnsmasq-dns" containerID="cri-o://97f98ebf726b3b17ee80d198485eb9354459be38d160912c70ba7b948dd66281" gracePeriod=10 Dec 06 04:01:24 crc kubenswrapper[4802]: I1206 04:01:24.272324 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bbpx8\" (UniqueName: \"kubernetes.io/projected/bb7d4cc2-ea97-4675-bc65-8803fa1e3708-kube-api-access-bbpx8\") pod \"bb7d4cc2-ea97-4675-bc65-8803fa1e3708\" (UID: \"bb7d4cc2-ea97-4675-bc65-8803fa1e3708\") " Dec 06 04:01:24 crc kubenswrapper[4802]: I1206 04:01:24.272396 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb7d4cc2-ea97-4675-bc65-8803fa1e3708-combined-ca-bundle\") pod \"bb7d4cc2-ea97-4675-bc65-8803fa1e3708\" (UID: \"bb7d4cc2-ea97-4675-bc65-8803fa1e3708\") " Dec 06 04:01:24 crc kubenswrapper[4802]: I1206 04:01:24.272477 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb7d4cc2-ea97-4675-bc65-8803fa1e3708-scripts\") pod \"bb7d4cc2-ea97-4675-bc65-8803fa1e3708\" (UID: \"bb7d4cc2-ea97-4675-bc65-8803fa1e3708\") " Dec 06 04:01:24 crc kubenswrapper[4802]: I1206 04:01:24.272564 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb7d4cc2-ea97-4675-bc65-8803fa1e3708-config-data\") pod \"bb7d4cc2-ea97-4675-bc65-8803fa1e3708\" (UID: \"bb7d4cc2-ea97-4675-bc65-8803fa1e3708\") " Dec 06 04:01:24 crc kubenswrapper[4802]: I1206 04:01:24.291738 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb7d4cc2-ea97-4675-bc65-8803fa1e3708-scripts" (OuterVolumeSpecName: "scripts") pod "bb7d4cc2-ea97-4675-bc65-8803fa1e3708" (UID: "bb7d4cc2-ea97-4675-bc65-8803fa1e3708"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:01:24 crc kubenswrapper[4802]: I1206 04:01:24.296205 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb7d4cc2-ea97-4675-bc65-8803fa1e3708-kube-api-access-bbpx8" (OuterVolumeSpecName: "kube-api-access-bbpx8") pod "bb7d4cc2-ea97-4675-bc65-8803fa1e3708" (UID: "bb7d4cc2-ea97-4675-bc65-8803fa1e3708"). InnerVolumeSpecName "kube-api-access-bbpx8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:01:24 crc kubenswrapper[4802]: I1206 04:01:24.327512 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb7d4cc2-ea97-4675-bc65-8803fa1e3708-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bb7d4cc2-ea97-4675-bc65-8803fa1e3708" (UID: "bb7d4cc2-ea97-4675-bc65-8803fa1e3708"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:01:24 crc kubenswrapper[4802]: I1206 04:01:24.340401 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb7d4cc2-ea97-4675-bc65-8803fa1e3708-config-data" (OuterVolumeSpecName: "config-data") pod "bb7d4cc2-ea97-4675-bc65-8803fa1e3708" (UID: "bb7d4cc2-ea97-4675-bc65-8803fa1e3708"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:01:24 crc kubenswrapper[4802]: I1206 04:01:24.378459 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb7d4cc2-ea97-4675-bc65-8803fa1e3708-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:24 crc kubenswrapper[4802]: I1206 04:01:24.378498 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bbpx8\" (UniqueName: \"kubernetes.io/projected/bb7d4cc2-ea97-4675-bc65-8803fa1e3708-kube-api-access-bbpx8\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:24 crc kubenswrapper[4802]: I1206 04:01:24.378514 4802 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb7d4cc2-ea97-4675-bc65-8803fa1e3708-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:24 crc kubenswrapper[4802]: I1206 04:01:24.378526 4802 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb7d4cc2-ea97-4675-bc65-8803fa1e3708-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:24 crc kubenswrapper[4802]: I1206 04:01:24.731608 4802 generic.go:334] "Generic (PLEG): container finished" podID="aa11f460-8357-4771-9941-0d1682ed0b74" containerID="97f98ebf726b3b17ee80d198485eb9354459be38d160912c70ba7b948dd66281" exitCode=0 Dec 06 04:01:24 crc kubenswrapper[4802]: I1206 04:01:24.732153 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688b9f5b49-vxlxf" event={"ID":"aa11f460-8357-4771-9941-0d1682ed0b74","Type":"ContainerDied","Data":"97f98ebf726b3b17ee80d198485eb9354459be38d160912c70ba7b948dd66281"} Dec 06 04:01:24 crc kubenswrapper[4802]: I1206 04:01:24.741038 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-kx676" Dec 06 04:01:24 crc kubenswrapper[4802]: I1206 04:01:24.741136 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-kx676" event={"ID":"bb7d4cc2-ea97-4675-bc65-8803fa1e3708","Type":"ContainerDied","Data":"66747415eeef492e8b0853aca16ef0e036a53c19c47b149425105d835e5e55d3"} Dec 06 04:01:24 crc kubenswrapper[4802]: I1206 04:01:24.741178 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="66747415eeef492e8b0853aca16ef0e036a53c19c47b149425105d835e5e55d3" Dec 06 04:01:24 crc kubenswrapper[4802]: I1206 04:01:24.815946 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688b9f5b49-vxlxf" Dec 06 04:01:24 crc kubenswrapper[4802]: I1206 04:01:24.855889 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="05ec15f4-9b81-4585-adfd-c749c7b12c38" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.193:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 04:01:24 crc kubenswrapper[4802]: I1206 04:01:24.856189 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="05ec15f4-9b81-4585-adfd-c749c7b12c38" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.193:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 04:01:24 crc kubenswrapper[4802]: I1206 04:01:24.871094 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 06 04:01:24 crc kubenswrapper[4802]: I1206 04:01:24.877937 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 06 04:01:24 crc kubenswrapper[4802]: I1206 04:01:24.878119 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="05ec15f4-9b81-4585-adfd-c749c7b12c38" containerName="nova-api-log" containerID="cri-o://2dc54ae8aecbd3584f1f88ec065879db6466e3d5f5eae5ca7bfc561117d67fa2" gracePeriod=30 Dec 06 04:01:24 crc kubenswrapper[4802]: I1206 04:01:24.878223 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="05ec15f4-9b81-4585-adfd-c749c7b12c38" containerName="nova-api-api" containerID="cri-o://c8fb2250bcb32700a9bcaa59793b36842ad24f7af491161c04f30f0dcb9dbd1f" gracePeriod=30 Dec 06 04:01:24 crc kubenswrapper[4802]: I1206 04:01:24.901650 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 04:01:24 crc kubenswrapper[4802]: I1206 04:01:24.928662 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 04:01:24 crc kubenswrapper[4802]: I1206 04:01:24.928875 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2" containerName="nova-metadata-log" containerID="cri-o://db23039eb210af8755e0d9c3eead68049e65c14daae0069faf5d95ca6b676b98" gracePeriod=30 Dec 06 04:01:24 crc kubenswrapper[4802]: I1206 04:01:24.929364 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2" containerName="nova-metadata-metadata" containerID="cri-o://79fc6c27474d5f7851d99ee13c14be384c15c465af9da4f56b45151fc52221a5" gracePeriod=30 Dec 06 04:01:24 crc kubenswrapper[4802]: I1206 04:01:24.992055 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa11f460-8357-4771-9941-0d1682ed0b74-config\") pod \"aa11f460-8357-4771-9941-0d1682ed0b74\" (UID: \"aa11f460-8357-4771-9941-0d1682ed0b74\") " Dec 06 04:01:24 crc kubenswrapper[4802]: I1206 04:01:24.992190 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa11f460-8357-4771-9941-0d1682ed0b74-ovsdbserver-nb\") pod \"aa11f460-8357-4771-9941-0d1682ed0b74\" (UID: \"aa11f460-8357-4771-9941-0d1682ed0b74\") " Dec 06 04:01:24 crc kubenswrapper[4802]: I1206 04:01:24.992282 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa11f460-8357-4771-9941-0d1682ed0b74-ovsdbserver-sb\") pod \"aa11f460-8357-4771-9941-0d1682ed0b74\" (UID: \"aa11f460-8357-4771-9941-0d1682ed0b74\") " Dec 06 04:01:24 crc kubenswrapper[4802]: I1206 04:01:24.992401 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa11f460-8357-4771-9941-0d1682ed0b74-dns-svc\") pod \"aa11f460-8357-4771-9941-0d1682ed0b74\" (UID: \"aa11f460-8357-4771-9941-0d1682ed0b74\") " Dec 06 04:01:24 crc kubenswrapper[4802]: I1206 04:01:24.992469 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-67sbh\" (UniqueName: \"kubernetes.io/projected/aa11f460-8357-4771-9941-0d1682ed0b74-kube-api-access-67sbh\") pod \"aa11f460-8357-4771-9941-0d1682ed0b74\" (UID: \"aa11f460-8357-4771-9941-0d1682ed0b74\") " Dec 06 04:01:24 crc kubenswrapper[4802]: I1206 04:01:24.992549 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aa11f460-8357-4771-9941-0d1682ed0b74-dns-swift-storage-0\") pod \"aa11f460-8357-4771-9941-0d1682ed0b74\" (UID: \"aa11f460-8357-4771-9941-0d1682ed0b74\") " Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.000021 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa11f460-8357-4771-9941-0d1682ed0b74-kube-api-access-67sbh" (OuterVolumeSpecName: "kube-api-access-67sbh") pod "aa11f460-8357-4771-9941-0d1682ed0b74" (UID: "aa11f460-8357-4771-9941-0d1682ed0b74"). InnerVolumeSpecName "kube-api-access-67sbh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.085618 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa11f460-8357-4771-9941-0d1682ed0b74-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "aa11f460-8357-4771-9941-0d1682ed0b74" (UID: "aa11f460-8357-4771-9941-0d1682ed0b74"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.090251 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa11f460-8357-4771-9941-0d1682ed0b74-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "aa11f460-8357-4771-9941-0d1682ed0b74" (UID: "aa11f460-8357-4771-9941-0d1682ed0b74"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.094854 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa11f460-8357-4771-9941-0d1682ed0b74-config" (OuterVolumeSpecName: "config") pod "aa11f460-8357-4771-9941-0d1682ed0b74" (UID: "aa11f460-8357-4771-9941-0d1682ed0b74"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.095078 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-67sbh\" (UniqueName: \"kubernetes.io/projected/aa11f460-8357-4771-9941-0d1682ed0b74-kube-api-access-67sbh\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.095106 4802 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aa11f460-8357-4771-9941-0d1682ed0b74-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.095115 4802 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa11f460-8357-4771-9941-0d1682ed0b74-config\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.095124 4802 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa11f460-8357-4771-9941-0d1682ed0b74-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.121263 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa11f460-8357-4771-9941-0d1682ed0b74-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "aa11f460-8357-4771-9941-0d1682ed0b74" (UID: "aa11f460-8357-4771-9941-0d1682ed0b74"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.141786 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa11f460-8357-4771-9941-0d1682ed0b74-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "aa11f460-8357-4771-9941-0d1682ed0b74" (UID: "aa11f460-8357-4771-9941-0d1682ed0b74"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.198508 4802 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa11f460-8357-4771-9941-0d1682ed0b74-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.198540 4802 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa11f460-8357-4771-9941-0d1682ed0b74-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.352717 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-5sl8p" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.493816 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.507036 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/735dec51-8659-44bf-bd0b-afd1a1d78d86-scripts\") pod \"735dec51-8659-44bf-bd0b-afd1a1d78d86\" (UID: \"735dec51-8659-44bf-bd0b-afd1a1d78d86\") " Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.507360 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/735dec51-8659-44bf-bd0b-afd1a1d78d86-combined-ca-bundle\") pod \"735dec51-8659-44bf-bd0b-afd1a1d78d86\" (UID: \"735dec51-8659-44bf-bd0b-afd1a1d78d86\") " Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.507898 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/735dec51-8659-44bf-bd0b-afd1a1d78d86-config-data\") pod \"735dec51-8659-44bf-bd0b-afd1a1d78d86\" (UID: \"735dec51-8659-44bf-bd0b-afd1a1d78d86\") " Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.508066 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bxkwc\" (UniqueName: \"kubernetes.io/projected/735dec51-8659-44bf-bd0b-afd1a1d78d86-kube-api-access-bxkwc\") pod \"735dec51-8659-44bf-bd0b-afd1a1d78d86\" (UID: \"735dec51-8659-44bf-bd0b-afd1a1d78d86\") " Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.539576 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/735dec51-8659-44bf-bd0b-afd1a1d78d86-config-data" (OuterVolumeSpecName: "config-data") pod "735dec51-8659-44bf-bd0b-afd1a1d78d86" (UID: "735dec51-8659-44bf-bd0b-afd1a1d78d86"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.549016 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/735dec51-8659-44bf-bd0b-afd1a1d78d86-scripts" (OuterVolumeSpecName: "scripts") pod "735dec51-8659-44bf-bd0b-afd1a1d78d86" (UID: "735dec51-8659-44bf-bd0b-afd1a1d78d86"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.549330 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/735dec51-8659-44bf-bd0b-afd1a1d78d86-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "735dec51-8659-44bf-bd0b-afd1a1d78d86" (UID: "735dec51-8659-44bf-bd0b-afd1a1d78d86"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.562543 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/735dec51-8659-44bf-bd0b-afd1a1d78d86-kube-api-access-bxkwc" (OuterVolumeSpecName: "kube-api-access-bxkwc") pod "735dec51-8659-44bf-bd0b-afd1a1d78d86" (UID: "735dec51-8659-44bf-bd0b-afd1a1d78d86"). InnerVolumeSpecName "kube-api-access-bxkwc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.610471 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2-config-data\") pod \"82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2\" (UID: \"82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2\") " Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.610579 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2-combined-ca-bundle\") pod \"82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2\" (UID: \"82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2\") " Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.610670 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2-logs\") pod \"82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2\" (UID: \"82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2\") " Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.610854 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gls7m\" (UniqueName: \"kubernetes.io/projected/82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2-kube-api-access-gls7m\") pod \"82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2\" (UID: \"82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2\") " Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.610890 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2-nova-metadata-tls-certs\") pod \"82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2\" (UID: \"82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2\") " Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.611363 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2-logs" (OuterVolumeSpecName: "logs") pod "82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2" (UID: "82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.611858 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bxkwc\" (UniqueName: \"kubernetes.io/projected/735dec51-8659-44bf-bd0b-afd1a1d78d86-kube-api-access-bxkwc\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.611881 4802 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/735dec51-8659-44bf-bd0b-afd1a1d78d86-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.611893 4802 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/735dec51-8659-44bf-bd0b-afd1a1d78d86-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.611905 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/735dec51-8659-44bf-bd0b-afd1a1d78d86-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.611916 4802 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2-logs\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.619053 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2-kube-api-access-gls7m" (OuterVolumeSpecName: "kube-api-access-gls7m") pod "82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2" (UID: "82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2"). InnerVolumeSpecName "kube-api-access-gls7m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.639023 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2" (UID: "82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.651854 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2-config-data" (OuterVolumeSpecName: "config-data") pod "82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2" (UID: "82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.667454 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2" (UID: "82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.714015 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gls7m\" (UniqueName: \"kubernetes.io/projected/82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2-kube-api-access-gls7m\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.714054 4802 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.714067 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.714079 4802 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.748919 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-5sl8p" event={"ID":"735dec51-8659-44bf-bd0b-afd1a1d78d86","Type":"ContainerDied","Data":"73abe00dafcadcca6e927805a277f8eef49528e5ee0867642b61611f1961be67"} Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.748995 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="73abe00dafcadcca6e927805a277f8eef49528e5ee0867642b61611f1961be67" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.748966 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-5sl8p" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.751438 4802 generic.go:334] "Generic (PLEG): container finished" podID="05ec15f4-9b81-4585-adfd-c749c7b12c38" containerID="2dc54ae8aecbd3584f1f88ec065879db6466e3d5f5eae5ca7bfc561117d67fa2" exitCode=143 Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.751534 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"05ec15f4-9b81-4585-adfd-c749c7b12c38","Type":"ContainerDied","Data":"2dc54ae8aecbd3584f1f88ec065879db6466e3d5f5eae5ca7bfc561117d67fa2"} Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.753435 4802 generic.go:334] "Generic (PLEG): container finished" podID="82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2" containerID="79fc6c27474d5f7851d99ee13c14be384c15c465af9da4f56b45151fc52221a5" exitCode=0 Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.753468 4802 generic.go:334] "Generic (PLEG): container finished" podID="82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2" containerID="db23039eb210af8755e0d9c3eead68049e65c14daae0069faf5d95ca6b676b98" exitCode=143 Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.753474 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2","Type":"ContainerDied","Data":"79fc6c27474d5f7851d99ee13c14be384c15c465af9da4f56b45151fc52221a5"} Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.753510 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.753531 4802 scope.go:117] "RemoveContainer" containerID="79fc6c27474d5f7851d99ee13c14be384c15c465af9da4f56b45151fc52221a5" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.753517 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2","Type":"ContainerDied","Data":"db23039eb210af8755e0d9c3eead68049e65c14daae0069faf5d95ca6b676b98"} Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.753644 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2","Type":"ContainerDied","Data":"5d3e4e4a075f7d6eecf3bde4c5aa9801bb895ca436d2b35d2e5b9bb85972b5cd"} Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.755628 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688b9f5b49-vxlxf" event={"ID":"aa11f460-8357-4771-9941-0d1682ed0b74","Type":"ContainerDied","Data":"04057ba7dda93084cee2e5e79d0b64105810b94c48c817669bcabe78bfc85aeb"} Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.755881 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688b9f5b49-vxlxf" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.788647 4802 scope.go:117] "RemoveContainer" containerID="db23039eb210af8755e0d9c3eead68049e65c14daae0069faf5d95ca6b676b98" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.793361 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-688b9f5b49-vxlxf"] Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.808837 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-688b9f5b49-vxlxf"] Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.815394 4802 scope.go:117] "RemoveContainer" containerID="79fc6c27474d5f7851d99ee13c14be384c15c465af9da4f56b45151fc52221a5" Dec 06 04:01:25 crc kubenswrapper[4802]: E1206 04:01:25.815815 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79fc6c27474d5f7851d99ee13c14be384c15c465af9da4f56b45151fc52221a5\": container with ID starting with 79fc6c27474d5f7851d99ee13c14be384c15c465af9da4f56b45151fc52221a5 not found: ID does not exist" containerID="79fc6c27474d5f7851d99ee13c14be384c15c465af9da4f56b45151fc52221a5" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.815848 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79fc6c27474d5f7851d99ee13c14be384c15c465af9da4f56b45151fc52221a5"} err="failed to get container status \"79fc6c27474d5f7851d99ee13c14be384c15c465af9da4f56b45151fc52221a5\": rpc error: code = NotFound desc = could not find container \"79fc6c27474d5f7851d99ee13c14be384c15c465af9da4f56b45151fc52221a5\": container with ID starting with 79fc6c27474d5f7851d99ee13c14be384c15c465af9da4f56b45151fc52221a5 not found: ID does not exist" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.815869 4802 scope.go:117] "RemoveContainer" containerID="db23039eb210af8755e0d9c3eead68049e65c14daae0069faf5d95ca6b676b98" Dec 06 04:01:25 crc kubenswrapper[4802]: E1206 04:01:25.816076 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db23039eb210af8755e0d9c3eead68049e65c14daae0069faf5d95ca6b676b98\": container with ID starting with db23039eb210af8755e0d9c3eead68049e65c14daae0069faf5d95ca6b676b98 not found: ID does not exist" containerID="db23039eb210af8755e0d9c3eead68049e65c14daae0069faf5d95ca6b676b98" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.816096 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db23039eb210af8755e0d9c3eead68049e65c14daae0069faf5d95ca6b676b98"} err="failed to get container status \"db23039eb210af8755e0d9c3eead68049e65c14daae0069faf5d95ca6b676b98\": rpc error: code = NotFound desc = could not find container \"db23039eb210af8755e0d9c3eead68049e65c14daae0069faf5d95ca6b676b98\": container with ID starting with db23039eb210af8755e0d9c3eead68049e65c14daae0069faf5d95ca6b676b98 not found: ID does not exist" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.816108 4802 scope.go:117] "RemoveContainer" containerID="79fc6c27474d5f7851d99ee13c14be384c15c465af9da4f56b45151fc52221a5" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.816268 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79fc6c27474d5f7851d99ee13c14be384c15c465af9da4f56b45151fc52221a5"} err="failed to get container status \"79fc6c27474d5f7851d99ee13c14be384c15c465af9da4f56b45151fc52221a5\": rpc error: code = NotFound desc = could not find container \"79fc6c27474d5f7851d99ee13c14be384c15c465af9da4f56b45151fc52221a5\": container with ID starting with 79fc6c27474d5f7851d99ee13c14be384c15c465af9da4f56b45151fc52221a5 not found: ID does not exist" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.816281 4802 scope.go:117] "RemoveContainer" containerID="db23039eb210af8755e0d9c3eead68049e65c14daae0069faf5d95ca6b676b98" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.816422 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db23039eb210af8755e0d9c3eead68049e65c14daae0069faf5d95ca6b676b98"} err="failed to get container status \"db23039eb210af8755e0d9c3eead68049e65c14daae0069faf5d95ca6b676b98\": rpc error: code = NotFound desc = could not find container \"db23039eb210af8755e0d9c3eead68049e65c14daae0069faf5d95ca6b676b98\": container with ID starting with db23039eb210af8755e0d9c3eead68049e65c14daae0069faf5d95ca6b676b98 not found: ID does not exist" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.816435 4802 scope.go:117] "RemoveContainer" containerID="97f98ebf726b3b17ee80d198485eb9354459be38d160912c70ba7b948dd66281" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.844223 4802 scope.go:117] "RemoveContainer" containerID="f13ba3f09772ae13d7557ec46f47e01c1e51ec8e4be764cd3a1bd49bf62a18df" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.845033 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 06 04:01:25 crc kubenswrapper[4802]: E1206 04:01:25.845501 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb7d4cc2-ea97-4675-bc65-8803fa1e3708" containerName="nova-manage" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.845518 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb7d4cc2-ea97-4675-bc65-8803fa1e3708" containerName="nova-manage" Dec 06 04:01:25 crc kubenswrapper[4802]: E1206 04:01:25.845545 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa11f460-8357-4771-9941-0d1682ed0b74" containerName="init" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.845553 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa11f460-8357-4771-9941-0d1682ed0b74" containerName="init" Dec 06 04:01:25 crc kubenswrapper[4802]: E1206 04:01:25.845569 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="735dec51-8659-44bf-bd0b-afd1a1d78d86" containerName="nova-cell1-conductor-db-sync" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.845577 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="735dec51-8659-44bf-bd0b-afd1a1d78d86" containerName="nova-cell1-conductor-db-sync" Dec 06 04:01:25 crc kubenswrapper[4802]: E1206 04:01:25.845594 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa11f460-8357-4771-9941-0d1682ed0b74" containerName="dnsmasq-dns" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.845601 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa11f460-8357-4771-9941-0d1682ed0b74" containerName="dnsmasq-dns" Dec 06 04:01:25 crc kubenswrapper[4802]: E1206 04:01:25.845613 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2" containerName="nova-metadata-log" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.845623 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2" containerName="nova-metadata-log" Dec 06 04:01:25 crc kubenswrapper[4802]: E1206 04:01:25.845647 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2" containerName="nova-metadata-metadata" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.845656 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2" containerName="nova-metadata-metadata" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.845898 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2" containerName="nova-metadata-log" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.845924 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="735dec51-8659-44bf-bd0b-afd1a1d78d86" containerName="nova-cell1-conductor-db-sync" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.845935 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2" containerName="nova-metadata-metadata" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.845956 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa11f460-8357-4771-9941-0d1682ed0b74" containerName="dnsmasq-dns" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.845972 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb7d4cc2-ea97-4675-bc65-8803fa1e3708" containerName="nova-manage" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.846733 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.852466 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.866091 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.880144 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.912272 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.926150 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.927882 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.930173 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.930216 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 06 04:01:25 crc kubenswrapper[4802]: I1206 04:01:25.935797 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 04:01:26 crc kubenswrapper[4802]: I1206 04:01:26.019025 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be\") " pod="openstack/nova-metadata-0" Dec 06 04:01:26 crc kubenswrapper[4802]: I1206 04:01:26.019393 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be-config-data\") pod \"nova-metadata-0\" (UID: \"7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be\") " pod="openstack/nova-metadata-0" Dec 06 04:01:26 crc kubenswrapper[4802]: I1206 04:01:26.019492 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaeadc44-2583-4e10-b0d3-15b754927734-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"eaeadc44-2583-4e10-b0d3-15b754927734\") " pod="openstack/nova-cell1-conductor-0" Dec 06 04:01:26 crc kubenswrapper[4802]: I1206 04:01:26.019610 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be-logs\") pod \"nova-metadata-0\" (UID: \"7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be\") " pod="openstack/nova-metadata-0" Dec 06 04:01:26 crc kubenswrapper[4802]: I1206 04:01:26.019835 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be\") " pod="openstack/nova-metadata-0" Dec 06 04:01:26 crc kubenswrapper[4802]: I1206 04:01:26.020096 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktkcf\" (UniqueName: \"kubernetes.io/projected/eaeadc44-2583-4e10-b0d3-15b754927734-kube-api-access-ktkcf\") pod \"nova-cell1-conductor-0\" (UID: \"eaeadc44-2583-4e10-b0d3-15b754927734\") " pod="openstack/nova-cell1-conductor-0" Dec 06 04:01:26 crc kubenswrapper[4802]: I1206 04:01:26.020197 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dcqg\" (UniqueName: \"kubernetes.io/projected/7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be-kube-api-access-2dcqg\") pod \"nova-metadata-0\" (UID: \"7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be\") " pod="openstack/nova-metadata-0" Dec 06 04:01:26 crc kubenswrapper[4802]: I1206 04:01:26.020278 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaeadc44-2583-4e10-b0d3-15b754927734-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"eaeadc44-2583-4e10-b0d3-15b754927734\") " pod="openstack/nova-cell1-conductor-0" Dec 06 04:01:26 crc kubenswrapper[4802]: I1206 04:01:26.121930 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be\") " pod="openstack/nova-metadata-0" Dec 06 04:01:26 crc kubenswrapper[4802]: I1206 04:01:26.121978 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be-config-data\") pod \"nova-metadata-0\" (UID: \"7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be\") " pod="openstack/nova-metadata-0" Dec 06 04:01:26 crc kubenswrapper[4802]: I1206 04:01:26.122001 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaeadc44-2583-4e10-b0d3-15b754927734-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"eaeadc44-2583-4e10-b0d3-15b754927734\") " pod="openstack/nova-cell1-conductor-0" Dec 06 04:01:26 crc kubenswrapper[4802]: I1206 04:01:26.122061 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be-logs\") pod \"nova-metadata-0\" (UID: \"7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be\") " pod="openstack/nova-metadata-0" Dec 06 04:01:26 crc kubenswrapper[4802]: I1206 04:01:26.122093 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be\") " pod="openstack/nova-metadata-0" Dec 06 04:01:26 crc kubenswrapper[4802]: I1206 04:01:26.122120 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktkcf\" (UniqueName: \"kubernetes.io/projected/eaeadc44-2583-4e10-b0d3-15b754927734-kube-api-access-ktkcf\") pod \"nova-cell1-conductor-0\" (UID: \"eaeadc44-2583-4e10-b0d3-15b754927734\") " pod="openstack/nova-cell1-conductor-0" Dec 06 04:01:26 crc kubenswrapper[4802]: I1206 04:01:26.122147 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dcqg\" (UniqueName: \"kubernetes.io/projected/7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be-kube-api-access-2dcqg\") pod \"nova-metadata-0\" (UID: \"7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be\") " pod="openstack/nova-metadata-0" Dec 06 04:01:26 crc kubenswrapper[4802]: I1206 04:01:26.122175 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaeadc44-2583-4e10-b0d3-15b754927734-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"eaeadc44-2583-4e10-b0d3-15b754927734\") " pod="openstack/nova-cell1-conductor-0" Dec 06 04:01:26 crc kubenswrapper[4802]: I1206 04:01:26.122839 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be-logs\") pod \"nova-metadata-0\" (UID: \"7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be\") " pod="openstack/nova-metadata-0" Dec 06 04:01:26 crc kubenswrapper[4802]: I1206 04:01:26.125513 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be\") " pod="openstack/nova-metadata-0" Dec 06 04:01:26 crc kubenswrapper[4802]: I1206 04:01:26.125573 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaeadc44-2583-4e10-b0d3-15b754927734-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"eaeadc44-2583-4e10-b0d3-15b754927734\") " pod="openstack/nova-cell1-conductor-0" Dec 06 04:01:26 crc kubenswrapper[4802]: I1206 04:01:26.126816 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be-config-data\") pod \"nova-metadata-0\" (UID: \"7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be\") " pod="openstack/nova-metadata-0" Dec 06 04:01:26 crc kubenswrapper[4802]: I1206 04:01:26.127291 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be\") " pod="openstack/nova-metadata-0" Dec 06 04:01:26 crc kubenswrapper[4802]: I1206 04:01:26.135549 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaeadc44-2583-4e10-b0d3-15b754927734-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"eaeadc44-2583-4e10-b0d3-15b754927734\") " pod="openstack/nova-cell1-conductor-0" Dec 06 04:01:26 crc kubenswrapper[4802]: I1206 04:01:26.147329 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktkcf\" (UniqueName: \"kubernetes.io/projected/eaeadc44-2583-4e10-b0d3-15b754927734-kube-api-access-ktkcf\") pod \"nova-cell1-conductor-0\" (UID: \"eaeadc44-2583-4e10-b0d3-15b754927734\") " pod="openstack/nova-cell1-conductor-0" Dec 06 04:01:26 crc kubenswrapper[4802]: I1206 04:01:26.154451 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dcqg\" (UniqueName: \"kubernetes.io/projected/7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be-kube-api-access-2dcqg\") pod \"nova-metadata-0\" (UID: \"7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be\") " pod="openstack/nova-metadata-0" Dec 06 04:01:26 crc kubenswrapper[4802]: I1206 04:01:26.175307 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 06 04:01:26 crc kubenswrapper[4802]: I1206 04:01:26.246270 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 04:01:26 crc kubenswrapper[4802]: I1206 04:01:26.665345 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 06 04:01:26 crc kubenswrapper[4802]: W1206 04:01:26.788452 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7655cfe4_be90_4fd1_8bdc_e9bd6aa0c2be.slice/crio-1afe2766859689dc8280a7c0d1c65f783dd603bb4800bbea8e85aaab909a0789 WatchSource:0}: Error finding container 1afe2766859689dc8280a7c0d1c65f783dd603bb4800bbea8e85aaab909a0789: Status 404 returned error can't find the container with id 1afe2766859689dc8280a7c0d1c65f783dd603bb4800bbea8e85aaab909a0789 Dec 06 04:01:26 crc kubenswrapper[4802]: I1206 04:01:26.788464 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 04:01:26 crc kubenswrapper[4802]: I1206 04:01:26.792063 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="71c5fc79-7c68-4406-b02a-6cf325075a7a" containerName="nova-scheduler-scheduler" containerID="cri-o://24648326d0ed4b1db9306783eeae2a39a0b46c7d4973ed6d5ad68bfc74434b67" gracePeriod=30 Dec 06 04:01:26 crc kubenswrapper[4802]: I1206 04:01:26.792414 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"eaeadc44-2583-4e10-b0d3-15b754927734","Type":"ContainerStarted","Data":"82d97162df5d8635b1562d4502f741717415254eb4a56d0c46d0013e5f8b3fb2"} Dec 06 04:01:27 crc kubenswrapper[4802]: I1206 04:01:27.463123 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2" path="/var/lib/kubelet/pods/82d07cd3-a5d6-4f9c-9bb4-db54f654d6a2/volumes" Dec 06 04:01:27 crc kubenswrapper[4802]: I1206 04:01:27.464035 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa11f460-8357-4771-9941-0d1682ed0b74" path="/var/lib/kubelet/pods/aa11f460-8357-4771-9941-0d1682ed0b74/volumes" Dec 06 04:01:27 crc kubenswrapper[4802]: I1206 04:01:27.848071 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"eaeadc44-2583-4e10-b0d3-15b754927734","Type":"ContainerStarted","Data":"ccfd1346d11050e4f9d08856dba18906033cb6fe102f1e4e32747ffe34ff150d"} Dec 06 04:01:27 crc kubenswrapper[4802]: I1206 04:01:27.848183 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 06 04:01:27 crc kubenswrapper[4802]: I1206 04:01:27.861193 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be","Type":"ContainerStarted","Data":"f39aa5495b856420e48e53837f48c1f867617f09aaa212ea86ad61f7409b64b6"} Dec 06 04:01:27 crc kubenswrapper[4802]: I1206 04:01:27.861437 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be","Type":"ContainerStarted","Data":"d10f8eeaa0106a33b52fd1a34f7ab3007212756b42626cc4ae7d14a99433a0aa"} Dec 06 04:01:27 crc kubenswrapper[4802]: I1206 04:01:27.861446 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be","Type":"ContainerStarted","Data":"1afe2766859689dc8280a7c0d1c65f783dd603bb4800bbea8e85aaab909a0789"} Dec 06 04:01:27 crc kubenswrapper[4802]: I1206 04:01:27.903202 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.903176523 podStartE2EDuration="2.903176523s" podCreationTimestamp="2025-12-06 04:01:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:01:27.899636846 +0000 UTC m=+1280.771545998" watchObservedRunningTime="2025-12-06 04:01:27.903176523 +0000 UTC m=+1280.775085675" Dec 06 04:01:27 crc kubenswrapper[4802]: I1206 04:01:27.911409 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.911383758 podStartE2EDuration="2.911383758s" podCreationTimestamp="2025-12-06 04:01:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:01:27.880434941 +0000 UTC m=+1280.752344093" watchObservedRunningTime="2025-12-06 04:01:27.911383758 +0000 UTC m=+1280.783292910" Dec 06 04:01:29 crc kubenswrapper[4802]: E1206 04:01:29.120918 4802 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="24648326d0ed4b1db9306783eeae2a39a0b46c7d4973ed6d5ad68bfc74434b67" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 06 04:01:29 crc kubenswrapper[4802]: E1206 04:01:29.122440 4802 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="24648326d0ed4b1db9306783eeae2a39a0b46c7d4973ed6d5ad68bfc74434b67" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 06 04:01:29 crc kubenswrapper[4802]: E1206 04:01:29.124059 4802 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="24648326d0ed4b1db9306783eeae2a39a0b46c7d4973ed6d5ad68bfc74434b67" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 06 04:01:29 crc kubenswrapper[4802]: E1206 04:01:29.124092 4802 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="71c5fc79-7c68-4406-b02a-6cf325075a7a" containerName="nova-scheduler-scheduler" Dec 06 04:01:30 crc kubenswrapper[4802]: I1206 04:01:30.767186 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 06 04:01:30 crc kubenswrapper[4802]: I1206 04:01:30.877900 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 04:01:30 crc kubenswrapper[4802]: I1206 04:01:30.893145 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 04:01:30 crc kubenswrapper[4802]: I1206 04:01:30.906559 4802 generic.go:334] "Generic (PLEG): container finished" podID="05ec15f4-9b81-4585-adfd-c749c7b12c38" containerID="c8fb2250bcb32700a9bcaa59793b36842ad24f7af491161c04f30f0dcb9dbd1f" exitCode=0 Dec 06 04:01:30 crc kubenswrapper[4802]: I1206 04:01:30.906654 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"05ec15f4-9b81-4585-adfd-c749c7b12c38","Type":"ContainerDied","Data":"c8fb2250bcb32700a9bcaa59793b36842ad24f7af491161c04f30f0dcb9dbd1f"} Dec 06 04:01:30 crc kubenswrapper[4802]: I1206 04:01:30.906965 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"05ec15f4-9b81-4585-adfd-c749c7b12c38","Type":"ContainerDied","Data":"e977cdc34bb1a7fb53f49652f283594fa977f41520a14d51526cb52935405b1f"} Dec 06 04:01:30 crc kubenswrapper[4802]: I1206 04:01:30.906706 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 04:01:30 crc kubenswrapper[4802]: I1206 04:01:30.907031 4802 scope.go:117] "RemoveContainer" containerID="c8fb2250bcb32700a9bcaa59793b36842ad24f7af491161c04f30f0dcb9dbd1f" Dec 06 04:01:30 crc kubenswrapper[4802]: I1206 04:01:30.909622 4802 generic.go:334] "Generic (PLEG): container finished" podID="71c5fc79-7c68-4406-b02a-6cf325075a7a" containerID="24648326d0ed4b1db9306783eeae2a39a0b46c7d4973ed6d5ad68bfc74434b67" exitCode=0 Dec 06 04:01:30 crc kubenswrapper[4802]: I1206 04:01:30.909867 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"71c5fc79-7c68-4406-b02a-6cf325075a7a","Type":"ContainerDied","Data":"24648326d0ed4b1db9306783eeae2a39a0b46c7d4973ed6d5ad68bfc74434b67"} Dec 06 04:01:30 crc kubenswrapper[4802]: I1206 04:01:30.909984 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"71c5fc79-7c68-4406-b02a-6cf325075a7a","Type":"ContainerDied","Data":"475402fe59aa0ad464150ed550eefa2db54290355afdce7f5fc3f71dc2f5873b"} Dec 06 04:01:30 crc kubenswrapper[4802]: I1206 04:01:30.910371 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 04:01:30 crc kubenswrapper[4802]: I1206 04:01:30.947573 4802 scope.go:117] "RemoveContainer" containerID="2dc54ae8aecbd3584f1f88ec065879db6466e3d5f5eae5ca7bfc561117d67fa2" Dec 06 04:01:30 crc kubenswrapper[4802]: I1206 04:01:30.966883 4802 scope.go:117] "RemoveContainer" containerID="c8fb2250bcb32700a9bcaa59793b36842ad24f7af491161c04f30f0dcb9dbd1f" Dec 06 04:01:30 crc kubenswrapper[4802]: E1206 04:01:30.967410 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8fb2250bcb32700a9bcaa59793b36842ad24f7af491161c04f30f0dcb9dbd1f\": container with ID starting with c8fb2250bcb32700a9bcaa59793b36842ad24f7af491161c04f30f0dcb9dbd1f not found: ID does not exist" containerID="c8fb2250bcb32700a9bcaa59793b36842ad24f7af491161c04f30f0dcb9dbd1f" Dec 06 04:01:30 crc kubenswrapper[4802]: I1206 04:01:30.967469 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8fb2250bcb32700a9bcaa59793b36842ad24f7af491161c04f30f0dcb9dbd1f"} err="failed to get container status \"c8fb2250bcb32700a9bcaa59793b36842ad24f7af491161c04f30f0dcb9dbd1f\": rpc error: code = NotFound desc = could not find container \"c8fb2250bcb32700a9bcaa59793b36842ad24f7af491161c04f30f0dcb9dbd1f\": container with ID starting with c8fb2250bcb32700a9bcaa59793b36842ad24f7af491161c04f30f0dcb9dbd1f not found: ID does not exist" Dec 06 04:01:30 crc kubenswrapper[4802]: I1206 04:01:30.967503 4802 scope.go:117] "RemoveContainer" containerID="2dc54ae8aecbd3584f1f88ec065879db6466e3d5f5eae5ca7bfc561117d67fa2" Dec 06 04:01:30 crc kubenswrapper[4802]: E1206 04:01:30.967878 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2dc54ae8aecbd3584f1f88ec065879db6466e3d5f5eae5ca7bfc561117d67fa2\": container with ID starting with 2dc54ae8aecbd3584f1f88ec065879db6466e3d5f5eae5ca7bfc561117d67fa2 not found: ID does not exist" containerID="2dc54ae8aecbd3584f1f88ec065879db6466e3d5f5eae5ca7bfc561117d67fa2" Dec 06 04:01:30 crc kubenswrapper[4802]: I1206 04:01:30.967899 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2dc54ae8aecbd3584f1f88ec065879db6466e3d5f5eae5ca7bfc561117d67fa2"} err="failed to get container status \"2dc54ae8aecbd3584f1f88ec065879db6466e3d5f5eae5ca7bfc561117d67fa2\": rpc error: code = NotFound desc = could not find container \"2dc54ae8aecbd3584f1f88ec065879db6466e3d5f5eae5ca7bfc561117d67fa2\": container with ID starting with 2dc54ae8aecbd3584f1f88ec065879db6466e3d5f5eae5ca7bfc561117d67fa2 not found: ID does not exist" Dec 06 04:01:30 crc kubenswrapper[4802]: I1206 04:01:30.967912 4802 scope.go:117] "RemoveContainer" containerID="24648326d0ed4b1db9306783eeae2a39a0b46c7d4973ed6d5ad68bfc74434b67" Dec 06 04:01:30 crc kubenswrapper[4802]: I1206 04:01:30.992167 4802 scope.go:117] "RemoveContainer" containerID="24648326d0ed4b1db9306783eeae2a39a0b46c7d4973ed6d5ad68bfc74434b67" Dec 06 04:01:30 crc kubenswrapper[4802]: E1206 04:01:30.992676 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24648326d0ed4b1db9306783eeae2a39a0b46c7d4973ed6d5ad68bfc74434b67\": container with ID starting with 24648326d0ed4b1db9306783eeae2a39a0b46c7d4973ed6d5ad68bfc74434b67 not found: ID does not exist" containerID="24648326d0ed4b1db9306783eeae2a39a0b46c7d4973ed6d5ad68bfc74434b67" Dec 06 04:01:30 crc kubenswrapper[4802]: I1206 04:01:30.992718 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24648326d0ed4b1db9306783eeae2a39a0b46c7d4973ed6d5ad68bfc74434b67"} err="failed to get container status \"24648326d0ed4b1db9306783eeae2a39a0b46c7d4973ed6d5ad68bfc74434b67\": rpc error: code = NotFound desc = could not find container \"24648326d0ed4b1db9306783eeae2a39a0b46c7d4973ed6d5ad68bfc74434b67\": container with ID starting with 24648326d0ed4b1db9306783eeae2a39a0b46c7d4973ed6d5ad68bfc74434b67 not found: ID does not exist" Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.034659 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05ec15f4-9b81-4585-adfd-c749c7b12c38-config-data\") pod \"05ec15f4-9b81-4585-adfd-c749c7b12c38\" (UID: \"05ec15f4-9b81-4585-adfd-c749c7b12c38\") " Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.034780 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71c5fc79-7c68-4406-b02a-6cf325075a7a-config-data\") pod \"71c5fc79-7c68-4406-b02a-6cf325075a7a\" (UID: \"71c5fc79-7c68-4406-b02a-6cf325075a7a\") " Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.034850 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8fs2n\" (UniqueName: \"kubernetes.io/projected/05ec15f4-9b81-4585-adfd-c749c7b12c38-kube-api-access-8fs2n\") pod \"05ec15f4-9b81-4585-adfd-c749c7b12c38\" (UID: \"05ec15f4-9b81-4585-adfd-c749c7b12c38\") " Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.034948 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05ec15f4-9b81-4585-adfd-c749c7b12c38-combined-ca-bundle\") pod \"05ec15f4-9b81-4585-adfd-c749c7b12c38\" (UID: \"05ec15f4-9b81-4585-adfd-c749c7b12c38\") " Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.035017 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71c5fc79-7c68-4406-b02a-6cf325075a7a-combined-ca-bundle\") pod \"71c5fc79-7c68-4406-b02a-6cf325075a7a\" (UID: \"71c5fc79-7c68-4406-b02a-6cf325075a7a\") " Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.035080 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/05ec15f4-9b81-4585-adfd-c749c7b12c38-logs\") pod \"05ec15f4-9b81-4585-adfd-c749c7b12c38\" (UID: \"05ec15f4-9b81-4585-adfd-c749c7b12c38\") " Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.035114 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n87f5\" (UniqueName: \"kubernetes.io/projected/71c5fc79-7c68-4406-b02a-6cf325075a7a-kube-api-access-n87f5\") pod \"71c5fc79-7c68-4406-b02a-6cf325075a7a\" (UID: \"71c5fc79-7c68-4406-b02a-6cf325075a7a\") " Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.035893 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05ec15f4-9b81-4585-adfd-c749c7b12c38-logs" (OuterVolumeSpecName: "logs") pod "05ec15f4-9b81-4585-adfd-c749c7b12c38" (UID: "05ec15f4-9b81-4585-adfd-c749c7b12c38"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.041692 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05ec15f4-9b81-4585-adfd-c749c7b12c38-kube-api-access-8fs2n" (OuterVolumeSpecName: "kube-api-access-8fs2n") pod "05ec15f4-9b81-4585-adfd-c749c7b12c38" (UID: "05ec15f4-9b81-4585-adfd-c749c7b12c38"). InnerVolumeSpecName "kube-api-access-8fs2n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.041750 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71c5fc79-7c68-4406-b02a-6cf325075a7a-kube-api-access-n87f5" (OuterVolumeSpecName: "kube-api-access-n87f5") pod "71c5fc79-7c68-4406-b02a-6cf325075a7a" (UID: "71c5fc79-7c68-4406-b02a-6cf325075a7a"). InnerVolumeSpecName "kube-api-access-n87f5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.070485 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71c5fc79-7c68-4406-b02a-6cf325075a7a-config-data" (OuterVolumeSpecName: "config-data") pod "71c5fc79-7c68-4406-b02a-6cf325075a7a" (UID: "71c5fc79-7c68-4406-b02a-6cf325075a7a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.070896 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71c5fc79-7c68-4406-b02a-6cf325075a7a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "71c5fc79-7c68-4406-b02a-6cf325075a7a" (UID: "71c5fc79-7c68-4406-b02a-6cf325075a7a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.075180 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05ec15f4-9b81-4585-adfd-c749c7b12c38-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "05ec15f4-9b81-4585-adfd-c749c7b12c38" (UID: "05ec15f4-9b81-4585-adfd-c749c7b12c38"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.084050 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05ec15f4-9b81-4585-adfd-c749c7b12c38-config-data" (OuterVolumeSpecName: "config-data") pod "05ec15f4-9b81-4585-adfd-c749c7b12c38" (UID: "05ec15f4-9b81-4585-adfd-c749c7b12c38"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.139047 4802 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05ec15f4-9b81-4585-adfd-c749c7b12c38-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.139074 4802 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71c5fc79-7c68-4406-b02a-6cf325075a7a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.139083 4802 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/05ec15f4-9b81-4585-adfd-c749c7b12c38-logs\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.139092 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n87f5\" (UniqueName: \"kubernetes.io/projected/71c5fc79-7c68-4406-b02a-6cf325075a7a-kube-api-access-n87f5\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.139102 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05ec15f4-9b81-4585-adfd-c749c7b12c38-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.139110 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71c5fc79-7c68-4406-b02a-6cf325075a7a-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.139118 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8fs2n\" (UniqueName: \"kubernetes.io/projected/05ec15f4-9b81-4585-adfd-c749c7b12c38-kube-api-access-8fs2n\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.239122 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.246377 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.246634 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.248891 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.261150 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.271555 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.280447 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 06 04:01:31 crc kubenswrapper[4802]: E1206 04:01:31.281057 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05ec15f4-9b81-4585-adfd-c749c7b12c38" containerName="nova-api-api" Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.281125 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="05ec15f4-9b81-4585-adfd-c749c7b12c38" containerName="nova-api-api" Dec 06 04:01:31 crc kubenswrapper[4802]: E1206 04:01:31.281368 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05ec15f4-9b81-4585-adfd-c749c7b12c38" containerName="nova-api-log" Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.281473 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="05ec15f4-9b81-4585-adfd-c749c7b12c38" containerName="nova-api-log" Dec 06 04:01:31 crc kubenswrapper[4802]: E1206 04:01:31.281559 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71c5fc79-7c68-4406-b02a-6cf325075a7a" containerName="nova-scheduler-scheduler" Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.281613 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="71c5fc79-7c68-4406-b02a-6cf325075a7a" containerName="nova-scheduler-scheduler" Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.281878 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="05ec15f4-9b81-4585-adfd-c749c7b12c38" containerName="nova-api-log" Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.281968 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="05ec15f4-9b81-4585-adfd-c749c7b12c38" containerName="nova-api-api" Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.282031 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="71c5fc79-7c68-4406-b02a-6cf325075a7a" containerName="nova-scheduler-scheduler" Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.283079 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.285533 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.290699 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.291930 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.297050 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.300476 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.321874 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.443808 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zck9m\" (UniqueName: \"kubernetes.io/projected/cb0d4e0d-9954-49c8-b05a-118fb157c87e-kube-api-access-zck9m\") pod \"nova-scheduler-0\" (UID: \"cb0d4e0d-9954-49c8-b05a-118fb157c87e\") " pod="openstack/nova-scheduler-0" Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.443867 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee1f735d-c438-4b44-be26-cd9b4a1b7906-config-data\") pod \"nova-api-0\" (UID: \"ee1f735d-c438-4b44-be26-cd9b4a1b7906\") " pod="openstack/nova-api-0" Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.443898 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ee1f735d-c438-4b44-be26-cd9b4a1b7906-logs\") pod \"nova-api-0\" (UID: \"ee1f735d-c438-4b44-be26-cd9b4a1b7906\") " pod="openstack/nova-api-0" Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.444028 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb0d4e0d-9954-49c8-b05a-118fb157c87e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"cb0d4e0d-9954-49c8-b05a-118fb157c87e\") " pod="openstack/nova-scheduler-0" Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.444078 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb0d4e0d-9954-49c8-b05a-118fb157c87e-config-data\") pod \"nova-scheduler-0\" (UID: \"cb0d4e0d-9954-49c8-b05a-118fb157c87e\") " pod="openstack/nova-scheduler-0" Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.444172 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee1f735d-c438-4b44-be26-cd9b4a1b7906-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ee1f735d-c438-4b44-be26-cd9b4a1b7906\") " pod="openstack/nova-api-0" Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.444280 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wc7r5\" (UniqueName: \"kubernetes.io/projected/ee1f735d-c438-4b44-be26-cd9b4a1b7906-kube-api-access-wc7r5\") pod \"nova-api-0\" (UID: \"ee1f735d-c438-4b44-be26-cd9b4a1b7906\") " pod="openstack/nova-api-0" Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.460430 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05ec15f4-9b81-4585-adfd-c749c7b12c38" path="/var/lib/kubelet/pods/05ec15f4-9b81-4585-adfd-c749c7b12c38/volumes" Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.461112 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71c5fc79-7c68-4406-b02a-6cf325075a7a" path="/var/lib/kubelet/pods/71c5fc79-7c68-4406-b02a-6cf325075a7a/volumes" Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.545870 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wc7r5\" (UniqueName: \"kubernetes.io/projected/ee1f735d-c438-4b44-be26-cd9b4a1b7906-kube-api-access-wc7r5\") pod \"nova-api-0\" (UID: \"ee1f735d-c438-4b44-be26-cd9b4a1b7906\") " pod="openstack/nova-api-0" Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.545979 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zck9m\" (UniqueName: \"kubernetes.io/projected/cb0d4e0d-9954-49c8-b05a-118fb157c87e-kube-api-access-zck9m\") pod \"nova-scheduler-0\" (UID: \"cb0d4e0d-9954-49c8-b05a-118fb157c87e\") " pod="openstack/nova-scheduler-0" Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.546024 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee1f735d-c438-4b44-be26-cd9b4a1b7906-config-data\") pod \"nova-api-0\" (UID: \"ee1f735d-c438-4b44-be26-cd9b4a1b7906\") " pod="openstack/nova-api-0" Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.546070 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ee1f735d-c438-4b44-be26-cd9b4a1b7906-logs\") pod \"nova-api-0\" (UID: \"ee1f735d-c438-4b44-be26-cd9b4a1b7906\") " pod="openstack/nova-api-0" Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.546207 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb0d4e0d-9954-49c8-b05a-118fb157c87e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"cb0d4e0d-9954-49c8-b05a-118fb157c87e\") " pod="openstack/nova-scheduler-0" Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.546707 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb0d4e0d-9954-49c8-b05a-118fb157c87e-config-data\") pod \"nova-scheduler-0\" (UID: \"cb0d4e0d-9954-49c8-b05a-118fb157c87e\") " pod="openstack/nova-scheduler-0" Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.546828 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ee1f735d-c438-4b44-be26-cd9b4a1b7906-logs\") pod \"nova-api-0\" (UID: \"ee1f735d-c438-4b44-be26-cd9b4a1b7906\") " pod="openstack/nova-api-0" Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.547305 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee1f735d-c438-4b44-be26-cd9b4a1b7906-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ee1f735d-c438-4b44-be26-cd9b4a1b7906\") " pod="openstack/nova-api-0" Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.551329 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee1f735d-c438-4b44-be26-cd9b4a1b7906-config-data\") pod \"nova-api-0\" (UID: \"ee1f735d-c438-4b44-be26-cd9b4a1b7906\") " pod="openstack/nova-api-0" Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.551373 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb0d4e0d-9954-49c8-b05a-118fb157c87e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"cb0d4e0d-9954-49c8-b05a-118fb157c87e\") " pod="openstack/nova-scheduler-0" Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.551627 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb0d4e0d-9954-49c8-b05a-118fb157c87e-config-data\") pod \"nova-scheduler-0\" (UID: \"cb0d4e0d-9954-49c8-b05a-118fb157c87e\") " pod="openstack/nova-scheduler-0" Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.552216 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee1f735d-c438-4b44-be26-cd9b4a1b7906-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ee1f735d-c438-4b44-be26-cd9b4a1b7906\") " pod="openstack/nova-api-0" Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.565299 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zck9m\" (UniqueName: \"kubernetes.io/projected/cb0d4e0d-9954-49c8-b05a-118fb157c87e-kube-api-access-zck9m\") pod \"nova-scheduler-0\" (UID: \"cb0d4e0d-9954-49c8-b05a-118fb157c87e\") " pod="openstack/nova-scheduler-0" Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.575449 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wc7r5\" (UniqueName: \"kubernetes.io/projected/ee1f735d-c438-4b44-be26-cd9b4a1b7906-kube-api-access-wc7r5\") pod \"nova-api-0\" (UID: \"ee1f735d-c438-4b44-be26-cd9b4a1b7906\") " pod="openstack/nova-api-0" Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.598197 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 04:01:31 crc kubenswrapper[4802]: I1206 04:01:31.607183 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 04:01:32 crc kubenswrapper[4802]: I1206 04:01:32.175367 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 06 04:01:32 crc kubenswrapper[4802]: I1206 04:01:32.248412 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 04:01:32 crc kubenswrapper[4802]: W1206 04:01:32.251568 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcb0d4e0d_9954_49c8_b05a_118fb157c87e.slice/crio-774f7ead2ef7c54ec80dedaaa235e905afc165b55be17fefce1a4ffc790bea8c WatchSource:0}: Error finding container 774f7ead2ef7c54ec80dedaaa235e905afc165b55be17fefce1a4ffc790bea8c: Status 404 returned error can't find the container with id 774f7ead2ef7c54ec80dedaaa235e905afc165b55be17fefce1a4ffc790bea8c Dec 06 04:01:32 crc kubenswrapper[4802]: I1206 04:01:32.947352 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ee1f735d-c438-4b44-be26-cd9b4a1b7906","Type":"ContainerStarted","Data":"6bec787747e626d59fa9cfe8bfac461664d5016602e4ae4ac2a77193d517b046"} Dec 06 04:01:32 crc kubenswrapper[4802]: I1206 04:01:32.947688 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ee1f735d-c438-4b44-be26-cd9b4a1b7906","Type":"ContainerStarted","Data":"5c17109631f662ffba99b75d4d915c33f7c4dd58a15f40333caeaf79ddfccb03"} Dec 06 04:01:32 crc kubenswrapper[4802]: I1206 04:01:32.947705 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ee1f735d-c438-4b44-be26-cd9b4a1b7906","Type":"ContainerStarted","Data":"f8349daf8172aec143b601a5b9bad0e1000c3f3921950fb0e77b47365ffe0052"} Dec 06 04:01:32 crc kubenswrapper[4802]: I1206 04:01:32.949166 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cb0d4e0d-9954-49c8-b05a-118fb157c87e","Type":"ContainerStarted","Data":"74685c7dbe3d20aba42d93effcc6a18d5f191b2974cf99959fe1b09051c22626"} Dec 06 04:01:32 crc kubenswrapper[4802]: I1206 04:01:32.949227 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cb0d4e0d-9954-49c8-b05a-118fb157c87e","Type":"ContainerStarted","Data":"774f7ead2ef7c54ec80dedaaa235e905afc165b55be17fefce1a4ffc790bea8c"} Dec 06 04:01:32 crc kubenswrapper[4802]: I1206 04:01:32.972400 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=1.97237966 podStartE2EDuration="1.97237966s" podCreationTimestamp="2025-12-06 04:01:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:01:32.967569719 +0000 UTC m=+1285.839478871" watchObservedRunningTime="2025-12-06 04:01:32.97237966 +0000 UTC m=+1285.844288812" Dec 06 04:01:32 crc kubenswrapper[4802]: I1206 04:01:32.988993 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.9889758039999998 podStartE2EDuration="1.988975804s" podCreationTimestamp="2025-12-06 04:01:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:01:32.983113984 +0000 UTC m=+1285.855023126" watchObservedRunningTime="2025-12-06 04:01:32.988975804 +0000 UTC m=+1285.860884956" Dec 06 04:01:34 crc kubenswrapper[4802]: I1206 04:01:34.593654 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 04:01:34 crc kubenswrapper[4802]: I1206 04:01:34.596807 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="b96722fa-dc6a-44c3-9460-0ce247056613" containerName="kube-state-metrics" containerID="cri-o://44f31f10d271b5bc0656d1123c02d51de86d32004e13b130eeb6e5eb02a544e8" gracePeriod=30 Dec 06 04:01:34 crc kubenswrapper[4802]: I1206 04:01:34.971903 4802 generic.go:334] "Generic (PLEG): container finished" podID="b96722fa-dc6a-44c3-9460-0ce247056613" containerID="44f31f10d271b5bc0656d1123c02d51de86d32004e13b130eeb6e5eb02a544e8" exitCode=2 Dec 06 04:01:34 crc kubenswrapper[4802]: I1206 04:01:34.971964 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"b96722fa-dc6a-44c3-9460-0ce247056613","Type":"ContainerDied","Data":"44f31f10d271b5bc0656d1123c02d51de86d32004e13b130eeb6e5eb02a544e8"} Dec 06 04:01:35 crc kubenswrapper[4802]: I1206 04:01:35.104644 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 06 04:01:35 crc kubenswrapper[4802]: I1206 04:01:35.225502 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fbzm2\" (UniqueName: \"kubernetes.io/projected/b96722fa-dc6a-44c3-9460-0ce247056613-kube-api-access-fbzm2\") pod \"b96722fa-dc6a-44c3-9460-0ce247056613\" (UID: \"b96722fa-dc6a-44c3-9460-0ce247056613\") " Dec 06 04:01:35 crc kubenswrapper[4802]: I1206 04:01:35.234389 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b96722fa-dc6a-44c3-9460-0ce247056613-kube-api-access-fbzm2" (OuterVolumeSpecName: "kube-api-access-fbzm2") pod "b96722fa-dc6a-44c3-9460-0ce247056613" (UID: "b96722fa-dc6a-44c3-9460-0ce247056613"). InnerVolumeSpecName "kube-api-access-fbzm2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:01:35 crc kubenswrapper[4802]: I1206 04:01:35.327212 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fbzm2\" (UniqueName: \"kubernetes.io/projected/b96722fa-dc6a-44c3-9460-0ce247056613-kube-api-access-fbzm2\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:35 crc kubenswrapper[4802]: I1206 04:01:35.982421 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"b96722fa-dc6a-44c3-9460-0ce247056613","Type":"ContainerDied","Data":"5224c4b5e7ebf432c7dfc363b364cfd34bf6890a993f567c73e1b1a4d4701034"} Dec 06 04:01:35 crc kubenswrapper[4802]: I1206 04:01:35.983394 4802 scope.go:117] "RemoveContainer" containerID="44f31f10d271b5bc0656d1123c02d51de86d32004e13b130eeb6e5eb02a544e8" Dec 06 04:01:35 crc kubenswrapper[4802]: I1206 04:01:35.982531 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 06 04:01:36 crc kubenswrapper[4802]: I1206 04:01:36.004225 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 04:01:36 crc kubenswrapper[4802]: I1206 04:01:36.013637 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 04:01:36 crc kubenswrapper[4802]: I1206 04:01:36.025554 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 04:01:36 crc kubenswrapper[4802]: E1206 04:01:36.026070 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b96722fa-dc6a-44c3-9460-0ce247056613" containerName="kube-state-metrics" Dec 06 04:01:36 crc kubenswrapper[4802]: I1206 04:01:36.026095 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="b96722fa-dc6a-44c3-9460-0ce247056613" containerName="kube-state-metrics" Dec 06 04:01:36 crc kubenswrapper[4802]: I1206 04:01:36.026376 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="b96722fa-dc6a-44c3-9460-0ce247056613" containerName="kube-state-metrics" Dec 06 04:01:36 crc kubenswrapper[4802]: I1206 04:01:36.027225 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 06 04:01:36 crc kubenswrapper[4802]: I1206 04:01:36.029211 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 06 04:01:36 crc kubenswrapper[4802]: I1206 04:01:36.029444 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 06 04:01:36 crc kubenswrapper[4802]: I1206 04:01:36.036289 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 04:01:36 crc kubenswrapper[4802]: I1206 04:01:36.143173 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9szxb\" (UniqueName: \"kubernetes.io/projected/5489ef4c-603a-44ce-9a0e-6ad56188e669-kube-api-access-9szxb\") pod \"kube-state-metrics-0\" (UID: \"5489ef4c-603a-44ce-9a0e-6ad56188e669\") " pod="openstack/kube-state-metrics-0" Dec 06 04:01:36 crc kubenswrapper[4802]: I1206 04:01:36.143512 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/5489ef4c-603a-44ce-9a0e-6ad56188e669-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"5489ef4c-603a-44ce-9a0e-6ad56188e669\") " pod="openstack/kube-state-metrics-0" Dec 06 04:01:36 crc kubenswrapper[4802]: I1206 04:01:36.143692 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5489ef4c-603a-44ce-9a0e-6ad56188e669-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"5489ef4c-603a-44ce-9a0e-6ad56188e669\") " pod="openstack/kube-state-metrics-0" Dec 06 04:01:36 crc kubenswrapper[4802]: I1206 04:01:36.143879 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/5489ef4c-603a-44ce-9a0e-6ad56188e669-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"5489ef4c-603a-44ce-9a0e-6ad56188e669\") " pod="openstack/kube-state-metrics-0" Dec 06 04:01:36 crc kubenswrapper[4802]: I1206 04:01:36.214347 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 06 04:01:36 crc kubenswrapper[4802]: I1206 04:01:36.246169 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/5489ef4c-603a-44ce-9a0e-6ad56188e669-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"5489ef4c-603a-44ce-9a0e-6ad56188e669\") " pod="openstack/kube-state-metrics-0" Dec 06 04:01:36 crc kubenswrapper[4802]: I1206 04:01:36.246233 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5489ef4c-603a-44ce-9a0e-6ad56188e669-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"5489ef4c-603a-44ce-9a0e-6ad56188e669\") " pod="openstack/kube-state-metrics-0" Dec 06 04:01:36 crc kubenswrapper[4802]: I1206 04:01:36.246263 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/5489ef4c-603a-44ce-9a0e-6ad56188e669-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"5489ef4c-603a-44ce-9a0e-6ad56188e669\") " pod="openstack/kube-state-metrics-0" Dec 06 04:01:36 crc kubenswrapper[4802]: I1206 04:01:36.246402 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9szxb\" (UniqueName: \"kubernetes.io/projected/5489ef4c-603a-44ce-9a0e-6ad56188e669-kube-api-access-9szxb\") pod \"kube-state-metrics-0\" (UID: \"5489ef4c-603a-44ce-9a0e-6ad56188e669\") " pod="openstack/kube-state-metrics-0" Dec 06 04:01:36 crc kubenswrapper[4802]: I1206 04:01:36.246834 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 06 04:01:36 crc kubenswrapper[4802]: I1206 04:01:36.246869 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 06 04:01:36 crc kubenswrapper[4802]: I1206 04:01:36.252556 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/5489ef4c-603a-44ce-9a0e-6ad56188e669-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"5489ef4c-603a-44ce-9a0e-6ad56188e669\") " pod="openstack/kube-state-metrics-0" Dec 06 04:01:36 crc kubenswrapper[4802]: I1206 04:01:36.254048 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/5489ef4c-603a-44ce-9a0e-6ad56188e669-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"5489ef4c-603a-44ce-9a0e-6ad56188e669\") " pod="openstack/kube-state-metrics-0" Dec 06 04:01:36 crc kubenswrapper[4802]: I1206 04:01:36.254353 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5489ef4c-603a-44ce-9a0e-6ad56188e669-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"5489ef4c-603a-44ce-9a0e-6ad56188e669\") " pod="openstack/kube-state-metrics-0" Dec 06 04:01:36 crc kubenswrapper[4802]: I1206 04:01:36.265617 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9szxb\" (UniqueName: \"kubernetes.io/projected/5489ef4c-603a-44ce-9a0e-6ad56188e669-kube-api-access-9szxb\") pod \"kube-state-metrics-0\" (UID: \"5489ef4c-603a-44ce-9a0e-6ad56188e669\") " pod="openstack/kube-state-metrics-0" Dec 06 04:01:36 crc kubenswrapper[4802]: I1206 04:01:36.346142 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 06 04:01:36 crc kubenswrapper[4802]: I1206 04:01:36.611991 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 06 04:01:36 crc kubenswrapper[4802]: I1206 04:01:36.626322 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 04:01:36 crc kubenswrapper[4802]: I1206 04:01:36.626628 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="30152cc8-8968-443f-923a-0ddca8c016ee" containerName="sg-core" containerID="cri-o://4e67921587bb5beca9f50d6d82a1e653bdf472c8cbd66272021421e0474dba82" gracePeriod=30 Dec 06 04:01:36 crc kubenswrapper[4802]: I1206 04:01:36.626658 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="30152cc8-8968-443f-923a-0ddca8c016ee" containerName="proxy-httpd" containerID="cri-o://7b9a8b8ec82130774b4701a4abc56aeb457b4266e4d179334ca9825fb673f20b" gracePeriod=30 Dec 06 04:01:36 crc kubenswrapper[4802]: I1206 04:01:36.626721 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="30152cc8-8968-443f-923a-0ddca8c016ee" containerName="ceilometer-notification-agent" containerID="cri-o://4b2f1959e7f989b0dc68503aec47cb31044c3a075f0bc892a91e9336697a3c15" gracePeriod=30 Dec 06 04:01:36 crc kubenswrapper[4802]: I1206 04:01:36.626589 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="30152cc8-8968-443f-923a-0ddca8c016ee" containerName="ceilometer-central-agent" containerID="cri-o://9419cab08a0d0f6b39eefd268cd4b3759dab87c76047fa89e3000f6949febb58" gracePeriod=30 Dec 06 04:01:36 crc kubenswrapper[4802]: I1206 04:01:36.815768 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 04:01:37 crc kubenswrapper[4802]: I1206 04:01:37.018591 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"5489ef4c-603a-44ce-9a0e-6ad56188e669","Type":"ContainerStarted","Data":"5e7af4f8a3a6b7e48289bf7955569cc2a895e1ad94ec7465b7c5d6c308c0dceb"} Dec 06 04:01:37 crc kubenswrapper[4802]: I1206 04:01:37.029024 4802 generic.go:334] "Generic (PLEG): container finished" podID="30152cc8-8968-443f-923a-0ddca8c016ee" containerID="7b9a8b8ec82130774b4701a4abc56aeb457b4266e4d179334ca9825fb673f20b" exitCode=0 Dec 06 04:01:37 crc kubenswrapper[4802]: I1206 04:01:37.029332 4802 generic.go:334] "Generic (PLEG): container finished" podID="30152cc8-8968-443f-923a-0ddca8c016ee" containerID="4e67921587bb5beca9f50d6d82a1e653bdf472c8cbd66272021421e0474dba82" exitCode=2 Dec 06 04:01:37 crc kubenswrapper[4802]: I1206 04:01:37.029426 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"30152cc8-8968-443f-923a-0ddca8c016ee","Type":"ContainerDied","Data":"7b9a8b8ec82130774b4701a4abc56aeb457b4266e4d179334ca9825fb673f20b"} Dec 06 04:01:37 crc kubenswrapper[4802]: I1206 04:01:37.029526 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"30152cc8-8968-443f-923a-0ddca8c016ee","Type":"ContainerDied","Data":"4e67921587bb5beca9f50d6d82a1e653bdf472c8cbd66272021421e0474dba82"} Dec 06 04:01:37 crc kubenswrapper[4802]: I1206 04:01:37.301353 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.201:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 04:01:37 crc kubenswrapper[4802]: I1206 04:01:37.301506 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.201:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 04:01:37 crc kubenswrapper[4802]: I1206 04:01:37.466325 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b96722fa-dc6a-44c3-9460-0ce247056613" path="/var/lib/kubelet/pods/b96722fa-dc6a-44c3-9460-0ce247056613/volumes" Dec 06 04:01:38 crc kubenswrapper[4802]: I1206 04:01:38.044749 4802 generic.go:334] "Generic (PLEG): container finished" podID="30152cc8-8968-443f-923a-0ddca8c016ee" containerID="9419cab08a0d0f6b39eefd268cd4b3759dab87c76047fa89e3000f6949febb58" exitCode=0 Dec 06 04:01:38 crc kubenswrapper[4802]: I1206 04:01:38.044790 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"30152cc8-8968-443f-923a-0ddca8c016ee","Type":"ContainerDied","Data":"9419cab08a0d0f6b39eefd268cd4b3759dab87c76047fa89e3000f6949febb58"} Dec 06 04:01:38 crc kubenswrapper[4802]: I1206 04:01:38.048189 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"5489ef4c-603a-44ce-9a0e-6ad56188e669","Type":"ContainerStarted","Data":"b92cfc4f47dcdcf2744cbfecdefa103988e66a2ad1526c23b33eccc47a00e58b"} Dec 06 04:01:38 crc kubenswrapper[4802]: I1206 04:01:38.049320 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 06 04:01:38 crc kubenswrapper[4802]: I1206 04:01:38.085646 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.701507938 podStartE2EDuration="2.085626593s" podCreationTimestamp="2025-12-06 04:01:36 +0000 UTC" firstStartedPulling="2025-12-06 04:01:36.817699692 +0000 UTC m=+1289.689608844" lastFinishedPulling="2025-12-06 04:01:37.201818347 +0000 UTC m=+1290.073727499" observedRunningTime="2025-12-06 04:01:38.062484909 +0000 UTC m=+1290.934394071" watchObservedRunningTime="2025-12-06 04:01:38.085626593 +0000 UTC m=+1290.957535735" Dec 06 04:01:39 crc kubenswrapper[4802]: I1206 04:01:39.725002 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 04:01:39 crc kubenswrapper[4802]: I1206 04:01:39.817478 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30152cc8-8968-443f-923a-0ddca8c016ee-scripts\") pod \"30152cc8-8968-443f-923a-0ddca8c016ee\" (UID: \"30152cc8-8968-443f-923a-0ddca8c016ee\") " Dec 06 04:01:39 crc kubenswrapper[4802]: I1206 04:01:39.817580 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30152cc8-8968-443f-923a-0ddca8c016ee-config-data\") pod \"30152cc8-8968-443f-923a-0ddca8c016ee\" (UID: \"30152cc8-8968-443f-923a-0ddca8c016ee\") " Dec 06 04:01:39 crc kubenswrapper[4802]: I1206 04:01:39.817640 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30152cc8-8968-443f-923a-0ddca8c016ee-combined-ca-bundle\") pod \"30152cc8-8968-443f-923a-0ddca8c016ee\" (UID: \"30152cc8-8968-443f-923a-0ddca8c016ee\") " Dec 06 04:01:39 crc kubenswrapper[4802]: I1206 04:01:39.817661 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-29vhd\" (UniqueName: \"kubernetes.io/projected/30152cc8-8968-443f-923a-0ddca8c016ee-kube-api-access-29vhd\") pod \"30152cc8-8968-443f-923a-0ddca8c016ee\" (UID: \"30152cc8-8968-443f-923a-0ddca8c016ee\") " Dec 06 04:01:39 crc kubenswrapper[4802]: I1206 04:01:39.817693 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/30152cc8-8968-443f-923a-0ddca8c016ee-sg-core-conf-yaml\") pod \"30152cc8-8968-443f-923a-0ddca8c016ee\" (UID: \"30152cc8-8968-443f-923a-0ddca8c016ee\") " Dec 06 04:01:39 crc kubenswrapper[4802]: I1206 04:01:39.817741 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/30152cc8-8968-443f-923a-0ddca8c016ee-log-httpd\") pod \"30152cc8-8968-443f-923a-0ddca8c016ee\" (UID: \"30152cc8-8968-443f-923a-0ddca8c016ee\") " Dec 06 04:01:39 crc kubenswrapper[4802]: I1206 04:01:39.817824 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/30152cc8-8968-443f-923a-0ddca8c016ee-run-httpd\") pod \"30152cc8-8968-443f-923a-0ddca8c016ee\" (UID: \"30152cc8-8968-443f-923a-0ddca8c016ee\") " Dec 06 04:01:39 crc kubenswrapper[4802]: I1206 04:01:39.818252 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30152cc8-8968-443f-923a-0ddca8c016ee-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "30152cc8-8968-443f-923a-0ddca8c016ee" (UID: "30152cc8-8968-443f-923a-0ddca8c016ee"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:01:39 crc kubenswrapper[4802]: I1206 04:01:39.818413 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30152cc8-8968-443f-923a-0ddca8c016ee-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "30152cc8-8968-443f-923a-0ddca8c016ee" (UID: "30152cc8-8968-443f-923a-0ddca8c016ee"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:01:39 crc kubenswrapper[4802]: I1206 04:01:39.823610 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30152cc8-8968-443f-923a-0ddca8c016ee-kube-api-access-29vhd" (OuterVolumeSpecName: "kube-api-access-29vhd") pod "30152cc8-8968-443f-923a-0ddca8c016ee" (UID: "30152cc8-8968-443f-923a-0ddca8c016ee"). InnerVolumeSpecName "kube-api-access-29vhd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:01:39 crc kubenswrapper[4802]: I1206 04:01:39.830083 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30152cc8-8968-443f-923a-0ddca8c016ee-scripts" (OuterVolumeSpecName: "scripts") pod "30152cc8-8968-443f-923a-0ddca8c016ee" (UID: "30152cc8-8968-443f-923a-0ddca8c016ee"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:01:39 crc kubenswrapper[4802]: I1206 04:01:39.845288 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30152cc8-8968-443f-923a-0ddca8c016ee-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "30152cc8-8968-443f-923a-0ddca8c016ee" (UID: "30152cc8-8968-443f-923a-0ddca8c016ee"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:01:39 crc kubenswrapper[4802]: I1206 04:01:39.894335 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30152cc8-8968-443f-923a-0ddca8c016ee-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "30152cc8-8968-443f-923a-0ddca8c016ee" (UID: "30152cc8-8968-443f-923a-0ddca8c016ee"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:01:39 crc kubenswrapper[4802]: I1206 04:01:39.916873 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30152cc8-8968-443f-923a-0ddca8c016ee-config-data" (OuterVolumeSpecName: "config-data") pod "30152cc8-8968-443f-923a-0ddca8c016ee" (UID: "30152cc8-8968-443f-923a-0ddca8c016ee"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:01:39 crc kubenswrapper[4802]: I1206 04:01:39.920069 4802 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30152cc8-8968-443f-923a-0ddca8c016ee-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:39 crc kubenswrapper[4802]: I1206 04:01:39.920095 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30152cc8-8968-443f-923a-0ddca8c016ee-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:39 crc kubenswrapper[4802]: I1206 04:01:39.920104 4802 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30152cc8-8968-443f-923a-0ddca8c016ee-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:39 crc kubenswrapper[4802]: I1206 04:01:39.920117 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-29vhd\" (UniqueName: \"kubernetes.io/projected/30152cc8-8968-443f-923a-0ddca8c016ee-kube-api-access-29vhd\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:39 crc kubenswrapper[4802]: I1206 04:01:39.920128 4802 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/30152cc8-8968-443f-923a-0ddca8c016ee-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:39 crc kubenswrapper[4802]: I1206 04:01:39.920136 4802 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/30152cc8-8968-443f-923a-0ddca8c016ee-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:39 crc kubenswrapper[4802]: I1206 04:01:39.920145 4802 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/30152cc8-8968-443f-923a-0ddca8c016ee-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:40 crc kubenswrapper[4802]: I1206 04:01:40.068980 4802 generic.go:334] "Generic (PLEG): container finished" podID="30152cc8-8968-443f-923a-0ddca8c016ee" containerID="4b2f1959e7f989b0dc68503aec47cb31044c3a075f0bc892a91e9336697a3c15" exitCode=0 Dec 06 04:01:40 crc kubenswrapper[4802]: I1206 04:01:40.069087 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"30152cc8-8968-443f-923a-0ddca8c016ee","Type":"ContainerDied","Data":"4b2f1959e7f989b0dc68503aec47cb31044c3a075f0bc892a91e9336697a3c15"} Dec 06 04:01:40 crc kubenswrapper[4802]: I1206 04:01:40.069137 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"30152cc8-8968-443f-923a-0ddca8c016ee","Type":"ContainerDied","Data":"2b782ebf701bf7af80fdaa69877b316efde31731a2bf48eee1e4c33c5d1a1917"} Dec 06 04:01:40 crc kubenswrapper[4802]: I1206 04:01:40.069161 4802 scope.go:117] "RemoveContainer" containerID="7b9a8b8ec82130774b4701a4abc56aeb457b4266e4d179334ca9825fb673f20b" Dec 06 04:01:40 crc kubenswrapper[4802]: I1206 04:01:40.069458 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 04:01:40 crc kubenswrapper[4802]: I1206 04:01:40.092739 4802 scope.go:117] "RemoveContainer" containerID="4e67921587bb5beca9f50d6d82a1e653bdf472c8cbd66272021421e0474dba82" Dec 06 04:01:40 crc kubenswrapper[4802]: I1206 04:01:40.112373 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 04:01:40 crc kubenswrapper[4802]: I1206 04:01:40.121021 4802 scope.go:117] "RemoveContainer" containerID="4b2f1959e7f989b0dc68503aec47cb31044c3a075f0bc892a91e9336697a3c15" Dec 06 04:01:40 crc kubenswrapper[4802]: I1206 04:01:40.132832 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 06 04:01:40 crc kubenswrapper[4802]: I1206 04:01:40.149556 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 06 04:01:40 crc kubenswrapper[4802]: E1206 04:01:40.150051 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30152cc8-8968-443f-923a-0ddca8c016ee" containerName="proxy-httpd" Dec 06 04:01:40 crc kubenswrapper[4802]: I1206 04:01:40.150068 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="30152cc8-8968-443f-923a-0ddca8c016ee" containerName="proxy-httpd" Dec 06 04:01:40 crc kubenswrapper[4802]: E1206 04:01:40.150097 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30152cc8-8968-443f-923a-0ddca8c016ee" containerName="ceilometer-central-agent" Dec 06 04:01:40 crc kubenswrapper[4802]: I1206 04:01:40.150106 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="30152cc8-8968-443f-923a-0ddca8c016ee" containerName="ceilometer-central-agent" Dec 06 04:01:40 crc kubenswrapper[4802]: E1206 04:01:40.150136 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30152cc8-8968-443f-923a-0ddca8c016ee" containerName="ceilometer-notification-agent" Dec 06 04:01:40 crc kubenswrapper[4802]: I1206 04:01:40.150146 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="30152cc8-8968-443f-923a-0ddca8c016ee" containerName="ceilometer-notification-agent" Dec 06 04:01:40 crc kubenswrapper[4802]: E1206 04:01:40.150154 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30152cc8-8968-443f-923a-0ddca8c016ee" containerName="sg-core" Dec 06 04:01:40 crc kubenswrapper[4802]: I1206 04:01:40.150162 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="30152cc8-8968-443f-923a-0ddca8c016ee" containerName="sg-core" Dec 06 04:01:40 crc kubenswrapper[4802]: I1206 04:01:40.150394 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="30152cc8-8968-443f-923a-0ddca8c016ee" containerName="ceilometer-central-agent" Dec 06 04:01:40 crc kubenswrapper[4802]: I1206 04:01:40.150413 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="30152cc8-8968-443f-923a-0ddca8c016ee" containerName="sg-core" Dec 06 04:01:40 crc kubenswrapper[4802]: I1206 04:01:40.150434 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="30152cc8-8968-443f-923a-0ddca8c016ee" containerName="proxy-httpd" Dec 06 04:01:40 crc kubenswrapper[4802]: I1206 04:01:40.150450 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="30152cc8-8968-443f-923a-0ddca8c016ee" containerName="ceilometer-notification-agent" Dec 06 04:01:40 crc kubenswrapper[4802]: I1206 04:01:40.151855 4802 scope.go:117] "RemoveContainer" containerID="9419cab08a0d0f6b39eefd268cd4b3759dab87c76047fa89e3000f6949febb58" Dec 06 04:01:40 crc kubenswrapper[4802]: I1206 04:01:40.152648 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 04:01:40 crc kubenswrapper[4802]: I1206 04:01:40.155199 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 06 04:01:40 crc kubenswrapper[4802]: I1206 04:01:40.155399 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 06 04:01:40 crc kubenswrapper[4802]: I1206 04:01:40.155588 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 06 04:01:40 crc kubenswrapper[4802]: I1206 04:01:40.161982 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 04:01:40 crc kubenswrapper[4802]: I1206 04:01:40.195537 4802 scope.go:117] "RemoveContainer" containerID="7b9a8b8ec82130774b4701a4abc56aeb457b4266e4d179334ca9825fb673f20b" Dec 06 04:01:40 crc kubenswrapper[4802]: E1206 04:01:40.196473 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b9a8b8ec82130774b4701a4abc56aeb457b4266e4d179334ca9825fb673f20b\": container with ID starting with 7b9a8b8ec82130774b4701a4abc56aeb457b4266e4d179334ca9825fb673f20b not found: ID does not exist" containerID="7b9a8b8ec82130774b4701a4abc56aeb457b4266e4d179334ca9825fb673f20b" Dec 06 04:01:40 crc kubenswrapper[4802]: I1206 04:01:40.196522 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b9a8b8ec82130774b4701a4abc56aeb457b4266e4d179334ca9825fb673f20b"} err="failed to get container status \"7b9a8b8ec82130774b4701a4abc56aeb457b4266e4d179334ca9825fb673f20b\": rpc error: code = NotFound desc = could not find container \"7b9a8b8ec82130774b4701a4abc56aeb457b4266e4d179334ca9825fb673f20b\": container with ID starting with 7b9a8b8ec82130774b4701a4abc56aeb457b4266e4d179334ca9825fb673f20b not found: ID does not exist" Dec 06 04:01:40 crc kubenswrapper[4802]: I1206 04:01:40.196547 4802 scope.go:117] "RemoveContainer" containerID="4e67921587bb5beca9f50d6d82a1e653bdf472c8cbd66272021421e0474dba82" Dec 06 04:01:40 crc kubenswrapper[4802]: E1206 04:01:40.196991 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e67921587bb5beca9f50d6d82a1e653bdf472c8cbd66272021421e0474dba82\": container with ID starting with 4e67921587bb5beca9f50d6d82a1e653bdf472c8cbd66272021421e0474dba82 not found: ID does not exist" containerID="4e67921587bb5beca9f50d6d82a1e653bdf472c8cbd66272021421e0474dba82" Dec 06 04:01:40 crc kubenswrapper[4802]: I1206 04:01:40.197020 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e67921587bb5beca9f50d6d82a1e653bdf472c8cbd66272021421e0474dba82"} err="failed to get container status \"4e67921587bb5beca9f50d6d82a1e653bdf472c8cbd66272021421e0474dba82\": rpc error: code = NotFound desc = could not find container \"4e67921587bb5beca9f50d6d82a1e653bdf472c8cbd66272021421e0474dba82\": container with ID starting with 4e67921587bb5beca9f50d6d82a1e653bdf472c8cbd66272021421e0474dba82 not found: ID does not exist" Dec 06 04:01:40 crc kubenswrapper[4802]: I1206 04:01:40.197040 4802 scope.go:117] "RemoveContainer" containerID="4b2f1959e7f989b0dc68503aec47cb31044c3a075f0bc892a91e9336697a3c15" Dec 06 04:01:40 crc kubenswrapper[4802]: E1206 04:01:40.197359 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b2f1959e7f989b0dc68503aec47cb31044c3a075f0bc892a91e9336697a3c15\": container with ID starting with 4b2f1959e7f989b0dc68503aec47cb31044c3a075f0bc892a91e9336697a3c15 not found: ID does not exist" containerID="4b2f1959e7f989b0dc68503aec47cb31044c3a075f0bc892a91e9336697a3c15" Dec 06 04:01:40 crc kubenswrapper[4802]: I1206 04:01:40.197377 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b2f1959e7f989b0dc68503aec47cb31044c3a075f0bc892a91e9336697a3c15"} err="failed to get container status \"4b2f1959e7f989b0dc68503aec47cb31044c3a075f0bc892a91e9336697a3c15\": rpc error: code = NotFound desc = could not find container \"4b2f1959e7f989b0dc68503aec47cb31044c3a075f0bc892a91e9336697a3c15\": container with ID starting with 4b2f1959e7f989b0dc68503aec47cb31044c3a075f0bc892a91e9336697a3c15 not found: ID does not exist" Dec 06 04:01:40 crc kubenswrapper[4802]: I1206 04:01:40.197389 4802 scope.go:117] "RemoveContainer" containerID="9419cab08a0d0f6b39eefd268cd4b3759dab87c76047fa89e3000f6949febb58" Dec 06 04:01:40 crc kubenswrapper[4802]: E1206 04:01:40.197566 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9419cab08a0d0f6b39eefd268cd4b3759dab87c76047fa89e3000f6949febb58\": container with ID starting with 9419cab08a0d0f6b39eefd268cd4b3759dab87c76047fa89e3000f6949febb58 not found: ID does not exist" containerID="9419cab08a0d0f6b39eefd268cd4b3759dab87c76047fa89e3000f6949febb58" Dec 06 04:01:40 crc kubenswrapper[4802]: I1206 04:01:40.197584 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9419cab08a0d0f6b39eefd268cd4b3759dab87c76047fa89e3000f6949febb58"} err="failed to get container status \"9419cab08a0d0f6b39eefd268cd4b3759dab87c76047fa89e3000f6949febb58\": rpc error: code = NotFound desc = could not find container \"9419cab08a0d0f6b39eefd268cd4b3759dab87c76047fa89e3000f6949febb58\": container with ID starting with 9419cab08a0d0f6b39eefd268cd4b3759dab87c76047fa89e3000f6949febb58 not found: ID does not exist" Dec 06 04:01:40 crc kubenswrapper[4802]: I1206 04:01:40.328514 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b08151ce-4f6b-4817-a2f4-8d7eb3102ce7-scripts\") pod \"ceilometer-0\" (UID: \"b08151ce-4f6b-4817-a2f4-8d7eb3102ce7\") " pod="openstack/ceilometer-0" Dec 06 04:01:40 crc kubenswrapper[4802]: I1206 04:01:40.328564 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b08151ce-4f6b-4817-a2f4-8d7eb3102ce7-log-httpd\") pod \"ceilometer-0\" (UID: \"b08151ce-4f6b-4817-a2f4-8d7eb3102ce7\") " pod="openstack/ceilometer-0" Dec 06 04:01:40 crc kubenswrapper[4802]: I1206 04:01:40.328617 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b08151ce-4f6b-4817-a2f4-8d7eb3102ce7-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"b08151ce-4f6b-4817-a2f4-8d7eb3102ce7\") " pod="openstack/ceilometer-0" Dec 06 04:01:40 crc kubenswrapper[4802]: I1206 04:01:40.328635 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbwmm\" (UniqueName: \"kubernetes.io/projected/b08151ce-4f6b-4817-a2f4-8d7eb3102ce7-kube-api-access-hbwmm\") pod \"ceilometer-0\" (UID: \"b08151ce-4f6b-4817-a2f4-8d7eb3102ce7\") " pod="openstack/ceilometer-0" Dec 06 04:01:40 crc kubenswrapper[4802]: I1206 04:01:40.328660 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b08151ce-4f6b-4817-a2f4-8d7eb3102ce7-run-httpd\") pod \"ceilometer-0\" (UID: \"b08151ce-4f6b-4817-a2f4-8d7eb3102ce7\") " pod="openstack/ceilometer-0" Dec 06 04:01:40 crc kubenswrapper[4802]: I1206 04:01:40.328704 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b08151ce-4f6b-4817-a2f4-8d7eb3102ce7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b08151ce-4f6b-4817-a2f4-8d7eb3102ce7\") " pod="openstack/ceilometer-0" Dec 06 04:01:40 crc kubenswrapper[4802]: I1206 04:01:40.328728 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b08151ce-4f6b-4817-a2f4-8d7eb3102ce7-config-data\") pod \"ceilometer-0\" (UID: \"b08151ce-4f6b-4817-a2f4-8d7eb3102ce7\") " pod="openstack/ceilometer-0" Dec 06 04:01:40 crc kubenswrapper[4802]: I1206 04:01:40.328839 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b08151ce-4f6b-4817-a2f4-8d7eb3102ce7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b08151ce-4f6b-4817-a2f4-8d7eb3102ce7\") " pod="openstack/ceilometer-0" Dec 06 04:01:40 crc kubenswrapper[4802]: I1206 04:01:40.430618 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b08151ce-4f6b-4817-a2f4-8d7eb3102ce7-config-data\") pod \"ceilometer-0\" (UID: \"b08151ce-4f6b-4817-a2f4-8d7eb3102ce7\") " pod="openstack/ceilometer-0" Dec 06 04:01:40 crc kubenswrapper[4802]: I1206 04:01:40.430668 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b08151ce-4f6b-4817-a2f4-8d7eb3102ce7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b08151ce-4f6b-4817-a2f4-8d7eb3102ce7\") " pod="openstack/ceilometer-0" Dec 06 04:01:40 crc kubenswrapper[4802]: I1206 04:01:40.430779 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b08151ce-4f6b-4817-a2f4-8d7eb3102ce7-scripts\") pod \"ceilometer-0\" (UID: \"b08151ce-4f6b-4817-a2f4-8d7eb3102ce7\") " pod="openstack/ceilometer-0" Dec 06 04:01:40 crc kubenswrapper[4802]: I1206 04:01:40.430815 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b08151ce-4f6b-4817-a2f4-8d7eb3102ce7-log-httpd\") pod \"ceilometer-0\" (UID: \"b08151ce-4f6b-4817-a2f4-8d7eb3102ce7\") " pod="openstack/ceilometer-0" Dec 06 04:01:40 crc kubenswrapper[4802]: I1206 04:01:40.430872 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b08151ce-4f6b-4817-a2f4-8d7eb3102ce7-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"b08151ce-4f6b-4817-a2f4-8d7eb3102ce7\") " pod="openstack/ceilometer-0" Dec 06 04:01:40 crc kubenswrapper[4802]: I1206 04:01:40.430896 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbwmm\" (UniqueName: \"kubernetes.io/projected/b08151ce-4f6b-4817-a2f4-8d7eb3102ce7-kube-api-access-hbwmm\") pod \"ceilometer-0\" (UID: \"b08151ce-4f6b-4817-a2f4-8d7eb3102ce7\") " pod="openstack/ceilometer-0" Dec 06 04:01:40 crc kubenswrapper[4802]: I1206 04:01:40.430929 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b08151ce-4f6b-4817-a2f4-8d7eb3102ce7-run-httpd\") pod \"ceilometer-0\" (UID: \"b08151ce-4f6b-4817-a2f4-8d7eb3102ce7\") " pod="openstack/ceilometer-0" Dec 06 04:01:40 crc kubenswrapper[4802]: I1206 04:01:40.430986 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b08151ce-4f6b-4817-a2f4-8d7eb3102ce7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b08151ce-4f6b-4817-a2f4-8d7eb3102ce7\") " pod="openstack/ceilometer-0" Dec 06 04:01:40 crc kubenswrapper[4802]: I1206 04:01:40.431975 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b08151ce-4f6b-4817-a2f4-8d7eb3102ce7-log-httpd\") pod \"ceilometer-0\" (UID: \"b08151ce-4f6b-4817-a2f4-8d7eb3102ce7\") " pod="openstack/ceilometer-0" Dec 06 04:01:40 crc kubenswrapper[4802]: I1206 04:01:40.432412 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b08151ce-4f6b-4817-a2f4-8d7eb3102ce7-run-httpd\") pod \"ceilometer-0\" (UID: \"b08151ce-4f6b-4817-a2f4-8d7eb3102ce7\") " pod="openstack/ceilometer-0" Dec 06 04:01:40 crc kubenswrapper[4802]: I1206 04:01:40.434458 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b08151ce-4f6b-4817-a2f4-8d7eb3102ce7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b08151ce-4f6b-4817-a2f4-8d7eb3102ce7\") " pod="openstack/ceilometer-0" Dec 06 04:01:40 crc kubenswrapper[4802]: I1206 04:01:40.435065 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b08151ce-4f6b-4817-a2f4-8d7eb3102ce7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b08151ce-4f6b-4817-a2f4-8d7eb3102ce7\") " pod="openstack/ceilometer-0" Dec 06 04:01:40 crc kubenswrapper[4802]: I1206 04:01:40.435355 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b08151ce-4f6b-4817-a2f4-8d7eb3102ce7-scripts\") pod \"ceilometer-0\" (UID: \"b08151ce-4f6b-4817-a2f4-8d7eb3102ce7\") " pod="openstack/ceilometer-0" Dec 06 04:01:40 crc kubenswrapper[4802]: I1206 04:01:40.436379 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b08151ce-4f6b-4817-a2f4-8d7eb3102ce7-config-data\") pod \"ceilometer-0\" (UID: \"b08151ce-4f6b-4817-a2f4-8d7eb3102ce7\") " pod="openstack/ceilometer-0" Dec 06 04:01:40 crc kubenswrapper[4802]: I1206 04:01:40.438453 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b08151ce-4f6b-4817-a2f4-8d7eb3102ce7-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"b08151ce-4f6b-4817-a2f4-8d7eb3102ce7\") " pod="openstack/ceilometer-0" Dec 06 04:01:40 crc kubenswrapper[4802]: I1206 04:01:40.449629 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbwmm\" (UniqueName: \"kubernetes.io/projected/b08151ce-4f6b-4817-a2f4-8d7eb3102ce7-kube-api-access-hbwmm\") pod \"ceilometer-0\" (UID: \"b08151ce-4f6b-4817-a2f4-8d7eb3102ce7\") " pod="openstack/ceilometer-0" Dec 06 04:01:40 crc kubenswrapper[4802]: I1206 04:01:40.476422 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 04:01:41 crc kubenswrapper[4802]: I1206 04:01:41.023061 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 04:01:41 crc kubenswrapper[4802]: W1206 04:01:41.026649 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb08151ce_4f6b_4817_a2f4_8d7eb3102ce7.slice/crio-1e1abe516cd539856d1b1e1bb73ad1b318058826d2b4e1cd677f911c70ec46b7 WatchSource:0}: Error finding container 1e1abe516cd539856d1b1e1bb73ad1b318058826d2b4e1cd677f911c70ec46b7: Status 404 returned error can't find the container with id 1e1abe516cd539856d1b1e1bb73ad1b318058826d2b4e1cd677f911c70ec46b7 Dec 06 04:01:41 crc kubenswrapper[4802]: I1206 04:01:41.078410 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b08151ce-4f6b-4817-a2f4-8d7eb3102ce7","Type":"ContainerStarted","Data":"1e1abe516cd539856d1b1e1bb73ad1b318058826d2b4e1cd677f911c70ec46b7"} Dec 06 04:01:41 crc kubenswrapper[4802]: I1206 04:01:41.461202 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30152cc8-8968-443f-923a-0ddca8c016ee" path="/var/lib/kubelet/pods/30152cc8-8968-443f-923a-0ddca8c016ee/volumes" Dec 06 04:01:41 crc kubenswrapper[4802]: I1206 04:01:41.599287 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 06 04:01:41 crc kubenswrapper[4802]: I1206 04:01:41.599360 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 06 04:01:41 crc kubenswrapper[4802]: I1206 04:01:41.608812 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 06 04:01:41 crc kubenswrapper[4802]: I1206 04:01:41.664594 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 06 04:01:42 crc kubenswrapper[4802]: I1206 04:01:42.122039 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 06 04:01:42 crc kubenswrapper[4802]: I1206 04:01:42.681279 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ee1f735d-c438-4b44-be26-cd9b4a1b7906" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.202:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 04:01:42 crc kubenswrapper[4802]: I1206 04:01:42.682074 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ee1f735d-c438-4b44-be26-cd9b4a1b7906" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.202:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 04:01:43 crc kubenswrapper[4802]: I1206 04:01:43.098589 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b08151ce-4f6b-4817-a2f4-8d7eb3102ce7","Type":"ContainerStarted","Data":"2820c9995ecbda04e1b26777b309755cbfe16dc0e7d2db9f3f465cce37e3e884"} Dec 06 04:01:43 crc kubenswrapper[4802]: I1206 04:01:43.098919 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b08151ce-4f6b-4817-a2f4-8d7eb3102ce7","Type":"ContainerStarted","Data":"698c9c18b88905aca9a2c9a61bc63976565cc974e35e045171604f2d2dcbbe06"} Dec 06 04:01:43 crc kubenswrapper[4802]: I1206 04:01:43.283181 4802 patch_prober.go:28] interesting pod/machine-config-daemon-zpxxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:01:43 crc kubenswrapper[4802]: I1206 04:01:43.283258 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:01:44 crc kubenswrapper[4802]: I1206 04:01:44.108986 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b08151ce-4f6b-4817-a2f4-8d7eb3102ce7","Type":"ContainerStarted","Data":"c3547662bdaf8f7ab818f100714f18c0aa113040e437ae41ca1dcb58231774f3"} Dec 06 04:01:46 crc kubenswrapper[4802]: I1206 04:01:46.140505 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b08151ce-4f6b-4817-a2f4-8d7eb3102ce7","Type":"ContainerStarted","Data":"85b0f8aedc15de534c8216b1fccadeeab3ca2478f155f6b7b6e9a9d3329a792f"} Dec 06 04:01:46 crc kubenswrapper[4802]: I1206 04:01:46.141355 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 06 04:01:46 crc kubenswrapper[4802]: I1206 04:01:46.173849 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.16560409 podStartE2EDuration="6.173830231s" podCreationTimestamp="2025-12-06 04:01:40 +0000 UTC" firstStartedPulling="2025-12-06 04:01:41.029708722 +0000 UTC m=+1293.901617874" lastFinishedPulling="2025-12-06 04:01:45.037934863 +0000 UTC m=+1297.909844015" observedRunningTime="2025-12-06 04:01:46.16832056 +0000 UTC m=+1299.040229722" watchObservedRunningTime="2025-12-06 04:01:46.173830231 +0000 UTC m=+1299.045739383" Dec 06 04:01:46 crc kubenswrapper[4802]: I1206 04:01:46.252927 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 06 04:01:46 crc kubenswrapper[4802]: I1206 04:01:46.259058 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 06 04:01:46 crc kubenswrapper[4802]: I1206 04:01:46.263310 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 06 04:01:46 crc kubenswrapper[4802]: I1206 04:01:46.357282 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 06 04:01:47 crc kubenswrapper[4802]: I1206 04:01:47.154604 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 06 04:01:50 crc kubenswrapper[4802]: I1206 04:01:50.176465 4802 generic.go:334] "Generic (PLEG): container finished" podID="ab782538-46b6-4426-9bb4-e6aba3007da6" containerID="72d35a154ed99c981e13031caa104f30cd26dbfbe75fcce0172353ebb84806eb" exitCode=137 Dec 06 04:01:50 crc kubenswrapper[4802]: I1206 04:01:50.176633 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ab782538-46b6-4426-9bb4-e6aba3007da6","Type":"ContainerDied","Data":"72d35a154ed99c981e13031caa104f30cd26dbfbe75fcce0172353ebb84806eb"} Dec 06 04:01:50 crc kubenswrapper[4802]: I1206 04:01:50.550589 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 06 04:01:50 crc kubenswrapper[4802]: I1206 04:01:50.638251 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab782538-46b6-4426-9bb4-e6aba3007da6-config-data\") pod \"ab782538-46b6-4426-9bb4-e6aba3007da6\" (UID: \"ab782538-46b6-4426-9bb4-e6aba3007da6\") " Dec 06 04:01:50 crc kubenswrapper[4802]: I1206 04:01:50.638291 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qhsd4\" (UniqueName: \"kubernetes.io/projected/ab782538-46b6-4426-9bb4-e6aba3007da6-kube-api-access-qhsd4\") pod \"ab782538-46b6-4426-9bb4-e6aba3007da6\" (UID: \"ab782538-46b6-4426-9bb4-e6aba3007da6\") " Dec 06 04:01:50 crc kubenswrapper[4802]: I1206 04:01:50.638500 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab782538-46b6-4426-9bb4-e6aba3007da6-combined-ca-bundle\") pod \"ab782538-46b6-4426-9bb4-e6aba3007da6\" (UID: \"ab782538-46b6-4426-9bb4-e6aba3007da6\") " Dec 06 04:01:50 crc kubenswrapper[4802]: I1206 04:01:50.644720 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab782538-46b6-4426-9bb4-e6aba3007da6-kube-api-access-qhsd4" (OuterVolumeSpecName: "kube-api-access-qhsd4") pod "ab782538-46b6-4426-9bb4-e6aba3007da6" (UID: "ab782538-46b6-4426-9bb4-e6aba3007da6"). InnerVolumeSpecName "kube-api-access-qhsd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:01:50 crc kubenswrapper[4802]: I1206 04:01:50.666077 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab782538-46b6-4426-9bb4-e6aba3007da6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ab782538-46b6-4426-9bb4-e6aba3007da6" (UID: "ab782538-46b6-4426-9bb4-e6aba3007da6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:01:50 crc kubenswrapper[4802]: I1206 04:01:50.678129 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab782538-46b6-4426-9bb4-e6aba3007da6-config-data" (OuterVolumeSpecName: "config-data") pod "ab782538-46b6-4426-9bb4-e6aba3007da6" (UID: "ab782538-46b6-4426-9bb4-e6aba3007da6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:01:50 crc kubenswrapper[4802]: I1206 04:01:50.740701 4802 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab782538-46b6-4426-9bb4-e6aba3007da6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:50 crc kubenswrapper[4802]: I1206 04:01:50.740732 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab782538-46b6-4426-9bb4-e6aba3007da6-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:50 crc kubenswrapper[4802]: I1206 04:01:50.740741 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qhsd4\" (UniqueName: \"kubernetes.io/projected/ab782538-46b6-4426-9bb4-e6aba3007da6-kube-api-access-qhsd4\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:51 crc kubenswrapper[4802]: I1206 04:01:51.192180 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ab782538-46b6-4426-9bb4-e6aba3007da6","Type":"ContainerDied","Data":"4ff9e1520cc1aca2a35c58fafa7add7422a74d0b6dbebcb5b02df899ec2306c0"} Dec 06 04:01:51 crc kubenswrapper[4802]: I1206 04:01:51.192235 4802 scope.go:117] "RemoveContainer" containerID="72d35a154ed99c981e13031caa104f30cd26dbfbe75fcce0172353ebb84806eb" Dec 06 04:01:51 crc kubenswrapper[4802]: I1206 04:01:51.192249 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 06 04:01:51 crc kubenswrapper[4802]: I1206 04:01:51.240087 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 04:01:51 crc kubenswrapper[4802]: I1206 04:01:51.252955 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 04:01:51 crc kubenswrapper[4802]: I1206 04:01:51.280651 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 04:01:51 crc kubenswrapper[4802]: E1206 04:01:51.281157 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab782538-46b6-4426-9bb4-e6aba3007da6" containerName="nova-cell1-novncproxy-novncproxy" Dec 06 04:01:51 crc kubenswrapper[4802]: I1206 04:01:51.281178 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab782538-46b6-4426-9bb4-e6aba3007da6" containerName="nova-cell1-novncproxy-novncproxy" Dec 06 04:01:51 crc kubenswrapper[4802]: I1206 04:01:51.281384 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab782538-46b6-4426-9bb4-e6aba3007da6" containerName="nova-cell1-novncproxy-novncproxy" Dec 06 04:01:51 crc kubenswrapper[4802]: I1206 04:01:51.281946 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 06 04:01:51 crc kubenswrapper[4802]: I1206 04:01:51.284026 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 06 04:01:51 crc kubenswrapper[4802]: I1206 04:01:51.284484 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 06 04:01:51 crc kubenswrapper[4802]: I1206 04:01:51.284642 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 06 04:01:51 crc kubenswrapper[4802]: I1206 04:01:51.292896 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 04:01:51 crc kubenswrapper[4802]: I1206 04:01:51.350054 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ede13563-3589-4889-a568-55af534782c5-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ede13563-3589-4889-a568-55af534782c5\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 04:01:51 crc kubenswrapper[4802]: I1206 04:01:51.350128 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/ede13563-3589-4889-a568-55af534782c5-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ede13563-3589-4889-a568-55af534782c5\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 04:01:51 crc kubenswrapper[4802]: I1206 04:01:51.350336 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/ede13563-3589-4889-a568-55af534782c5-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ede13563-3589-4889-a568-55af534782c5\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 04:01:51 crc kubenswrapper[4802]: I1206 04:01:51.350436 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ztndn\" (UniqueName: \"kubernetes.io/projected/ede13563-3589-4889-a568-55af534782c5-kube-api-access-ztndn\") pod \"nova-cell1-novncproxy-0\" (UID: \"ede13563-3589-4889-a568-55af534782c5\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 04:01:51 crc kubenswrapper[4802]: I1206 04:01:51.350496 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ede13563-3589-4889-a568-55af534782c5-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ede13563-3589-4889-a568-55af534782c5\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 04:01:51 crc kubenswrapper[4802]: I1206 04:01:51.451625 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ede13563-3589-4889-a568-55af534782c5-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ede13563-3589-4889-a568-55af534782c5\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 04:01:51 crc kubenswrapper[4802]: I1206 04:01:51.451687 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/ede13563-3589-4889-a568-55af534782c5-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ede13563-3589-4889-a568-55af534782c5\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 04:01:51 crc kubenswrapper[4802]: I1206 04:01:51.451719 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/ede13563-3589-4889-a568-55af534782c5-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ede13563-3589-4889-a568-55af534782c5\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 04:01:51 crc kubenswrapper[4802]: I1206 04:01:51.451764 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ztndn\" (UniqueName: \"kubernetes.io/projected/ede13563-3589-4889-a568-55af534782c5-kube-api-access-ztndn\") pod \"nova-cell1-novncproxy-0\" (UID: \"ede13563-3589-4889-a568-55af534782c5\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 04:01:51 crc kubenswrapper[4802]: I1206 04:01:51.451810 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ede13563-3589-4889-a568-55af534782c5-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ede13563-3589-4889-a568-55af534782c5\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 04:01:51 crc kubenswrapper[4802]: I1206 04:01:51.455770 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ede13563-3589-4889-a568-55af534782c5-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ede13563-3589-4889-a568-55af534782c5\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 04:01:51 crc kubenswrapper[4802]: I1206 04:01:51.456094 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/ede13563-3589-4889-a568-55af534782c5-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ede13563-3589-4889-a568-55af534782c5\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 04:01:51 crc kubenswrapper[4802]: I1206 04:01:51.456209 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/ede13563-3589-4889-a568-55af534782c5-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ede13563-3589-4889-a568-55af534782c5\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 04:01:51 crc kubenswrapper[4802]: I1206 04:01:51.458164 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ede13563-3589-4889-a568-55af534782c5-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ede13563-3589-4889-a568-55af534782c5\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 04:01:51 crc kubenswrapper[4802]: I1206 04:01:51.478332 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab782538-46b6-4426-9bb4-e6aba3007da6" path="/var/lib/kubelet/pods/ab782538-46b6-4426-9bb4-e6aba3007da6/volumes" Dec 06 04:01:51 crc kubenswrapper[4802]: I1206 04:01:51.479612 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ztndn\" (UniqueName: \"kubernetes.io/projected/ede13563-3589-4889-a568-55af534782c5-kube-api-access-ztndn\") pod \"nova-cell1-novncproxy-0\" (UID: \"ede13563-3589-4889-a568-55af534782c5\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 04:01:51 crc kubenswrapper[4802]: I1206 04:01:51.597523 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 06 04:01:51 crc kubenswrapper[4802]: I1206 04:01:51.684085 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 06 04:01:51 crc kubenswrapper[4802]: I1206 04:01:51.686052 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 06 04:01:51 crc kubenswrapper[4802]: I1206 04:01:51.723695 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 06 04:01:51 crc kubenswrapper[4802]: I1206 04:01:51.744270 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 06 04:01:52 crc kubenswrapper[4802]: W1206 04:01:52.150220 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podede13563_3589_4889_a568_55af534782c5.slice/crio-d448926b8d0bfc3b9e5750e88f78b0d56c5d3f907497bf13ae635c8761c1d90a WatchSource:0}: Error finding container d448926b8d0bfc3b9e5750e88f78b0d56c5d3f907497bf13ae635c8761c1d90a: Status 404 returned error can't find the container with id d448926b8d0bfc3b9e5750e88f78b0d56c5d3f907497bf13ae635c8761c1d90a Dec 06 04:01:52 crc kubenswrapper[4802]: I1206 04:01:52.153844 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 04:01:52 crc kubenswrapper[4802]: I1206 04:01:52.206146 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ede13563-3589-4889-a568-55af534782c5","Type":"ContainerStarted","Data":"d448926b8d0bfc3b9e5750e88f78b0d56c5d3f907497bf13ae635c8761c1d90a"} Dec 06 04:01:52 crc kubenswrapper[4802]: I1206 04:01:52.206196 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 06 04:01:52 crc kubenswrapper[4802]: I1206 04:01:52.213009 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 06 04:01:52 crc kubenswrapper[4802]: I1206 04:01:52.360812 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-f84f9ccf-fc42d"] Dec 06 04:01:52 crc kubenswrapper[4802]: I1206 04:01:52.363606 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f84f9ccf-fc42d" Dec 06 04:01:52 crc kubenswrapper[4802]: I1206 04:01:52.389171 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f84f9ccf-fc42d"] Dec 06 04:01:52 crc kubenswrapper[4802]: I1206 04:01:52.469028 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d01257b6-a8d4-41e1-b0a0-894cba15e527-ovsdbserver-sb\") pod \"dnsmasq-dns-f84f9ccf-fc42d\" (UID: \"d01257b6-a8d4-41e1-b0a0-894cba15e527\") " pod="openstack/dnsmasq-dns-f84f9ccf-fc42d" Dec 06 04:01:52 crc kubenswrapper[4802]: I1206 04:01:52.469136 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d01257b6-a8d4-41e1-b0a0-894cba15e527-dns-swift-storage-0\") pod \"dnsmasq-dns-f84f9ccf-fc42d\" (UID: \"d01257b6-a8d4-41e1-b0a0-894cba15e527\") " pod="openstack/dnsmasq-dns-f84f9ccf-fc42d" Dec 06 04:01:52 crc kubenswrapper[4802]: I1206 04:01:52.469171 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d01257b6-a8d4-41e1-b0a0-894cba15e527-config\") pod \"dnsmasq-dns-f84f9ccf-fc42d\" (UID: \"d01257b6-a8d4-41e1-b0a0-894cba15e527\") " pod="openstack/dnsmasq-dns-f84f9ccf-fc42d" Dec 06 04:01:52 crc kubenswrapper[4802]: I1206 04:01:52.469195 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d01257b6-a8d4-41e1-b0a0-894cba15e527-dns-svc\") pod \"dnsmasq-dns-f84f9ccf-fc42d\" (UID: \"d01257b6-a8d4-41e1-b0a0-894cba15e527\") " pod="openstack/dnsmasq-dns-f84f9ccf-fc42d" Dec 06 04:01:52 crc kubenswrapper[4802]: I1206 04:01:52.469231 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d01257b6-a8d4-41e1-b0a0-894cba15e527-ovsdbserver-nb\") pod \"dnsmasq-dns-f84f9ccf-fc42d\" (UID: \"d01257b6-a8d4-41e1-b0a0-894cba15e527\") " pod="openstack/dnsmasq-dns-f84f9ccf-fc42d" Dec 06 04:01:52 crc kubenswrapper[4802]: I1206 04:01:52.469252 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqzk7\" (UniqueName: \"kubernetes.io/projected/d01257b6-a8d4-41e1-b0a0-894cba15e527-kube-api-access-jqzk7\") pod \"dnsmasq-dns-f84f9ccf-fc42d\" (UID: \"d01257b6-a8d4-41e1-b0a0-894cba15e527\") " pod="openstack/dnsmasq-dns-f84f9ccf-fc42d" Dec 06 04:01:52 crc kubenswrapper[4802]: I1206 04:01:52.570561 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d01257b6-a8d4-41e1-b0a0-894cba15e527-ovsdbserver-sb\") pod \"dnsmasq-dns-f84f9ccf-fc42d\" (UID: \"d01257b6-a8d4-41e1-b0a0-894cba15e527\") " pod="openstack/dnsmasq-dns-f84f9ccf-fc42d" Dec 06 04:01:52 crc kubenswrapper[4802]: I1206 04:01:52.570676 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d01257b6-a8d4-41e1-b0a0-894cba15e527-dns-swift-storage-0\") pod \"dnsmasq-dns-f84f9ccf-fc42d\" (UID: \"d01257b6-a8d4-41e1-b0a0-894cba15e527\") " pod="openstack/dnsmasq-dns-f84f9ccf-fc42d" Dec 06 04:01:52 crc kubenswrapper[4802]: I1206 04:01:52.570709 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d01257b6-a8d4-41e1-b0a0-894cba15e527-config\") pod \"dnsmasq-dns-f84f9ccf-fc42d\" (UID: \"d01257b6-a8d4-41e1-b0a0-894cba15e527\") " pod="openstack/dnsmasq-dns-f84f9ccf-fc42d" Dec 06 04:01:52 crc kubenswrapper[4802]: I1206 04:01:52.570734 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d01257b6-a8d4-41e1-b0a0-894cba15e527-dns-svc\") pod \"dnsmasq-dns-f84f9ccf-fc42d\" (UID: \"d01257b6-a8d4-41e1-b0a0-894cba15e527\") " pod="openstack/dnsmasq-dns-f84f9ccf-fc42d" Dec 06 04:01:52 crc kubenswrapper[4802]: I1206 04:01:52.570785 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d01257b6-a8d4-41e1-b0a0-894cba15e527-ovsdbserver-nb\") pod \"dnsmasq-dns-f84f9ccf-fc42d\" (UID: \"d01257b6-a8d4-41e1-b0a0-894cba15e527\") " pod="openstack/dnsmasq-dns-f84f9ccf-fc42d" Dec 06 04:01:52 crc kubenswrapper[4802]: I1206 04:01:52.570822 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqzk7\" (UniqueName: \"kubernetes.io/projected/d01257b6-a8d4-41e1-b0a0-894cba15e527-kube-api-access-jqzk7\") pod \"dnsmasq-dns-f84f9ccf-fc42d\" (UID: \"d01257b6-a8d4-41e1-b0a0-894cba15e527\") " pod="openstack/dnsmasq-dns-f84f9ccf-fc42d" Dec 06 04:01:52 crc kubenswrapper[4802]: I1206 04:01:52.571915 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d01257b6-a8d4-41e1-b0a0-894cba15e527-dns-svc\") pod \"dnsmasq-dns-f84f9ccf-fc42d\" (UID: \"d01257b6-a8d4-41e1-b0a0-894cba15e527\") " pod="openstack/dnsmasq-dns-f84f9ccf-fc42d" Dec 06 04:01:52 crc kubenswrapper[4802]: I1206 04:01:52.572100 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d01257b6-a8d4-41e1-b0a0-894cba15e527-ovsdbserver-nb\") pod \"dnsmasq-dns-f84f9ccf-fc42d\" (UID: \"d01257b6-a8d4-41e1-b0a0-894cba15e527\") " pod="openstack/dnsmasq-dns-f84f9ccf-fc42d" Dec 06 04:01:52 crc kubenswrapper[4802]: I1206 04:01:52.572207 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d01257b6-a8d4-41e1-b0a0-894cba15e527-ovsdbserver-sb\") pod \"dnsmasq-dns-f84f9ccf-fc42d\" (UID: \"d01257b6-a8d4-41e1-b0a0-894cba15e527\") " pod="openstack/dnsmasq-dns-f84f9ccf-fc42d" Dec 06 04:01:52 crc kubenswrapper[4802]: I1206 04:01:52.572285 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d01257b6-a8d4-41e1-b0a0-894cba15e527-dns-swift-storage-0\") pod \"dnsmasq-dns-f84f9ccf-fc42d\" (UID: \"d01257b6-a8d4-41e1-b0a0-894cba15e527\") " pod="openstack/dnsmasq-dns-f84f9ccf-fc42d" Dec 06 04:01:52 crc kubenswrapper[4802]: I1206 04:01:52.572370 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d01257b6-a8d4-41e1-b0a0-894cba15e527-config\") pod \"dnsmasq-dns-f84f9ccf-fc42d\" (UID: \"d01257b6-a8d4-41e1-b0a0-894cba15e527\") " pod="openstack/dnsmasq-dns-f84f9ccf-fc42d" Dec 06 04:01:52 crc kubenswrapper[4802]: I1206 04:01:52.587460 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqzk7\" (UniqueName: \"kubernetes.io/projected/d01257b6-a8d4-41e1-b0a0-894cba15e527-kube-api-access-jqzk7\") pod \"dnsmasq-dns-f84f9ccf-fc42d\" (UID: \"d01257b6-a8d4-41e1-b0a0-894cba15e527\") " pod="openstack/dnsmasq-dns-f84f9ccf-fc42d" Dec 06 04:01:52 crc kubenswrapper[4802]: I1206 04:01:52.718118 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f84f9ccf-fc42d" Dec 06 04:01:53 crc kubenswrapper[4802]: I1206 04:01:53.206971 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f84f9ccf-fc42d"] Dec 06 04:01:53 crc kubenswrapper[4802]: I1206 04:01:53.236939 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ede13563-3589-4889-a568-55af534782c5","Type":"ContainerStarted","Data":"96c1e7ec095556decea946294d85e8985980a50576c2ce6f9d9762a48f750c79"} Dec 06 04:01:53 crc kubenswrapper[4802]: I1206 04:01:53.271086 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.271065138 podStartE2EDuration="2.271065138s" podCreationTimestamp="2025-12-06 04:01:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:01:53.254983708 +0000 UTC m=+1306.126892860" watchObservedRunningTime="2025-12-06 04:01:53.271065138 +0000 UTC m=+1306.142974290" Dec 06 04:01:54 crc kubenswrapper[4802]: I1206 04:01:54.249277 4802 generic.go:334] "Generic (PLEG): container finished" podID="d01257b6-a8d4-41e1-b0a0-894cba15e527" containerID="e407afae41d5750349a249f62b1c86ddc5f47cb7d5a7e26a5747658c452b8613" exitCode=0 Dec 06 04:01:54 crc kubenswrapper[4802]: I1206 04:01:54.249536 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f84f9ccf-fc42d" event={"ID":"d01257b6-a8d4-41e1-b0a0-894cba15e527","Type":"ContainerDied","Data":"e407afae41d5750349a249f62b1c86ddc5f47cb7d5a7e26a5747658c452b8613"} Dec 06 04:01:54 crc kubenswrapper[4802]: I1206 04:01:54.249711 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f84f9ccf-fc42d" event={"ID":"d01257b6-a8d4-41e1-b0a0-894cba15e527","Type":"ContainerStarted","Data":"2640aee878c0a7c70a9250585c44ea065e7ca7fdb683890442847ce021c6de29"} Dec 06 04:01:54 crc kubenswrapper[4802]: I1206 04:01:54.463720 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 04:01:54 crc kubenswrapper[4802]: I1206 04:01:54.464238 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b08151ce-4f6b-4817-a2f4-8d7eb3102ce7" containerName="ceilometer-central-agent" containerID="cri-o://698c9c18b88905aca9a2c9a61bc63976565cc974e35e045171604f2d2dcbbe06" gracePeriod=30 Dec 06 04:01:54 crc kubenswrapper[4802]: I1206 04:01:54.464303 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b08151ce-4f6b-4817-a2f4-8d7eb3102ce7" containerName="proxy-httpd" containerID="cri-o://85b0f8aedc15de534c8216b1fccadeeab3ca2478f155f6b7b6e9a9d3329a792f" gracePeriod=30 Dec 06 04:01:54 crc kubenswrapper[4802]: I1206 04:01:54.464350 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b08151ce-4f6b-4817-a2f4-8d7eb3102ce7" containerName="ceilometer-notification-agent" containerID="cri-o://2820c9995ecbda04e1b26777b309755cbfe16dc0e7d2db9f3f465cce37e3e884" gracePeriod=30 Dec 06 04:01:54 crc kubenswrapper[4802]: I1206 04:01:54.464353 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b08151ce-4f6b-4817-a2f4-8d7eb3102ce7" containerName="sg-core" containerID="cri-o://c3547662bdaf8f7ab818f100714f18c0aa113040e437ae41ca1dcb58231774f3" gracePeriod=30 Dec 06 04:01:54 crc kubenswrapper[4802]: I1206 04:01:54.987976 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 06 04:01:55 crc kubenswrapper[4802]: I1206 04:01:55.260132 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f84f9ccf-fc42d" event={"ID":"d01257b6-a8d4-41e1-b0a0-894cba15e527","Type":"ContainerStarted","Data":"3b8fd0883832b00c8e9451fb374667099110f5304d39b4dcaeac3002f8ef78c9"} Dec 06 04:01:55 crc kubenswrapper[4802]: I1206 04:01:55.260499 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-f84f9ccf-fc42d" Dec 06 04:01:55 crc kubenswrapper[4802]: I1206 04:01:55.262849 4802 generic.go:334] "Generic (PLEG): container finished" podID="b08151ce-4f6b-4817-a2f4-8d7eb3102ce7" containerID="85b0f8aedc15de534c8216b1fccadeeab3ca2478f155f6b7b6e9a9d3329a792f" exitCode=0 Dec 06 04:01:55 crc kubenswrapper[4802]: I1206 04:01:55.262878 4802 generic.go:334] "Generic (PLEG): container finished" podID="b08151ce-4f6b-4817-a2f4-8d7eb3102ce7" containerID="c3547662bdaf8f7ab818f100714f18c0aa113040e437ae41ca1dcb58231774f3" exitCode=2 Dec 06 04:01:55 crc kubenswrapper[4802]: I1206 04:01:55.262886 4802 generic.go:334] "Generic (PLEG): container finished" podID="b08151ce-4f6b-4817-a2f4-8d7eb3102ce7" containerID="698c9c18b88905aca9a2c9a61bc63976565cc974e35e045171604f2d2dcbbe06" exitCode=0 Dec 06 04:01:55 crc kubenswrapper[4802]: I1206 04:01:55.262916 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b08151ce-4f6b-4817-a2f4-8d7eb3102ce7","Type":"ContainerDied","Data":"85b0f8aedc15de534c8216b1fccadeeab3ca2478f155f6b7b6e9a9d3329a792f"} Dec 06 04:01:55 crc kubenswrapper[4802]: I1206 04:01:55.262951 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b08151ce-4f6b-4817-a2f4-8d7eb3102ce7","Type":"ContainerDied","Data":"c3547662bdaf8f7ab818f100714f18c0aa113040e437ae41ca1dcb58231774f3"} Dec 06 04:01:55 crc kubenswrapper[4802]: I1206 04:01:55.262964 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b08151ce-4f6b-4817-a2f4-8d7eb3102ce7","Type":"ContainerDied","Data":"698c9c18b88905aca9a2c9a61bc63976565cc974e35e045171604f2d2dcbbe06"} Dec 06 04:01:55 crc kubenswrapper[4802]: I1206 04:01:55.263058 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ee1f735d-c438-4b44-be26-cd9b4a1b7906" containerName="nova-api-log" containerID="cri-o://5c17109631f662ffba99b75d4d915c33f7c4dd58a15f40333caeaf79ddfccb03" gracePeriod=30 Dec 06 04:01:55 crc kubenswrapper[4802]: I1206 04:01:55.263101 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ee1f735d-c438-4b44-be26-cd9b4a1b7906" containerName="nova-api-api" containerID="cri-o://6bec787747e626d59fa9cfe8bfac461664d5016602e4ae4ac2a77193d517b046" gracePeriod=30 Dec 06 04:01:55 crc kubenswrapper[4802]: I1206 04:01:55.288338 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-f84f9ccf-fc42d" podStartSLOduration=3.288320622 podStartE2EDuration="3.288320622s" podCreationTimestamp="2025-12-06 04:01:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:01:55.285047373 +0000 UTC m=+1308.156956545" watchObservedRunningTime="2025-12-06 04:01:55.288320622 +0000 UTC m=+1308.160229784" Dec 06 04:01:56 crc kubenswrapper[4802]: I1206 04:01:56.285883 4802 generic.go:334] "Generic (PLEG): container finished" podID="ee1f735d-c438-4b44-be26-cd9b4a1b7906" containerID="5c17109631f662ffba99b75d4d915c33f7c4dd58a15f40333caeaf79ddfccb03" exitCode=143 Dec 06 04:01:56 crc kubenswrapper[4802]: I1206 04:01:56.285942 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ee1f735d-c438-4b44-be26-cd9b4a1b7906","Type":"ContainerDied","Data":"5c17109631f662ffba99b75d4d915c33f7c4dd58a15f40333caeaf79ddfccb03"} Dec 06 04:01:56 crc kubenswrapper[4802]: I1206 04:01:56.597839 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.108293 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.226641 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.292901 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wc7r5\" (UniqueName: \"kubernetes.io/projected/ee1f735d-c438-4b44-be26-cd9b4a1b7906-kube-api-access-wc7r5\") pod \"ee1f735d-c438-4b44-be26-cd9b4a1b7906\" (UID: \"ee1f735d-c438-4b44-be26-cd9b4a1b7906\") " Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.293320 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ee1f735d-c438-4b44-be26-cd9b4a1b7906-logs\") pod \"ee1f735d-c438-4b44-be26-cd9b4a1b7906\" (UID: \"ee1f735d-c438-4b44-be26-cd9b4a1b7906\") " Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.293367 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee1f735d-c438-4b44-be26-cd9b4a1b7906-combined-ca-bundle\") pod \"ee1f735d-c438-4b44-be26-cd9b4a1b7906\" (UID: \"ee1f735d-c438-4b44-be26-cd9b4a1b7906\") " Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.293384 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee1f735d-c438-4b44-be26-cd9b4a1b7906-config-data\") pod \"ee1f735d-c438-4b44-be26-cd9b4a1b7906\" (UID: \"ee1f735d-c438-4b44-be26-cd9b4a1b7906\") " Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.293799 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee1f735d-c438-4b44-be26-cd9b4a1b7906-logs" (OuterVolumeSpecName: "logs") pod "ee1f735d-c438-4b44-be26-cd9b4a1b7906" (UID: "ee1f735d-c438-4b44-be26-cd9b4a1b7906"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.298696 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee1f735d-c438-4b44-be26-cd9b4a1b7906-kube-api-access-wc7r5" (OuterVolumeSpecName: "kube-api-access-wc7r5") pod "ee1f735d-c438-4b44-be26-cd9b4a1b7906" (UID: "ee1f735d-c438-4b44-be26-cd9b4a1b7906"). InnerVolumeSpecName "kube-api-access-wc7r5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.326273 4802 generic.go:334] "Generic (PLEG): container finished" podID="ee1f735d-c438-4b44-be26-cd9b4a1b7906" containerID="6bec787747e626d59fa9cfe8bfac461664d5016602e4ae4ac2a77193d517b046" exitCode=0 Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.326336 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ee1f735d-c438-4b44-be26-cd9b4a1b7906","Type":"ContainerDied","Data":"6bec787747e626d59fa9cfe8bfac461664d5016602e4ae4ac2a77193d517b046"} Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.326406 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ee1f735d-c438-4b44-be26-cd9b4a1b7906","Type":"ContainerDied","Data":"f8349daf8172aec143b601a5b9bad0e1000c3f3921950fb0e77b47365ffe0052"} Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.326357 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.326436 4802 scope.go:117] "RemoveContainer" containerID="6bec787747e626d59fa9cfe8bfac461664d5016602e4ae4ac2a77193d517b046" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.327143 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee1f735d-c438-4b44-be26-cd9b4a1b7906-config-data" (OuterVolumeSpecName: "config-data") pod "ee1f735d-c438-4b44-be26-cd9b4a1b7906" (UID: "ee1f735d-c438-4b44-be26-cd9b4a1b7906"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.327733 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee1f735d-c438-4b44-be26-cd9b4a1b7906-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ee1f735d-c438-4b44-be26-cd9b4a1b7906" (UID: "ee1f735d-c438-4b44-be26-cd9b4a1b7906"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.329519 4802 generic.go:334] "Generic (PLEG): container finished" podID="b08151ce-4f6b-4817-a2f4-8d7eb3102ce7" containerID="2820c9995ecbda04e1b26777b309755cbfe16dc0e7d2db9f3f465cce37e3e884" exitCode=0 Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.329548 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b08151ce-4f6b-4817-a2f4-8d7eb3102ce7","Type":"ContainerDied","Data":"2820c9995ecbda04e1b26777b309755cbfe16dc0e7d2db9f3f465cce37e3e884"} Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.329572 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b08151ce-4f6b-4817-a2f4-8d7eb3102ce7","Type":"ContainerDied","Data":"1e1abe516cd539856d1b1e1bb73ad1b318058826d2b4e1cd677f911c70ec46b7"} Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.329633 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.345476 4802 scope.go:117] "RemoveContainer" containerID="5c17109631f662ffba99b75d4d915c33f7c4dd58a15f40333caeaf79ddfccb03" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.360323 4802 scope.go:117] "RemoveContainer" containerID="6bec787747e626d59fa9cfe8bfac461664d5016602e4ae4ac2a77193d517b046" Dec 06 04:01:59 crc kubenswrapper[4802]: E1206 04:01:59.360730 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6bec787747e626d59fa9cfe8bfac461664d5016602e4ae4ac2a77193d517b046\": container with ID starting with 6bec787747e626d59fa9cfe8bfac461664d5016602e4ae4ac2a77193d517b046 not found: ID does not exist" containerID="6bec787747e626d59fa9cfe8bfac461664d5016602e4ae4ac2a77193d517b046" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.360779 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6bec787747e626d59fa9cfe8bfac461664d5016602e4ae4ac2a77193d517b046"} err="failed to get container status \"6bec787747e626d59fa9cfe8bfac461664d5016602e4ae4ac2a77193d517b046\": rpc error: code = NotFound desc = could not find container \"6bec787747e626d59fa9cfe8bfac461664d5016602e4ae4ac2a77193d517b046\": container with ID starting with 6bec787747e626d59fa9cfe8bfac461664d5016602e4ae4ac2a77193d517b046 not found: ID does not exist" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.360808 4802 scope.go:117] "RemoveContainer" containerID="5c17109631f662ffba99b75d4d915c33f7c4dd58a15f40333caeaf79ddfccb03" Dec 06 04:01:59 crc kubenswrapper[4802]: E1206 04:01:59.361225 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c17109631f662ffba99b75d4d915c33f7c4dd58a15f40333caeaf79ddfccb03\": container with ID starting with 5c17109631f662ffba99b75d4d915c33f7c4dd58a15f40333caeaf79ddfccb03 not found: ID does not exist" containerID="5c17109631f662ffba99b75d4d915c33f7c4dd58a15f40333caeaf79ddfccb03" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.361250 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c17109631f662ffba99b75d4d915c33f7c4dd58a15f40333caeaf79ddfccb03"} err="failed to get container status \"5c17109631f662ffba99b75d4d915c33f7c4dd58a15f40333caeaf79ddfccb03\": rpc error: code = NotFound desc = could not find container \"5c17109631f662ffba99b75d4d915c33f7c4dd58a15f40333caeaf79ddfccb03\": container with ID starting with 5c17109631f662ffba99b75d4d915c33f7c4dd58a15f40333caeaf79ddfccb03 not found: ID does not exist" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.361264 4802 scope.go:117] "RemoveContainer" containerID="85b0f8aedc15de534c8216b1fccadeeab3ca2478f155f6b7b6e9a9d3329a792f" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.376896 4802 scope.go:117] "RemoveContainer" containerID="c3547662bdaf8f7ab818f100714f18c0aa113040e437ae41ca1dcb58231774f3" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.394706 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b08151ce-4f6b-4817-a2f4-8d7eb3102ce7-scripts\") pod \"b08151ce-4f6b-4817-a2f4-8d7eb3102ce7\" (UID: \"b08151ce-4f6b-4817-a2f4-8d7eb3102ce7\") " Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.394783 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b08151ce-4f6b-4817-a2f4-8d7eb3102ce7-ceilometer-tls-certs\") pod \"b08151ce-4f6b-4817-a2f4-8d7eb3102ce7\" (UID: \"b08151ce-4f6b-4817-a2f4-8d7eb3102ce7\") " Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.394842 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hbwmm\" (UniqueName: \"kubernetes.io/projected/b08151ce-4f6b-4817-a2f4-8d7eb3102ce7-kube-api-access-hbwmm\") pod \"b08151ce-4f6b-4817-a2f4-8d7eb3102ce7\" (UID: \"b08151ce-4f6b-4817-a2f4-8d7eb3102ce7\") " Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.394866 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b08151ce-4f6b-4817-a2f4-8d7eb3102ce7-combined-ca-bundle\") pod \"b08151ce-4f6b-4817-a2f4-8d7eb3102ce7\" (UID: \"b08151ce-4f6b-4817-a2f4-8d7eb3102ce7\") " Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.394890 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b08151ce-4f6b-4817-a2f4-8d7eb3102ce7-sg-core-conf-yaml\") pod \"b08151ce-4f6b-4817-a2f4-8d7eb3102ce7\" (UID: \"b08151ce-4f6b-4817-a2f4-8d7eb3102ce7\") " Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.394959 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b08151ce-4f6b-4817-a2f4-8d7eb3102ce7-config-data\") pod \"b08151ce-4f6b-4817-a2f4-8d7eb3102ce7\" (UID: \"b08151ce-4f6b-4817-a2f4-8d7eb3102ce7\") " Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.395103 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b08151ce-4f6b-4817-a2f4-8d7eb3102ce7-log-httpd\") pod \"b08151ce-4f6b-4817-a2f4-8d7eb3102ce7\" (UID: \"b08151ce-4f6b-4817-a2f4-8d7eb3102ce7\") " Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.395126 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b08151ce-4f6b-4817-a2f4-8d7eb3102ce7-run-httpd\") pod \"b08151ce-4f6b-4817-a2f4-8d7eb3102ce7\" (UID: \"b08151ce-4f6b-4817-a2f4-8d7eb3102ce7\") " Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.395536 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wc7r5\" (UniqueName: \"kubernetes.io/projected/ee1f735d-c438-4b44-be26-cd9b4a1b7906-kube-api-access-wc7r5\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.395559 4802 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ee1f735d-c438-4b44-be26-cd9b4a1b7906-logs\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.395574 4802 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee1f735d-c438-4b44-be26-cd9b4a1b7906-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.395586 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee1f735d-c438-4b44-be26-cd9b4a1b7906-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.396120 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b08151ce-4f6b-4817-a2f4-8d7eb3102ce7-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "b08151ce-4f6b-4817-a2f4-8d7eb3102ce7" (UID: "b08151ce-4f6b-4817-a2f4-8d7eb3102ce7"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.396192 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b08151ce-4f6b-4817-a2f4-8d7eb3102ce7-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "b08151ce-4f6b-4817-a2f4-8d7eb3102ce7" (UID: "b08151ce-4f6b-4817-a2f4-8d7eb3102ce7"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.399314 4802 scope.go:117] "RemoveContainer" containerID="2820c9995ecbda04e1b26777b309755cbfe16dc0e7d2db9f3f465cce37e3e884" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.399426 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b08151ce-4f6b-4817-a2f4-8d7eb3102ce7-scripts" (OuterVolumeSpecName: "scripts") pod "b08151ce-4f6b-4817-a2f4-8d7eb3102ce7" (UID: "b08151ce-4f6b-4817-a2f4-8d7eb3102ce7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.402992 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b08151ce-4f6b-4817-a2f4-8d7eb3102ce7-kube-api-access-hbwmm" (OuterVolumeSpecName: "kube-api-access-hbwmm") pod "b08151ce-4f6b-4817-a2f4-8d7eb3102ce7" (UID: "b08151ce-4f6b-4817-a2f4-8d7eb3102ce7"). InnerVolumeSpecName "kube-api-access-hbwmm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.416515 4802 scope.go:117] "RemoveContainer" containerID="698c9c18b88905aca9a2c9a61bc63976565cc974e35e045171604f2d2dcbbe06" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.427160 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b08151ce-4f6b-4817-a2f4-8d7eb3102ce7-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "b08151ce-4f6b-4817-a2f4-8d7eb3102ce7" (UID: "b08151ce-4f6b-4817-a2f4-8d7eb3102ce7"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.437056 4802 scope.go:117] "RemoveContainer" containerID="85b0f8aedc15de534c8216b1fccadeeab3ca2478f155f6b7b6e9a9d3329a792f" Dec 06 04:01:59 crc kubenswrapper[4802]: E1206 04:01:59.437540 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85b0f8aedc15de534c8216b1fccadeeab3ca2478f155f6b7b6e9a9d3329a792f\": container with ID starting with 85b0f8aedc15de534c8216b1fccadeeab3ca2478f155f6b7b6e9a9d3329a792f not found: ID does not exist" containerID="85b0f8aedc15de534c8216b1fccadeeab3ca2478f155f6b7b6e9a9d3329a792f" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.437582 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85b0f8aedc15de534c8216b1fccadeeab3ca2478f155f6b7b6e9a9d3329a792f"} err="failed to get container status \"85b0f8aedc15de534c8216b1fccadeeab3ca2478f155f6b7b6e9a9d3329a792f\": rpc error: code = NotFound desc = could not find container \"85b0f8aedc15de534c8216b1fccadeeab3ca2478f155f6b7b6e9a9d3329a792f\": container with ID starting with 85b0f8aedc15de534c8216b1fccadeeab3ca2478f155f6b7b6e9a9d3329a792f not found: ID does not exist" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.437610 4802 scope.go:117] "RemoveContainer" containerID="c3547662bdaf8f7ab818f100714f18c0aa113040e437ae41ca1dcb58231774f3" Dec 06 04:01:59 crc kubenswrapper[4802]: E1206 04:01:59.439147 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3547662bdaf8f7ab818f100714f18c0aa113040e437ae41ca1dcb58231774f3\": container with ID starting with c3547662bdaf8f7ab818f100714f18c0aa113040e437ae41ca1dcb58231774f3 not found: ID does not exist" containerID="c3547662bdaf8f7ab818f100714f18c0aa113040e437ae41ca1dcb58231774f3" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.439202 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3547662bdaf8f7ab818f100714f18c0aa113040e437ae41ca1dcb58231774f3"} err="failed to get container status \"c3547662bdaf8f7ab818f100714f18c0aa113040e437ae41ca1dcb58231774f3\": rpc error: code = NotFound desc = could not find container \"c3547662bdaf8f7ab818f100714f18c0aa113040e437ae41ca1dcb58231774f3\": container with ID starting with c3547662bdaf8f7ab818f100714f18c0aa113040e437ae41ca1dcb58231774f3 not found: ID does not exist" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.439233 4802 scope.go:117] "RemoveContainer" containerID="2820c9995ecbda04e1b26777b309755cbfe16dc0e7d2db9f3f465cce37e3e884" Dec 06 04:01:59 crc kubenswrapper[4802]: E1206 04:01:59.439537 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2820c9995ecbda04e1b26777b309755cbfe16dc0e7d2db9f3f465cce37e3e884\": container with ID starting with 2820c9995ecbda04e1b26777b309755cbfe16dc0e7d2db9f3f465cce37e3e884 not found: ID does not exist" containerID="2820c9995ecbda04e1b26777b309755cbfe16dc0e7d2db9f3f465cce37e3e884" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.439561 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2820c9995ecbda04e1b26777b309755cbfe16dc0e7d2db9f3f465cce37e3e884"} err="failed to get container status \"2820c9995ecbda04e1b26777b309755cbfe16dc0e7d2db9f3f465cce37e3e884\": rpc error: code = NotFound desc = could not find container \"2820c9995ecbda04e1b26777b309755cbfe16dc0e7d2db9f3f465cce37e3e884\": container with ID starting with 2820c9995ecbda04e1b26777b309755cbfe16dc0e7d2db9f3f465cce37e3e884 not found: ID does not exist" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.439577 4802 scope.go:117] "RemoveContainer" containerID="698c9c18b88905aca9a2c9a61bc63976565cc974e35e045171604f2d2dcbbe06" Dec 06 04:01:59 crc kubenswrapper[4802]: E1206 04:01:59.439815 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"698c9c18b88905aca9a2c9a61bc63976565cc974e35e045171604f2d2dcbbe06\": container with ID starting with 698c9c18b88905aca9a2c9a61bc63976565cc974e35e045171604f2d2dcbbe06 not found: ID does not exist" containerID="698c9c18b88905aca9a2c9a61bc63976565cc974e35e045171604f2d2dcbbe06" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.439839 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"698c9c18b88905aca9a2c9a61bc63976565cc974e35e045171604f2d2dcbbe06"} err="failed to get container status \"698c9c18b88905aca9a2c9a61bc63976565cc974e35e045171604f2d2dcbbe06\": rpc error: code = NotFound desc = could not find container \"698c9c18b88905aca9a2c9a61bc63976565cc974e35e045171604f2d2dcbbe06\": container with ID starting with 698c9c18b88905aca9a2c9a61bc63976565cc974e35e045171604f2d2dcbbe06 not found: ID does not exist" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.459563 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b08151ce-4f6b-4817-a2f4-8d7eb3102ce7-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "b08151ce-4f6b-4817-a2f4-8d7eb3102ce7" (UID: "b08151ce-4f6b-4817-a2f4-8d7eb3102ce7"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.485654 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b08151ce-4f6b-4817-a2f4-8d7eb3102ce7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b08151ce-4f6b-4817-a2f4-8d7eb3102ce7" (UID: "b08151ce-4f6b-4817-a2f4-8d7eb3102ce7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.496913 4802 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b08151ce-4f6b-4817-a2f4-8d7eb3102ce7-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.496944 4802 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b08151ce-4f6b-4817-a2f4-8d7eb3102ce7-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.496957 4802 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b08151ce-4f6b-4817-a2f4-8d7eb3102ce7-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.496968 4802 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b08151ce-4f6b-4817-a2f4-8d7eb3102ce7-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.496981 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hbwmm\" (UniqueName: \"kubernetes.io/projected/b08151ce-4f6b-4817-a2f4-8d7eb3102ce7-kube-api-access-hbwmm\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.496992 4802 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b08151ce-4f6b-4817-a2f4-8d7eb3102ce7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.497003 4802 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b08151ce-4f6b-4817-a2f4-8d7eb3102ce7-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.526106 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b08151ce-4f6b-4817-a2f4-8d7eb3102ce7-config-data" (OuterVolumeSpecName: "config-data") pod "b08151ce-4f6b-4817-a2f4-8d7eb3102ce7" (UID: "b08151ce-4f6b-4817-a2f4-8d7eb3102ce7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.599324 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b08151ce-4f6b-4817-a2f4-8d7eb3102ce7-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.649108 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.662020 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.680534 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.693858 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.701827 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 06 04:01:59 crc kubenswrapper[4802]: E1206 04:01:59.702191 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee1f735d-c438-4b44-be26-cd9b4a1b7906" containerName="nova-api-api" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.702204 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee1f735d-c438-4b44-be26-cd9b4a1b7906" containerName="nova-api-api" Dec 06 04:01:59 crc kubenswrapper[4802]: E1206 04:01:59.702218 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee1f735d-c438-4b44-be26-cd9b4a1b7906" containerName="nova-api-log" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.702225 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee1f735d-c438-4b44-be26-cd9b4a1b7906" containerName="nova-api-log" Dec 06 04:01:59 crc kubenswrapper[4802]: E1206 04:01:59.702236 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b08151ce-4f6b-4817-a2f4-8d7eb3102ce7" containerName="sg-core" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.702242 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="b08151ce-4f6b-4817-a2f4-8d7eb3102ce7" containerName="sg-core" Dec 06 04:01:59 crc kubenswrapper[4802]: E1206 04:01:59.702268 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b08151ce-4f6b-4817-a2f4-8d7eb3102ce7" containerName="ceilometer-central-agent" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.702274 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="b08151ce-4f6b-4817-a2f4-8d7eb3102ce7" containerName="ceilometer-central-agent" Dec 06 04:01:59 crc kubenswrapper[4802]: E1206 04:01:59.702286 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b08151ce-4f6b-4817-a2f4-8d7eb3102ce7" containerName="proxy-httpd" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.702291 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="b08151ce-4f6b-4817-a2f4-8d7eb3102ce7" containerName="proxy-httpd" Dec 06 04:01:59 crc kubenswrapper[4802]: E1206 04:01:59.702309 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b08151ce-4f6b-4817-a2f4-8d7eb3102ce7" containerName="ceilometer-notification-agent" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.702314 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="b08151ce-4f6b-4817-a2f4-8d7eb3102ce7" containerName="ceilometer-notification-agent" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.702471 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="b08151ce-4f6b-4817-a2f4-8d7eb3102ce7" containerName="ceilometer-notification-agent" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.702490 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee1f735d-c438-4b44-be26-cd9b4a1b7906" containerName="nova-api-log" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.702499 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="b08151ce-4f6b-4817-a2f4-8d7eb3102ce7" containerName="ceilometer-central-agent" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.702505 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="b08151ce-4f6b-4817-a2f4-8d7eb3102ce7" containerName="sg-core" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.702514 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee1f735d-c438-4b44-be26-cd9b4a1b7906" containerName="nova-api-api" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.702523 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="b08151ce-4f6b-4817-a2f4-8d7eb3102ce7" containerName="proxy-httpd" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.703444 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.719193 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.719456 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.719697 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.721672 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6242db88-a9c1-403e-8de6-d162448e7b81-config-data\") pod \"nova-api-0\" (UID: \"6242db88-a9c1-403e-8de6-d162448e7b81\") " pod="openstack/nova-api-0" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.721726 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ls7hb\" (UniqueName: \"kubernetes.io/projected/6242db88-a9c1-403e-8de6-d162448e7b81-kube-api-access-ls7hb\") pod \"nova-api-0\" (UID: \"6242db88-a9c1-403e-8de6-d162448e7b81\") " pod="openstack/nova-api-0" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.721844 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6242db88-a9c1-403e-8de6-d162448e7b81-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6242db88-a9c1-403e-8de6-d162448e7b81\") " pod="openstack/nova-api-0" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.721901 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6242db88-a9c1-403e-8de6-d162448e7b81-internal-tls-certs\") pod \"nova-api-0\" (UID: \"6242db88-a9c1-403e-8de6-d162448e7b81\") " pod="openstack/nova-api-0" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.721921 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6242db88-a9c1-403e-8de6-d162448e7b81-public-tls-certs\") pod \"nova-api-0\" (UID: \"6242db88-a9c1-403e-8de6-d162448e7b81\") " pod="openstack/nova-api-0" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.722053 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6242db88-a9c1-403e-8de6-d162448e7b81-logs\") pod \"nova-api-0\" (UID: \"6242db88-a9c1-403e-8de6-d162448e7b81\") " pod="openstack/nova-api-0" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.739472 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.759709 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.762654 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.764928 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.764993 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.765232 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.774844 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.825500 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6242db88-a9c1-403e-8de6-d162448e7b81-logs\") pod \"nova-api-0\" (UID: \"6242db88-a9c1-403e-8de6-d162448e7b81\") " pod="openstack/nova-api-0" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.825565 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6242db88-a9c1-403e-8de6-d162448e7b81-config-data\") pod \"nova-api-0\" (UID: \"6242db88-a9c1-403e-8de6-d162448e7b81\") " pod="openstack/nova-api-0" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.825597 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ls7hb\" (UniqueName: \"kubernetes.io/projected/6242db88-a9c1-403e-8de6-d162448e7b81-kube-api-access-ls7hb\") pod \"nova-api-0\" (UID: \"6242db88-a9c1-403e-8de6-d162448e7b81\") " pod="openstack/nova-api-0" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.826298 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6242db88-a9c1-403e-8de6-d162448e7b81-logs\") pod \"nova-api-0\" (UID: \"6242db88-a9c1-403e-8de6-d162448e7b81\") " pod="openstack/nova-api-0" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.826295 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6242db88-a9c1-403e-8de6-d162448e7b81-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6242db88-a9c1-403e-8de6-d162448e7b81\") " pod="openstack/nova-api-0" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.826511 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6242db88-a9c1-403e-8de6-d162448e7b81-internal-tls-certs\") pod \"nova-api-0\" (UID: \"6242db88-a9c1-403e-8de6-d162448e7b81\") " pod="openstack/nova-api-0" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.826583 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6242db88-a9c1-403e-8de6-d162448e7b81-public-tls-certs\") pod \"nova-api-0\" (UID: \"6242db88-a9c1-403e-8de6-d162448e7b81\") " pod="openstack/nova-api-0" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.832938 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6242db88-a9c1-403e-8de6-d162448e7b81-public-tls-certs\") pod \"nova-api-0\" (UID: \"6242db88-a9c1-403e-8de6-d162448e7b81\") " pod="openstack/nova-api-0" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.832943 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6242db88-a9c1-403e-8de6-d162448e7b81-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6242db88-a9c1-403e-8de6-d162448e7b81\") " pod="openstack/nova-api-0" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.833197 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6242db88-a9c1-403e-8de6-d162448e7b81-config-data\") pod \"nova-api-0\" (UID: \"6242db88-a9c1-403e-8de6-d162448e7b81\") " pod="openstack/nova-api-0" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.838318 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6242db88-a9c1-403e-8de6-d162448e7b81-internal-tls-certs\") pod \"nova-api-0\" (UID: \"6242db88-a9c1-403e-8de6-d162448e7b81\") " pod="openstack/nova-api-0" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.840290 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ls7hb\" (UniqueName: \"kubernetes.io/projected/6242db88-a9c1-403e-8de6-d162448e7b81-kube-api-access-ls7hb\") pod \"nova-api-0\" (UID: \"6242db88-a9c1-403e-8de6-d162448e7b81\") " pod="openstack/nova-api-0" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.928452 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ede1365-588f-4048-9002-034d3dbd2ef3-run-httpd\") pod \"ceilometer-0\" (UID: \"3ede1365-588f-4048-9002-034d3dbd2ef3\") " pod="openstack/ceilometer-0" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.928494 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2pfb\" (UniqueName: \"kubernetes.io/projected/3ede1365-588f-4048-9002-034d3dbd2ef3-kube-api-access-d2pfb\") pod \"ceilometer-0\" (UID: \"3ede1365-588f-4048-9002-034d3dbd2ef3\") " pod="openstack/ceilometer-0" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.928602 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ede1365-588f-4048-9002-034d3dbd2ef3-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3ede1365-588f-4048-9002-034d3dbd2ef3\") " pod="openstack/ceilometer-0" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.928641 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ede1365-588f-4048-9002-034d3dbd2ef3-scripts\") pod \"ceilometer-0\" (UID: \"3ede1365-588f-4048-9002-034d3dbd2ef3\") " pod="openstack/ceilometer-0" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.928740 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ede1365-588f-4048-9002-034d3dbd2ef3-log-httpd\") pod \"ceilometer-0\" (UID: \"3ede1365-588f-4048-9002-034d3dbd2ef3\") " pod="openstack/ceilometer-0" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.928858 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ede1365-588f-4048-9002-034d3dbd2ef3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3ede1365-588f-4048-9002-034d3dbd2ef3\") " pod="openstack/ceilometer-0" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.928903 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3ede1365-588f-4048-9002-034d3dbd2ef3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3ede1365-588f-4048-9002-034d3dbd2ef3\") " pod="openstack/ceilometer-0" Dec 06 04:01:59 crc kubenswrapper[4802]: I1206 04:01:59.928934 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ede1365-588f-4048-9002-034d3dbd2ef3-config-data\") pod \"ceilometer-0\" (UID: \"3ede1365-588f-4048-9002-034d3dbd2ef3\") " pod="openstack/ceilometer-0" Dec 06 04:02:00 crc kubenswrapper[4802]: I1206 04:02:00.031283 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ede1365-588f-4048-9002-034d3dbd2ef3-log-httpd\") pod \"ceilometer-0\" (UID: \"3ede1365-588f-4048-9002-034d3dbd2ef3\") " pod="openstack/ceilometer-0" Dec 06 04:02:00 crc kubenswrapper[4802]: I1206 04:02:00.031385 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ede1365-588f-4048-9002-034d3dbd2ef3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3ede1365-588f-4048-9002-034d3dbd2ef3\") " pod="openstack/ceilometer-0" Dec 06 04:02:00 crc kubenswrapper[4802]: I1206 04:02:00.031439 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3ede1365-588f-4048-9002-034d3dbd2ef3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3ede1365-588f-4048-9002-034d3dbd2ef3\") " pod="openstack/ceilometer-0" Dec 06 04:02:00 crc kubenswrapper[4802]: I1206 04:02:00.031472 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ede1365-588f-4048-9002-034d3dbd2ef3-config-data\") pod \"ceilometer-0\" (UID: \"3ede1365-588f-4048-9002-034d3dbd2ef3\") " pod="openstack/ceilometer-0" Dec 06 04:02:00 crc kubenswrapper[4802]: I1206 04:02:00.031572 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ede1365-588f-4048-9002-034d3dbd2ef3-run-httpd\") pod \"ceilometer-0\" (UID: \"3ede1365-588f-4048-9002-034d3dbd2ef3\") " pod="openstack/ceilometer-0" Dec 06 04:02:00 crc kubenswrapper[4802]: I1206 04:02:00.031597 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2pfb\" (UniqueName: \"kubernetes.io/projected/3ede1365-588f-4048-9002-034d3dbd2ef3-kube-api-access-d2pfb\") pod \"ceilometer-0\" (UID: \"3ede1365-588f-4048-9002-034d3dbd2ef3\") " pod="openstack/ceilometer-0" Dec 06 04:02:00 crc kubenswrapper[4802]: I1206 04:02:00.031642 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ede1365-588f-4048-9002-034d3dbd2ef3-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3ede1365-588f-4048-9002-034d3dbd2ef3\") " pod="openstack/ceilometer-0" Dec 06 04:02:00 crc kubenswrapper[4802]: I1206 04:02:00.031665 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ede1365-588f-4048-9002-034d3dbd2ef3-scripts\") pod \"ceilometer-0\" (UID: \"3ede1365-588f-4048-9002-034d3dbd2ef3\") " pod="openstack/ceilometer-0" Dec 06 04:02:00 crc kubenswrapper[4802]: I1206 04:02:00.031826 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ede1365-588f-4048-9002-034d3dbd2ef3-log-httpd\") pod \"ceilometer-0\" (UID: \"3ede1365-588f-4048-9002-034d3dbd2ef3\") " pod="openstack/ceilometer-0" Dec 06 04:02:00 crc kubenswrapper[4802]: I1206 04:02:00.032102 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ede1365-588f-4048-9002-034d3dbd2ef3-run-httpd\") pod \"ceilometer-0\" (UID: \"3ede1365-588f-4048-9002-034d3dbd2ef3\") " pod="openstack/ceilometer-0" Dec 06 04:02:00 crc kubenswrapper[4802]: I1206 04:02:00.035091 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3ede1365-588f-4048-9002-034d3dbd2ef3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3ede1365-588f-4048-9002-034d3dbd2ef3\") " pod="openstack/ceilometer-0" Dec 06 04:02:00 crc kubenswrapper[4802]: I1206 04:02:00.035321 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ede1365-588f-4048-9002-034d3dbd2ef3-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3ede1365-588f-4048-9002-034d3dbd2ef3\") " pod="openstack/ceilometer-0" Dec 06 04:02:00 crc kubenswrapper[4802]: I1206 04:02:00.036333 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ede1365-588f-4048-9002-034d3dbd2ef3-config-data\") pod \"ceilometer-0\" (UID: \"3ede1365-588f-4048-9002-034d3dbd2ef3\") " pod="openstack/ceilometer-0" Dec 06 04:02:00 crc kubenswrapper[4802]: I1206 04:02:00.036662 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ede1365-588f-4048-9002-034d3dbd2ef3-scripts\") pod \"ceilometer-0\" (UID: \"3ede1365-588f-4048-9002-034d3dbd2ef3\") " pod="openstack/ceilometer-0" Dec 06 04:02:00 crc kubenswrapper[4802]: I1206 04:02:00.037170 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ede1365-588f-4048-9002-034d3dbd2ef3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3ede1365-588f-4048-9002-034d3dbd2ef3\") " pod="openstack/ceilometer-0" Dec 06 04:02:00 crc kubenswrapper[4802]: I1206 04:02:00.053187 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2pfb\" (UniqueName: \"kubernetes.io/projected/3ede1365-588f-4048-9002-034d3dbd2ef3-kube-api-access-d2pfb\") pod \"ceilometer-0\" (UID: \"3ede1365-588f-4048-9002-034d3dbd2ef3\") " pod="openstack/ceilometer-0" Dec 06 04:02:00 crc kubenswrapper[4802]: I1206 04:02:00.094701 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 04:02:00 crc kubenswrapper[4802]: I1206 04:02:00.102298 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 04:02:00 crc kubenswrapper[4802]: I1206 04:02:00.652222 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 04:02:00 crc kubenswrapper[4802]: I1206 04:02:00.665939 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 06 04:02:01 crc kubenswrapper[4802]: I1206 04:02:01.362032 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6242db88-a9c1-403e-8de6-d162448e7b81","Type":"ContainerStarted","Data":"b6b1e5ff1fdbce8fcfff793b384cfd348effcdfc95f8a10a61794c9b801f0ddf"} Dec 06 04:02:01 crc kubenswrapper[4802]: I1206 04:02:01.362581 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6242db88-a9c1-403e-8de6-d162448e7b81","Type":"ContainerStarted","Data":"a12f5de68366599bca0adb3315e75c4a95cce952726a6d08fce6030738b0db53"} Dec 06 04:02:01 crc kubenswrapper[4802]: I1206 04:02:01.362598 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6242db88-a9c1-403e-8de6-d162448e7b81","Type":"ContainerStarted","Data":"d407be472ab08028055e68b157d2d93f5e03b816e103359379e3e42071d60b34"} Dec 06 04:02:01 crc kubenswrapper[4802]: I1206 04:02:01.367192 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ede1365-588f-4048-9002-034d3dbd2ef3","Type":"ContainerStarted","Data":"73467bcc1325d3dbdc91642bf68d16425a685467b89396f5320d9c9801e89028"} Dec 06 04:02:01 crc kubenswrapper[4802]: I1206 04:02:01.367241 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ede1365-588f-4048-9002-034d3dbd2ef3","Type":"ContainerStarted","Data":"8f5e2251dbf24a8c8822b8ea2d9d09ad85b8f6350e2db03b10f8e07f8ea6d9fd"} Dec 06 04:02:01 crc kubenswrapper[4802]: I1206 04:02:01.382973 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.382954133 podStartE2EDuration="2.382954133s" podCreationTimestamp="2025-12-06 04:01:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:02:01.381676787 +0000 UTC m=+1314.253585939" watchObservedRunningTime="2025-12-06 04:02:01.382954133 +0000 UTC m=+1314.254863285" Dec 06 04:02:01 crc kubenswrapper[4802]: I1206 04:02:01.462993 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b08151ce-4f6b-4817-a2f4-8d7eb3102ce7" path="/var/lib/kubelet/pods/b08151ce-4f6b-4817-a2f4-8d7eb3102ce7/volumes" Dec 06 04:02:01 crc kubenswrapper[4802]: I1206 04:02:01.464221 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee1f735d-c438-4b44-be26-cd9b4a1b7906" path="/var/lib/kubelet/pods/ee1f735d-c438-4b44-be26-cd9b4a1b7906/volumes" Dec 06 04:02:01 crc kubenswrapper[4802]: I1206 04:02:01.597892 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 06 04:02:01 crc kubenswrapper[4802]: I1206 04:02:01.627790 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 06 04:02:02 crc kubenswrapper[4802]: I1206 04:02:02.384030 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ede1365-588f-4048-9002-034d3dbd2ef3","Type":"ContainerStarted","Data":"4fe9e6d39cbc7c61f0f9ca67e36005858564999f41954fa0ad1dbad42b1cc95b"} Dec 06 04:02:02 crc kubenswrapper[4802]: I1206 04:02:02.407474 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 06 04:02:02 crc kubenswrapper[4802]: I1206 04:02:02.622363 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-l4lf9"] Dec 06 04:02:02 crc kubenswrapper[4802]: I1206 04:02:02.623902 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-l4lf9" Dec 06 04:02:02 crc kubenswrapper[4802]: I1206 04:02:02.628305 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 06 04:02:02 crc kubenswrapper[4802]: I1206 04:02:02.628390 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 06 04:02:02 crc kubenswrapper[4802]: I1206 04:02:02.648140 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-l4lf9"] Dec 06 04:02:02 crc kubenswrapper[4802]: I1206 04:02:02.692330 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db51698a-4eea-44d9-bbe1-9dab86cd7a02-config-data\") pod \"nova-cell1-cell-mapping-l4lf9\" (UID: \"db51698a-4eea-44d9-bbe1-9dab86cd7a02\") " pod="openstack/nova-cell1-cell-mapping-l4lf9" Dec 06 04:02:02 crc kubenswrapper[4802]: I1206 04:02:02.692403 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db51698a-4eea-44d9-bbe1-9dab86cd7a02-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-l4lf9\" (UID: \"db51698a-4eea-44d9-bbe1-9dab86cd7a02\") " pod="openstack/nova-cell1-cell-mapping-l4lf9" Dec 06 04:02:02 crc kubenswrapper[4802]: I1206 04:02:02.692430 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4sfgt\" (UniqueName: \"kubernetes.io/projected/db51698a-4eea-44d9-bbe1-9dab86cd7a02-kube-api-access-4sfgt\") pod \"nova-cell1-cell-mapping-l4lf9\" (UID: \"db51698a-4eea-44d9-bbe1-9dab86cd7a02\") " pod="openstack/nova-cell1-cell-mapping-l4lf9" Dec 06 04:02:02 crc kubenswrapper[4802]: I1206 04:02:02.692529 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db51698a-4eea-44d9-bbe1-9dab86cd7a02-scripts\") pod \"nova-cell1-cell-mapping-l4lf9\" (UID: \"db51698a-4eea-44d9-bbe1-9dab86cd7a02\") " pod="openstack/nova-cell1-cell-mapping-l4lf9" Dec 06 04:02:02 crc kubenswrapper[4802]: I1206 04:02:02.719965 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-f84f9ccf-fc42d" Dec 06 04:02:02 crc kubenswrapper[4802]: I1206 04:02:02.794060 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db51698a-4eea-44d9-bbe1-9dab86cd7a02-config-data\") pod \"nova-cell1-cell-mapping-l4lf9\" (UID: \"db51698a-4eea-44d9-bbe1-9dab86cd7a02\") " pod="openstack/nova-cell1-cell-mapping-l4lf9" Dec 06 04:02:02 crc kubenswrapper[4802]: I1206 04:02:02.794119 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db51698a-4eea-44d9-bbe1-9dab86cd7a02-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-l4lf9\" (UID: \"db51698a-4eea-44d9-bbe1-9dab86cd7a02\") " pod="openstack/nova-cell1-cell-mapping-l4lf9" Dec 06 04:02:02 crc kubenswrapper[4802]: I1206 04:02:02.794140 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4sfgt\" (UniqueName: \"kubernetes.io/projected/db51698a-4eea-44d9-bbe1-9dab86cd7a02-kube-api-access-4sfgt\") pod \"nova-cell1-cell-mapping-l4lf9\" (UID: \"db51698a-4eea-44d9-bbe1-9dab86cd7a02\") " pod="openstack/nova-cell1-cell-mapping-l4lf9" Dec 06 04:02:02 crc kubenswrapper[4802]: I1206 04:02:02.794225 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db51698a-4eea-44d9-bbe1-9dab86cd7a02-scripts\") pod \"nova-cell1-cell-mapping-l4lf9\" (UID: \"db51698a-4eea-44d9-bbe1-9dab86cd7a02\") " pod="openstack/nova-cell1-cell-mapping-l4lf9" Dec 06 04:02:02 crc kubenswrapper[4802]: I1206 04:02:02.819943 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4sfgt\" (UniqueName: \"kubernetes.io/projected/db51698a-4eea-44d9-bbe1-9dab86cd7a02-kube-api-access-4sfgt\") pod \"nova-cell1-cell-mapping-l4lf9\" (UID: \"db51698a-4eea-44d9-bbe1-9dab86cd7a02\") " pod="openstack/nova-cell1-cell-mapping-l4lf9" Dec 06 04:02:02 crc kubenswrapper[4802]: I1206 04:02:02.828389 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db51698a-4eea-44d9-bbe1-9dab86cd7a02-scripts\") pod \"nova-cell1-cell-mapping-l4lf9\" (UID: \"db51698a-4eea-44d9-bbe1-9dab86cd7a02\") " pod="openstack/nova-cell1-cell-mapping-l4lf9" Dec 06 04:02:02 crc kubenswrapper[4802]: I1206 04:02:02.828732 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db51698a-4eea-44d9-bbe1-9dab86cd7a02-config-data\") pod \"nova-cell1-cell-mapping-l4lf9\" (UID: \"db51698a-4eea-44d9-bbe1-9dab86cd7a02\") " pod="openstack/nova-cell1-cell-mapping-l4lf9" Dec 06 04:02:02 crc kubenswrapper[4802]: I1206 04:02:02.836328 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-568d7fd7cf-246lw"] Dec 06 04:02:02 crc kubenswrapper[4802]: I1206 04:02:02.836594 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-568d7fd7cf-246lw" podUID="dd12e103-8229-466d-b8e2-5bf5eaa140bc" containerName="dnsmasq-dns" containerID="cri-o://55175fe1098f34cbf3f20d2c9e4fa9d2a30be8eba4ff9596438e223f98d62e83" gracePeriod=10 Dec 06 04:02:02 crc kubenswrapper[4802]: I1206 04:02:02.838451 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db51698a-4eea-44d9-bbe1-9dab86cd7a02-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-l4lf9\" (UID: \"db51698a-4eea-44d9-bbe1-9dab86cd7a02\") " pod="openstack/nova-cell1-cell-mapping-l4lf9" Dec 06 04:02:02 crc kubenswrapper[4802]: I1206 04:02:02.947148 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-l4lf9" Dec 06 04:02:03 crc kubenswrapper[4802]: I1206 04:02:03.309055 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-568d7fd7cf-246lw" Dec 06 04:02:03 crc kubenswrapper[4802]: I1206 04:02:03.392949 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ede1365-588f-4048-9002-034d3dbd2ef3","Type":"ContainerStarted","Data":"041555ee702efd03224ff3c77968272015be88981da6484df556a8225f6f3bfc"} Dec 06 04:02:03 crc kubenswrapper[4802]: I1206 04:02:03.394722 4802 generic.go:334] "Generic (PLEG): container finished" podID="dd12e103-8229-466d-b8e2-5bf5eaa140bc" containerID="55175fe1098f34cbf3f20d2c9e4fa9d2a30be8eba4ff9596438e223f98d62e83" exitCode=0 Dec 06 04:02:03 crc kubenswrapper[4802]: I1206 04:02:03.394960 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-568d7fd7cf-246lw" Dec 06 04:02:03 crc kubenswrapper[4802]: I1206 04:02:03.395556 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-568d7fd7cf-246lw" event={"ID":"dd12e103-8229-466d-b8e2-5bf5eaa140bc","Type":"ContainerDied","Data":"55175fe1098f34cbf3f20d2c9e4fa9d2a30be8eba4ff9596438e223f98d62e83"} Dec 06 04:02:03 crc kubenswrapper[4802]: I1206 04:02:03.395582 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-568d7fd7cf-246lw" event={"ID":"dd12e103-8229-466d-b8e2-5bf5eaa140bc","Type":"ContainerDied","Data":"6b9478bd0f2ad56fbb108028b8f264ba782adc418a18046814f3d26cff9b412b"} Dec 06 04:02:03 crc kubenswrapper[4802]: I1206 04:02:03.395599 4802 scope.go:117] "RemoveContainer" containerID="55175fe1098f34cbf3f20d2c9e4fa9d2a30be8eba4ff9596438e223f98d62e83" Dec 06 04:02:03 crc kubenswrapper[4802]: I1206 04:02:03.419566 4802 scope.go:117] "RemoveContainer" containerID="1c03aba4c50865b2bfc9d14891dbc7d7401c22bbdf3ca315d2b896283c9c70a0" Dec 06 04:02:03 crc kubenswrapper[4802]: I1206 04:02:03.446494 4802 scope.go:117] "RemoveContainer" containerID="55175fe1098f34cbf3f20d2c9e4fa9d2a30be8eba4ff9596438e223f98d62e83" Dec 06 04:02:03 crc kubenswrapper[4802]: E1206 04:02:03.446977 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55175fe1098f34cbf3f20d2c9e4fa9d2a30be8eba4ff9596438e223f98d62e83\": container with ID starting with 55175fe1098f34cbf3f20d2c9e4fa9d2a30be8eba4ff9596438e223f98d62e83 not found: ID does not exist" containerID="55175fe1098f34cbf3f20d2c9e4fa9d2a30be8eba4ff9596438e223f98d62e83" Dec 06 04:02:03 crc kubenswrapper[4802]: I1206 04:02:03.447015 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55175fe1098f34cbf3f20d2c9e4fa9d2a30be8eba4ff9596438e223f98d62e83"} err="failed to get container status \"55175fe1098f34cbf3f20d2c9e4fa9d2a30be8eba4ff9596438e223f98d62e83\": rpc error: code = NotFound desc = could not find container \"55175fe1098f34cbf3f20d2c9e4fa9d2a30be8eba4ff9596438e223f98d62e83\": container with ID starting with 55175fe1098f34cbf3f20d2c9e4fa9d2a30be8eba4ff9596438e223f98d62e83 not found: ID does not exist" Dec 06 04:02:03 crc kubenswrapper[4802]: I1206 04:02:03.447041 4802 scope.go:117] "RemoveContainer" containerID="1c03aba4c50865b2bfc9d14891dbc7d7401c22bbdf3ca315d2b896283c9c70a0" Dec 06 04:02:03 crc kubenswrapper[4802]: E1206 04:02:03.447499 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c03aba4c50865b2bfc9d14891dbc7d7401c22bbdf3ca315d2b896283c9c70a0\": container with ID starting with 1c03aba4c50865b2bfc9d14891dbc7d7401c22bbdf3ca315d2b896283c9c70a0 not found: ID does not exist" containerID="1c03aba4c50865b2bfc9d14891dbc7d7401c22bbdf3ca315d2b896283c9c70a0" Dec 06 04:02:03 crc kubenswrapper[4802]: I1206 04:02:03.447529 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c03aba4c50865b2bfc9d14891dbc7d7401c22bbdf3ca315d2b896283c9c70a0"} err="failed to get container status \"1c03aba4c50865b2bfc9d14891dbc7d7401c22bbdf3ca315d2b896283c9c70a0\": rpc error: code = NotFound desc = could not find container \"1c03aba4c50865b2bfc9d14891dbc7d7401c22bbdf3ca315d2b896283c9c70a0\": container with ID starting with 1c03aba4c50865b2bfc9d14891dbc7d7401c22bbdf3ca315d2b896283c9c70a0 not found: ID does not exist" Dec 06 04:02:03 crc kubenswrapper[4802]: I1206 04:02:03.501012 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-l4lf9"] Dec 06 04:02:03 crc kubenswrapper[4802]: W1206 04:02:03.502633 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddb51698a_4eea_44d9_bbe1_9dab86cd7a02.slice/crio-ac785aa1fc7da0140aef5cef2df58330a1b9ab9803b69faef062618c0a021c60 WatchSource:0}: Error finding container ac785aa1fc7da0140aef5cef2df58330a1b9ab9803b69faef062618c0a021c60: Status 404 returned error can't find the container with id ac785aa1fc7da0140aef5cef2df58330a1b9ab9803b69faef062618c0a021c60 Dec 06 04:02:03 crc kubenswrapper[4802]: I1206 04:02:03.509870 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dd12e103-8229-466d-b8e2-5bf5eaa140bc-ovsdbserver-sb\") pod \"dd12e103-8229-466d-b8e2-5bf5eaa140bc\" (UID: \"dd12e103-8229-466d-b8e2-5bf5eaa140bc\") " Dec 06 04:02:03 crc kubenswrapper[4802]: I1206 04:02:03.509916 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dd12e103-8229-466d-b8e2-5bf5eaa140bc-ovsdbserver-nb\") pod \"dd12e103-8229-466d-b8e2-5bf5eaa140bc\" (UID: \"dd12e103-8229-466d-b8e2-5bf5eaa140bc\") " Dec 06 04:02:03 crc kubenswrapper[4802]: I1206 04:02:03.509969 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dd12e103-8229-466d-b8e2-5bf5eaa140bc-dns-svc\") pod \"dd12e103-8229-466d-b8e2-5bf5eaa140bc\" (UID: \"dd12e103-8229-466d-b8e2-5bf5eaa140bc\") " Dec 06 04:02:03 crc kubenswrapper[4802]: I1206 04:02:03.510048 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wb66l\" (UniqueName: \"kubernetes.io/projected/dd12e103-8229-466d-b8e2-5bf5eaa140bc-kube-api-access-wb66l\") pod \"dd12e103-8229-466d-b8e2-5bf5eaa140bc\" (UID: \"dd12e103-8229-466d-b8e2-5bf5eaa140bc\") " Dec 06 04:02:03 crc kubenswrapper[4802]: I1206 04:02:03.510109 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd12e103-8229-466d-b8e2-5bf5eaa140bc-config\") pod \"dd12e103-8229-466d-b8e2-5bf5eaa140bc\" (UID: \"dd12e103-8229-466d-b8e2-5bf5eaa140bc\") " Dec 06 04:02:03 crc kubenswrapper[4802]: I1206 04:02:03.510212 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dd12e103-8229-466d-b8e2-5bf5eaa140bc-dns-swift-storage-0\") pod \"dd12e103-8229-466d-b8e2-5bf5eaa140bc\" (UID: \"dd12e103-8229-466d-b8e2-5bf5eaa140bc\") " Dec 06 04:02:03 crc kubenswrapper[4802]: I1206 04:02:03.515621 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd12e103-8229-466d-b8e2-5bf5eaa140bc-kube-api-access-wb66l" (OuterVolumeSpecName: "kube-api-access-wb66l") pod "dd12e103-8229-466d-b8e2-5bf5eaa140bc" (UID: "dd12e103-8229-466d-b8e2-5bf5eaa140bc"). InnerVolumeSpecName "kube-api-access-wb66l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:02:03 crc kubenswrapper[4802]: I1206 04:02:03.565353 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd12e103-8229-466d-b8e2-5bf5eaa140bc-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "dd12e103-8229-466d-b8e2-5bf5eaa140bc" (UID: "dd12e103-8229-466d-b8e2-5bf5eaa140bc"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:02:03 crc kubenswrapper[4802]: I1206 04:02:03.570133 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd12e103-8229-466d-b8e2-5bf5eaa140bc-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "dd12e103-8229-466d-b8e2-5bf5eaa140bc" (UID: "dd12e103-8229-466d-b8e2-5bf5eaa140bc"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:02:03 crc kubenswrapper[4802]: I1206 04:02:03.578793 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd12e103-8229-466d-b8e2-5bf5eaa140bc-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "dd12e103-8229-466d-b8e2-5bf5eaa140bc" (UID: "dd12e103-8229-466d-b8e2-5bf5eaa140bc"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:02:03 crc kubenswrapper[4802]: I1206 04:02:03.582240 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd12e103-8229-466d-b8e2-5bf5eaa140bc-config" (OuterVolumeSpecName: "config") pod "dd12e103-8229-466d-b8e2-5bf5eaa140bc" (UID: "dd12e103-8229-466d-b8e2-5bf5eaa140bc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:02:03 crc kubenswrapper[4802]: I1206 04:02:03.586524 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd12e103-8229-466d-b8e2-5bf5eaa140bc-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "dd12e103-8229-466d-b8e2-5bf5eaa140bc" (UID: "dd12e103-8229-466d-b8e2-5bf5eaa140bc"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:02:03 crc kubenswrapper[4802]: I1206 04:02:03.612072 4802 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dd12e103-8229-466d-b8e2-5bf5eaa140bc-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 04:02:03 crc kubenswrapper[4802]: I1206 04:02:03.612104 4802 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dd12e103-8229-466d-b8e2-5bf5eaa140bc-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 04:02:03 crc kubenswrapper[4802]: I1206 04:02:03.612115 4802 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dd12e103-8229-466d-b8e2-5bf5eaa140bc-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 04:02:03 crc kubenswrapper[4802]: I1206 04:02:03.612127 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wb66l\" (UniqueName: \"kubernetes.io/projected/dd12e103-8229-466d-b8e2-5bf5eaa140bc-kube-api-access-wb66l\") on node \"crc\" DevicePath \"\"" Dec 06 04:02:03 crc kubenswrapper[4802]: I1206 04:02:03.612136 4802 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd12e103-8229-466d-b8e2-5bf5eaa140bc-config\") on node \"crc\" DevicePath \"\"" Dec 06 04:02:03 crc kubenswrapper[4802]: I1206 04:02:03.612145 4802 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dd12e103-8229-466d-b8e2-5bf5eaa140bc-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 04:02:03 crc kubenswrapper[4802]: I1206 04:02:03.753186 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-568d7fd7cf-246lw"] Dec 06 04:02:03 crc kubenswrapper[4802]: I1206 04:02:03.763750 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-568d7fd7cf-246lw"] Dec 06 04:02:04 crc kubenswrapper[4802]: I1206 04:02:04.407569 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-l4lf9" event={"ID":"db51698a-4eea-44d9-bbe1-9dab86cd7a02","Type":"ContainerStarted","Data":"93295a23f9c3d49cfdc9948bfe05cc2796671712333ba33a3f9f42614887987b"} Dec 06 04:02:04 crc kubenswrapper[4802]: I1206 04:02:04.407837 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-l4lf9" event={"ID":"db51698a-4eea-44d9-bbe1-9dab86cd7a02","Type":"ContainerStarted","Data":"ac785aa1fc7da0140aef5cef2df58330a1b9ab9803b69faef062618c0a021c60"} Dec 06 04:02:04 crc kubenswrapper[4802]: I1206 04:02:04.412189 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ede1365-588f-4048-9002-034d3dbd2ef3","Type":"ContainerStarted","Data":"9adfd82b4aebb989dfd7e521fa4ceee9ae4fe88d73db2a14783d64678195dc58"} Dec 06 04:02:04 crc kubenswrapper[4802]: I1206 04:02:04.432440 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-l4lf9" podStartSLOduration=2.432419786 podStartE2EDuration="2.432419786s" podCreationTimestamp="2025-12-06 04:02:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:02:04.431311456 +0000 UTC m=+1317.303220618" watchObservedRunningTime="2025-12-06 04:02:04.432419786 +0000 UTC m=+1317.304328938" Dec 06 04:02:05 crc kubenswrapper[4802]: I1206 04:02:05.420150 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 06 04:02:05 crc kubenswrapper[4802]: I1206 04:02:05.449734 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.117224906 podStartE2EDuration="6.449716367s" podCreationTimestamp="2025-12-06 04:01:59 +0000 UTC" firstStartedPulling="2025-12-06 04:02:00.648611099 +0000 UTC m=+1313.520520251" lastFinishedPulling="2025-12-06 04:02:03.98110256 +0000 UTC m=+1316.853011712" observedRunningTime="2025-12-06 04:02:05.439230229 +0000 UTC m=+1318.311139381" watchObservedRunningTime="2025-12-06 04:02:05.449716367 +0000 UTC m=+1318.321625519" Dec 06 04:02:05 crc kubenswrapper[4802]: I1206 04:02:05.460085 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd12e103-8229-466d-b8e2-5bf5eaa140bc" path="/var/lib/kubelet/pods/dd12e103-8229-466d-b8e2-5bf5eaa140bc/volumes" Dec 06 04:02:09 crc kubenswrapper[4802]: I1206 04:02:09.473813 4802 generic.go:334] "Generic (PLEG): container finished" podID="db51698a-4eea-44d9-bbe1-9dab86cd7a02" containerID="93295a23f9c3d49cfdc9948bfe05cc2796671712333ba33a3f9f42614887987b" exitCode=0 Dec 06 04:02:09 crc kubenswrapper[4802]: I1206 04:02:09.473876 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-l4lf9" event={"ID":"db51698a-4eea-44d9-bbe1-9dab86cd7a02","Type":"ContainerDied","Data":"93295a23f9c3d49cfdc9948bfe05cc2796671712333ba33a3f9f42614887987b"} Dec 06 04:02:10 crc kubenswrapper[4802]: I1206 04:02:10.095495 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 06 04:02:10 crc kubenswrapper[4802]: I1206 04:02:10.095554 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 06 04:02:10 crc kubenswrapper[4802]: I1206 04:02:10.907811 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-l4lf9" Dec 06 04:02:11 crc kubenswrapper[4802]: I1206 04:02:11.079816 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4sfgt\" (UniqueName: \"kubernetes.io/projected/db51698a-4eea-44d9-bbe1-9dab86cd7a02-kube-api-access-4sfgt\") pod \"db51698a-4eea-44d9-bbe1-9dab86cd7a02\" (UID: \"db51698a-4eea-44d9-bbe1-9dab86cd7a02\") " Dec 06 04:02:11 crc kubenswrapper[4802]: I1206 04:02:11.079875 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db51698a-4eea-44d9-bbe1-9dab86cd7a02-scripts\") pod \"db51698a-4eea-44d9-bbe1-9dab86cd7a02\" (UID: \"db51698a-4eea-44d9-bbe1-9dab86cd7a02\") " Dec 06 04:02:11 crc kubenswrapper[4802]: I1206 04:02:11.079952 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db51698a-4eea-44d9-bbe1-9dab86cd7a02-combined-ca-bundle\") pod \"db51698a-4eea-44d9-bbe1-9dab86cd7a02\" (UID: \"db51698a-4eea-44d9-bbe1-9dab86cd7a02\") " Dec 06 04:02:11 crc kubenswrapper[4802]: I1206 04:02:11.080045 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db51698a-4eea-44d9-bbe1-9dab86cd7a02-config-data\") pod \"db51698a-4eea-44d9-bbe1-9dab86cd7a02\" (UID: \"db51698a-4eea-44d9-bbe1-9dab86cd7a02\") " Dec 06 04:02:11 crc kubenswrapper[4802]: I1206 04:02:11.087964 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db51698a-4eea-44d9-bbe1-9dab86cd7a02-kube-api-access-4sfgt" (OuterVolumeSpecName: "kube-api-access-4sfgt") pod "db51698a-4eea-44d9-bbe1-9dab86cd7a02" (UID: "db51698a-4eea-44d9-bbe1-9dab86cd7a02"). InnerVolumeSpecName "kube-api-access-4sfgt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:02:11 crc kubenswrapper[4802]: I1206 04:02:11.094530 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db51698a-4eea-44d9-bbe1-9dab86cd7a02-scripts" (OuterVolumeSpecName: "scripts") pod "db51698a-4eea-44d9-bbe1-9dab86cd7a02" (UID: "db51698a-4eea-44d9-bbe1-9dab86cd7a02"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:02:11 crc kubenswrapper[4802]: I1206 04:02:11.106979 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="6242db88-a9c1-403e-8de6-d162448e7b81" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.208:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 04:02:11 crc kubenswrapper[4802]: I1206 04:02:11.107391 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="6242db88-a9c1-403e-8de6-d162448e7b81" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.208:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 04:02:11 crc kubenswrapper[4802]: I1206 04:02:11.123328 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db51698a-4eea-44d9-bbe1-9dab86cd7a02-config-data" (OuterVolumeSpecName: "config-data") pod "db51698a-4eea-44d9-bbe1-9dab86cd7a02" (UID: "db51698a-4eea-44d9-bbe1-9dab86cd7a02"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:02:11 crc kubenswrapper[4802]: I1206 04:02:11.125149 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db51698a-4eea-44d9-bbe1-9dab86cd7a02-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "db51698a-4eea-44d9-bbe1-9dab86cd7a02" (UID: "db51698a-4eea-44d9-bbe1-9dab86cd7a02"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:02:11 crc kubenswrapper[4802]: I1206 04:02:11.184454 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4sfgt\" (UniqueName: \"kubernetes.io/projected/db51698a-4eea-44d9-bbe1-9dab86cd7a02-kube-api-access-4sfgt\") on node \"crc\" DevicePath \"\"" Dec 06 04:02:11 crc kubenswrapper[4802]: I1206 04:02:11.184488 4802 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db51698a-4eea-44d9-bbe1-9dab86cd7a02-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 04:02:11 crc kubenswrapper[4802]: I1206 04:02:11.184501 4802 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db51698a-4eea-44d9-bbe1-9dab86cd7a02-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:02:11 crc kubenswrapper[4802]: I1206 04:02:11.184514 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db51698a-4eea-44d9-bbe1-9dab86cd7a02-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 04:02:11 crc kubenswrapper[4802]: I1206 04:02:11.493072 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-l4lf9" event={"ID":"db51698a-4eea-44d9-bbe1-9dab86cd7a02","Type":"ContainerDied","Data":"ac785aa1fc7da0140aef5cef2df58330a1b9ab9803b69faef062618c0a021c60"} Dec 06 04:02:11 crc kubenswrapper[4802]: I1206 04:02:11.493109 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ac785aa1fc7da0140aef5cef2df58330a1b9ab9803b69faef062618c0a021c60" Dec 06 04:02:11 crc kubenswrapper[4802]: I1206 04:02:11.493127 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-l4lf9" Dec 06 04:02:11 crc kubenswrapper[4802]: I1206 04:02:11.689309 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 06 04:02:11 crc kubenswrapper[4802]: I1206 04:02:11.690332 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="6242db88-a9c1-403e-8de6-d162448e7b81" containerName="nova-api-log" containerID="cri-o://a12f5de68366599bca0adb3315e75c4a95cce952726a6d08fce6030738b0db53" gracePeriod=30 Dec 06 04:02:11 crc kubenswrapper[4802]: I1206 04:02:11.690861 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="6242db88-a9c1-403e-8de6-d162448e7b81" containerName="nova-api-api" containerID="cri-o://b6b1e5ff1fdbce8fcfff793b384cfd348effcdfc95f8a10a61794c9b801f0ddf" gracePeriod=30 Dec 06 04:02:11 crc kubenswrapper[4802]: I1206 04:02:11.732229 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 04:02:11 crc kubenswrapper[4802]: I1206 04:02:11.732412 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="cb0d4e0d-9954-49c8-b05a-118fb157c87e" containerName="nova-scheduler-scheduler" containerID="cri-o://74685c7dbe3d20aba42d93effcc6a18d5f191b2974cf99959fe1b09051c22626" gracePeriod=30 Dec 06 04:02:11 crc kubenswrapper[4802]: I1206 04:02:11.788661 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 04:02:11 crc kubenswrapper[4802]: I1206 04:02:11.788958 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be" containerName="nova-metadata-log" containerID="cri-o://d10f8eeaa0106a33b52fd1a34f7ab3007212756b42626cc4ae7d14a99433a0aa" gracePeriod=30 Dec 06 04:02:11 crc kubenswrapper[4802]: I1206 04:02:11.789325 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be" containerName="nova-metadata-metadata" containerID="cri-o://f39aa5495b856420e48e53837f48c1f867617f09aaa212ea86ad61f7409b64b6" gracePeriod=30 Dec 06 04:02:12 crc kubenswrapper[4802]: I1206 04:02:12.502953 4802 generic.go:334] "Generic (PLEG): container finished" podID="7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be" containerID="d10f8eeaa0106a33b52fd1a34f7ab3007212756b42626cc4ae7d14a99433a0aa" exitCode=143 Dec 06 04:02:12 crc kubenswrapper[4802]: I1206 04:02:12.503007 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be","Type":"ContainerDied","Data":"d10f8eeaa0106a33b52fd1a34f7ab3007212756b42626cc4ae7d14a99433a0aa"} Dec 06 04:02:12 crc kubenswrapper[4802]: I1206 04:02:12.506519 4802 generic.go:334] "Generic (PLEG): container finished" podID="6242db88-a9c1-403e-8de6-d162448e7b81" containerID="a12f5de68366599bca0adb3315e75c4a95cce952726a6d08fce6030738b0db53" exitCode=143 Dec 06 04:02:12 crc kubenswrapper[4802]: I1206 04:02:12.506571 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6242db88-a9c1-403e-8de6-d162448e7b81","Type":"ContainerDied","Data":"a12f5de68366599bca0adb3315e75c4a95cce952726a6d08fce6030738b0db53"} Dec 06 04:02:13 crc kubenswrapper[4802]: I1206 04:02:13.283112 4802 patch_prober.go:28] interesting pod/machine-config-daemon-zpxxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:02:13 crc kubenswrapper[4802]: I1206 04:02:13.283215 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:02:15 crc kubenswrapper[4802]: I1206 04:02:15.543057 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 04:02:15 crc kubenswrapper[4802]: I1206 04:02:15.545442 4802 generic.go:334] "Generic (PLEG): container finished" podID="7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be" containerID="f39aa5495b856420e48e53837f48c1f867617f09aaa212ea86ad61f7409b64b6" exitCode=0 Dec 06 04:02:15 crc kubenswrapper[4802]: I1206 04:02:15.545509 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be","Type":"ContainerDied","Data":"f39aa5495b856420e48e53837f48c1f867617f09aaa212ea86ad61f7409b64b6"} Dec 06 04:02:15 crc kubenswrapper[4802]: I1206 04:02:15.545554 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be","Type":"ContainerDied","Data":"1afe2766859689dc8280a7c0d1c65f783dd603bb4800bbea8e85aaab909a0789"} Dec 06 04:02:15 crc kubenswrapper[4802]: I1206 04:02:15.545570 4802 scope.go:117] "RemoveContainer" containerID="f39aa5495b856420e48e53837f48c1f867617f09aaa212ea86ad61f7409b64b6" Dec 06 04:02:15 crc kubenswrapper[4802]: I1206 04:02:15.548793 4802 generic.go:334] "Generic (PLEG): container finished" podID="cb0d4e0d-9954-49c8-b05a-118fb157c87e" containerID="74685c7dbe3d20aba42d93effcc6a18d5f191b2974cf99959fe1b09051c22626" exitCode=0 Dec 06 04:02:15 crc kubenswrapper[4802]: I1206 04:02:15.548822 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cb0d4e0d-9954-49c8-b05a-118fb157c87e","Type":"ContainerDied","Data":"74685c7dbe3d20aba42d93effcc6a18d5f191b2974cf99959fe1b09051c22626"} Dec 06 04:02:15 crc kubenswrapper[4802]: I1206 04:02:15.626510 4802 scope.go:117] "RemoveContainer" containerID="d10f8eeaa0106a33b52fd1a34f7ab3007212756b42626cc4ae7d14a99433a0aa" Dec 06 04:02:15 crc kubenswrapper[4802]: I1206 04:02:15.658802 4802 scope.go:117] "RemoveContainer" containerID="f39aa5495b856420e48e53837f48c1f867617f09aaa212ea86ad61f7409b64b6" Dec 06 04:02:15 crc kubenswrapper[4802]: E1206 04:02:15.659322 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f39aa5495b856420e48e53837f48c1f867617f09aaa212ea86ad61f7409b64b6\": container with ID starting with f39aa5495b856420e48e53837f48c1f867617f09aaa212ea86ad61f7409b64b6 not found: ID does not exist" containerID="f39aa5495b856420e48e53837f48c1f867617f09aaa212ea86ad61f7409b64b6" Dec 06 04:02:15 crc kubenswrapper[4802]: I1206 04:02:15.659349 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f39aa5495b856420e48e53837f48c1f867617f09aaa212ea86ad61f7409b64b6"} err="failed to get container status \"f39aa5495b856420e48e53837f48c1f867617f09aaa212ea86ad61f7409b64b6\": rpc error: code = NotFound desc = could not find container \"f39aa5495b856420e48e53837f48c1f867617f09aaa212ea86ad61f7409b64b6\": container with ID starting with f39aa5495b856420e48e53837f48c1f867617f09aaa212ea86ad61f7409b64b6 not found: ID does not exist" Dec 06 04:02:15 crc kubenswrapper[4802]: I1206 04:02:15.659370 4802 scope.go:117] "RemoveContainer" containerID="d10f8eeaa0106a33b52fd1a34f7ab3007212756b42626cc4ae7d14a99433a0aa" Dec 06 04:02:15 crc kubenswrapper[4802]: E1206 04:02:15.659707 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d10f8eeaa0106a33b52fd1a34f7ab3007212756b42626cc4ae7d14a99433a0aa\": container with ID starting with d10f8eeaa0106a33b52fd1a34f7ab3007212756b42626cc4ae7d14a99433a0aa not found: ID does not exist" containerID="d10f8eeaa0106a33b52fd1a34f7ab3007212756b42626cc4ae7d14a99433a0aa" Dec 06 04:02:15 crc kubenswrapper[4802]: I1206 04:02:15.659793 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d10f8eeaa0106a33b52fd1a34f7ab3007212756b42626cc4ae7d14a99433a0aa"} err="failed to get container status \"d10f8eeaa0106a33b52fd1a34f7ab3007212756b42626cc4ae7d14a99433a0aa\": rpc error: code = NotFound desc = could not find container \"d10f8eeaa0106a33b52fd1a34f7ab3007212756b42626cc4ae7d14a99433a0aa\": container with ID starting with d10f8eeaa0106a33b52fd1a34f7ab3007212756b42626cc4ae7d14a99433a0aa not found: ID does not exist" Dec 06 04:02:15 crc kubenswrapper[4802]: I1206 04:02:15.664077 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be-logs\") pod \"7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be\" (UID: \"7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be\") " Dec 06 04:02:15 crc kubenswrapper[4802]: I1206 04:02:15.664197 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be-combined-ca-bundle\") pod \"7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be\" (UID: \"7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be\") " Dec 06 04:02:15 crc kubenswrapper[4802]: I1206 04:02:15.664235 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2dcqg\" (UniqueName: \"kubernetes.io/projected/7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be-kube-api-access-2dcqg\") pod \"7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be\" (UID: \"7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be\") " Dec 06 04:02:15 crc kubenswrapper[4802]: I1206 04:02:15.664288 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be-nova-metadata-tls-certs\") pod \"7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be\" (UID: \"7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be\") " Dec 06 04:02:15 crc kubenswrapper[4802]: I1206 04:02:15.664340 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be-config-data\") pod \"7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be\" (UID: \"7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be\") " Dec 06 04:02:15 crc kubenswrapper[4802]: I1206 04:02:15.665578 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be-logs" (OuterVolumeSpecName: "logs") pod "7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be" (UID: "7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:02:15 crc kubenswrapper[4802]: I1206 04:02:15.672286 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be-kube-api-access-2dcqg" (OuterVolumeSpecName: "kube-api-access-2dcqg") pod "7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be" (UID: "7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be"). InnerVolumeSpecName "kube-api-access-2dcqg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:02:15 crc kubenswrapper[4802]: I1206 04:02:15.698630 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be" (UID: "7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:02:15 crc kubenswrapper[4802]: I1206 04:02:15.716209 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be-config-data" (OuterVolumeSpecName: "config-data") pod "7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be" (UID: "7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:02:15 crc kubenswrapper[4802]: I1206 04:02:15.738828 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be" (UID: "7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:02:15 crc kubenswrapper[4802]: I1206 04:02:15.767320 4802 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:02:15 crc kubenswrapper[4802]: I1206 04:02:15.767345 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2dcqg\" (UniqueName: \"kubernetes.io/projected/7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be-kube-api-access-2dcqg\") on node \"crc\" DevicePath \"\"" Dec 06 04:02:15 crc kubenswrapper[4802]: I1206 04:02:15.767356 4802 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 04:02:15 crc kubenswrapper[4802]: I1206 04:02:15.767365 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 04:02:15 crc kubenswrapper[4802]: I1206 04:02:15.767398 4802 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be-logs\") on node \"crc\" DevicePath \"\"" Dec 06 04:02:15 crc kubenswrapper[4802]: I1206 04:02:15.856331 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 04:02:15 crc kubenswrapper[4802]: I1206 04:02:15.969234 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb0d4e0d-9954-49c8-b05a-118fb157c87e-config-data\") pod \"cb0d4e0d-9954-49c8-b05a-118fb157c87e\" (UID: \"cb0d4e0d-9954-49c8-b05a-118fb157c87e\") " Dec 06 04:02:15 crc kubenswrapper[4802]: I1206 04:02:15.969322 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zck9m\" (UniqueName: \"kubernetes.io/projected/cb0d4e0d-9954-49c8-b05a-118fb157c87e-kube-api-access-zck9m\") pod \"cb0d4e0d-9954-49c8-b05a-118fb157c87e\" (UID: \"cb0d4e0d-9954-49c8-b05a-118fb157c87e\") " Dec 06 04:02:15 crc kubenswrapper[4802]: I1206 04:02:15.969401 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb0d4e0d-9954-49c8-b05a-118fb157c87e-combined-ca-bundle\") pod \"cb0d4e0d-9954-49c8-b05a-118fb157c87e\" (UID: \"cb0d4e0d-9954-49c8-b05a-118fb157c87e\") " Dec 06 04:02:15 crc kubenswrapper[4802]: I1206 04:02:15.973819 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb0d4e0d-9954-49c8-b05a-118fb157c87e-kube-api-access-zck9m" (OuterVolumeSpecName: "kube-api-access-zck9m") pod "cb0d4e0d-9954-49c8-b05a-118fb157c87e" (UID: "cb0d4e0d-9954-49c8-b05a-118fb157c87e"). InnerVolumeSpecName "kube-api-access-zck9m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.004642 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb0d4e0d-9954-49c8-b05a-118fb157c87e-config-data" (OuterVolumeSpecName: "config-data") pod "cb0d4e0d-9954-49c8-b05a-118fb157c87e" (UID: "cb0d4e0d-9954-49c8-b05a-118fb157c87e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.005613 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb0d4e0d-9954-49c8-b05a-118fb157c87e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cb0d4e0d-9954-49c8-b05a-118fb157c87e" (UID: "cb0d4e0d-9954-49c8-b05a-118fb157c87e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.071121 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zck9m\" (UniqueName: \"kubernetes.io/projected/cb0d4e0d-9954-49c8-b05a-118fb157c87e-kube-api-access-zck9m\") on node \"crc\" DevicePath \"\"" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.071162 4802 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb0d4e0d-9954-49c8-b05a-118fb157c87e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.071177 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb0d4e0d-9954-49c8-b05a-118fb157c87e-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.560791 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.562236 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cb0d4e0d-9954-49c8-b05a-118fb157c87e","Type":"ContainerDied","Data":"774f7ead2ef7c54ec80dedaaa235e905afc165b55be17fefce1a4ffc790bea8c"} Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.562314 4802 scope.go:117] "RemoveContainer" containerID="74685c7dbe3d20aba42d93effcc6a18d5f191b2974cf99959fe1b09051c22626" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.562734 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.565581 4802 generic.go:334] "Generic (PLEG): container finished" podID="6242db88-a9c1-403e-8de6-d162448e7b81" containerID="b6b1e5ff1fdbce8fcfff793b384cfd348effcdfc95f8a10a61794c9b801f0ddf" exitCode=0 Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.565670 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6242db88-a9c1-403e-8de6-d162448e7b81","Type":"ContainerDied","Data":"b6b1e5ff1fdbce8fcfff793b384cfd348effcdfc95f8a10a61794c9b801f0ddf"} Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.565714 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6242db88-a9c1-403e-8de6-d162448e7b81","Type":"ContainerDied","Data":"d407be472ab08028055e68b157d2d93f5e03b816e103359379e3e42071d60b34"} Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.565737 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d407be472ab08028055e68b157d2d93f5e03b816e103359379e3e42071d60b34" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.639517 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.688888 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.736568 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.745964 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.753807 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.762054 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 06 04:02:16 crc kubenswrapper[4802]: E1206 04:02:16.762583 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db51698a-4eea-44d9-bbe1-9dab86cd7a02" containerName="nova-manage" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.762611 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="db51698a-4eea-44d9-bbe1-9dab86cd7a02" containerName="nova-manage" Dec 06 04:02:16 crc kubenswrapper[4802]: E1206 04:02:16.762629 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6242db88-a9c1-403e-8de6-d162448e7b81" containerName="nova-api-api" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.762639 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="6242db88-a9c1-403e-8de6-d162448e7b81" containerName="nova-api-api" Dec 06 04:02:16 crc kubenswrapper[4802]: E1206 04:02:16.762658 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6242db88-a9c1-403e-8de6-d162448e7b81" containerName="nova-api-log" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.762666 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="6242db88-a9c1-403e-8de6-d162448e7b81" containerName="nova-api-log" Dec 06 04:02:16 crc kubenswrapper[4802]: E1206 04:02:16.762687 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd12e103-8229-466d-b8e2-5bf5eaa140bc" containerName="init" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.762695 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd12e103-8229-466d-b8e2-5bf5eaa140bc" containerName="init" Dec 06 04:02:16 crc kubenswrapper[4802]: E1206 04:02:16.762710 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be" containerName="nova-metadata-metadata" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.762718 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be" containerName="nova-metadata-metadata" Dec 06 04:02:16 crc kubenswrapper[4802]: E1206 04:02:16.762735 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb0d4e0d-9954-49c8-b05a-118fb157c87e" containerName="nova-scheduler-scheduler" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.762743 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb0d4e0d-9954-49c8-b05a-118fb157c87e" containerName="nova-scheduler-scheduler" Dec 06 04:02:16 crc kubenswrapper[4802]: E1206 04:02:16.762778 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd12e103-8229-466d-b8e2-5bf5eaa140bc" containerName="dnsmasq-dns" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.762787 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd12e103-8229-466d-b8e2-5bf5eaa140bc" containerName="dnsmasq-dns" Dec 06 04:02:16 crc kubenswrapper[4802]: E1206 04:02:16.762799 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be" containerName="nova-metadata-log" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.762808 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be" containerName="nova-metadata-log" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.763047 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb0d4e0d-9954-49c8-b05a-118fb157c87e" containerName="nova-scheduler-scheduler" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.763064 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd12e103-8229-466d-b8e2-5bf5eaa140bc" containerName="dnsmasq-dns" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.763073 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="6242db88-a9c1-403e-8de6-d162448e7b81" containerName="nova-api-api" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.763089 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="db51698a-4eea-44d9-bbe1-9dab86cd7a02" containerName="nova-manage" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.763109 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be" containerName="nova-metadata-metadata" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.763126 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="6242db88-a9c1-403e-8de6-d162448e7b81" containerName="nova-api-log" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.763140 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be" containerName="nova-metadata-log" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.764490 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.767554 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.767979 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.774481 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.784234 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.786086 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.786710 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6242db88-a9c1-403e-8de6-d162448e7b81-logs\") pod \"6242db88-a9c1-403e-8de6-d162448e7b81\" (UID: \"6242db88-a9c1-403e-8de6-d162448e7b81\") " Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.786881 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6242db88-a9c1-403e-8de6-d162448e7b81-internal-tls-certs\") pod \"6242db88-a9c1-403e-8de6-d162448e7b81\" (UID: \"6242db88-a9c1-403e-8de6-d162448e7b81\") " Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.787005 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ls7hb\" (UniqueName: \"kubernetes.io/projected/6242db88-a9c1-403e-8de6-d162448e7b81-kube-api-access-ls7hb\") pod \"6242db88-a9c1-403e-8de6-d162448e7b81\" (UID: \"6242db88-a9c1-403e-8de6-d162448e7b81\") " Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.788262 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.788577 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6242db88-a9c1-403e-8de6-d162448e7b81-logs" (OuterVolumeSpecName: "logs") pod "6242db88-a9c1-403e-8de6-d162448e7b81" (UID: "6242db88-a9c1-403e-8de6-d162448e7b81"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.790698 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6242db88-a9c1-403e-8de6-d162448e7b81-config-data\") pod \"6242db88-a9c1-403e-8de6-d162448e7b81\" (UID: \"6242db88-a9c1-403e-8de6-d162448e7b81\") " Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.790876 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6242db88-a9c1-403e-8de6-d162448e7b81-combined-ca-bundle\") pod \"6242db88-a9c1-403e-8de6-d162448e7b81\" (UID: \"6242db88-a9c1-403e-8de6-d162448e7b81\") " Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.791086 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6242db88-a9c1-403e-8de6-d162448e7b81-public-tls-certs\") pod \"6242db88-a9c1-403e-8de6-d162448e7b81\" (UID: \"6242db88-a9c1-403e-8de6-d162448e7b81\") " Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.791436 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1350889-6def-43ce-9827-761796bea156-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f1350889-6def-43ce-9827-761796bea156\") " pod="openstack/nova-metadata-0" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.792053 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvckz\" (UniqueName: \"kubernetes.io/projected/f1350889-6def-43ce-9827-761796bea156-kube-api-access-fvckz\") pod \"nova-metadata-0\" (UID: \"f1350889-6def-43ce-9827-761796bea156\") " pod="openstack/nova-metadata-0" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.792214 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1350889-6def-43ce-9827-761796bea156-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f1350889-6def-43ce-9827-761796bea156\") " pod="openstack/nova-metadata-0" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.792320 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjlsq\" (UniqueName: \"kubernetes.io/projected/723826fa-70b3-4be3-9380-42561ca2041b-kube-api-access-mjlsq\") pod \"nova-scheduler-0\" (UID: \"723826fa-70b3-4be3-9380-42561ca2041b\") " pod="openstack/nova-scheduler-0" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.792522 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/723826fa-70b3-4be3-9380-42561ca2041b-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"723826fa-70b3-4be3-9380-42561ca2041b\") " pod="openstack/nova-scheduler-0" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.792678 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1350889-6def-43ce-9827-761796bea156-logs\") pod \"nova-metadata-0\" (UID: \"f1350889-6def-43ce-9827-761796bea156\") " pod="openstack/nova-metadata-0" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.792906 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1350889-6def-43ce-9827-761796bea156-config-data\") pod \"nova-metadata-0\" (UID: \"f1350889-6def-43ce-9827-761796bea156\") " pod="openstack/nova-metadata-0" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.793042 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/723826fa-70b3-4be3-9380-42561ca2041b-config-data\") pod \"nova-scheduler-0\" (UID: \"723826fa-70b3-4be3-9380-42561ca2041b\") " pod="openstack/nova-scheduler-0" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.793684 4802 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6242db88-a9c1-403e-8de6-d162448e7b81-logs\") on node \"crc\" DevicePath \"\"" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.794720 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6242db88-a9c1-403e-8de6-d162448e7b81-kube-api-access-ls7hb" (OuterVolumeSpecName: "kube-api-access-ls7hb") pod "6242db88-a9c1-403e-8de6-d162448e7b81" (UID: "6242db88-a9c1-403e-8de6-d162448e7b81"). InnerVolumeSpecName "kube-api-access-ls7hb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.823017 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6242db88-a9c1-403e-8de6-d162448e7b81-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6242db88-a9c1-403e-8de6-d162448e7b81" (UID: "6242db88-a9c1-403e-8de6-d162448e7b81"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.841681 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6242db88-a9c1-403e-8de6-d162448e7b81-config-data" (OuterVolumeSpecName: "config-data") pod "6242db88-a9c1-403e-8de6-d162448e7b81" (UID: "6242db88-a9c1-403e-8de6-d162448e7b81"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.847264 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6242db88-a9c1-403e-8de6-d162448e7b81-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "6242db88-a9c1-403e-8de6-d162448e7b81" (UID: "6242db88-a9c1-403e-8de6-d162448e7b81"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.851248 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.865963 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6242db88-a9c1-403e-8de6-d162448e7b81-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "6242db88-a9c1-403e-8de6-d162448e7b81" (UID: "6242db88-a9c1-403e-8de6-d162448e7b81"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.896188 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1350889-6def-43ce-9827-761796bea156-logs\") pod \"nova-metadata-0\" (UID: \"f1350889-6def-43ce-9827-761796bea156\") " pod="openstack/nova-metadata-0" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.896267 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1350889-6def-43ce-9827-761796bea156-config-data\") pod \"nova-metadata-0\" (UID: \"f1350889-6def-43ce-9827-761796bea156\") " pod="openstack/nova-metadata-0" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.896289 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/723826fa-70b3-4be3-9380-42561ca2041b-config-data\") pod \"nova-scheduler-0\" (UID: \"723826fa-70b3-4be3-9380-42561ca2041b\") " pod="openstack/nova-scheduler-0" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.896337 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1350889-6def-43ce-9827-761796bea156-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f1350889-6def-43ce-9827-761796bea156\") " pod="openstack/nova-metadata-0" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.896371 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvckz\" (UniqueName: \"kubernetes.io/projected/f1350889-6def-43ce-9827-761796bea156-kube-api-access-fvckz\") pod \"nova-metadata-0\" (UID: \"f1350889-6def-43ce-9827-761796bea156\") " pod="openstack/nova-metadata-0" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.896400 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1350889-6def-43ce-9827-761796bea156-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f1350889-6def-43ce-9827-761796bea156\") " pod="openstack/nova-metadata-0" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.896417 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjlsq\" (UniqueName: \"kubernetes.io/projected/723826fa-70b3-4be3-9380-42561ca2041b-kube-api-access-mjlsq\") pod \"nova-scheduler-0\" (UID: \"723826fa-70b3-4be3-9380-42561ca2041b\") " pod="openstack/nova-scheduler-0" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.896463 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/723826fa-70b3-4be3-9380-42561ca2041b-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"723826fa-70b3-4be3-9380-42561ca2041b\") " pod="openstack/nova-scheduler-0" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.896569 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1350889-6def-43ce-9827-761796bea156-logs\") pod \"nova-metadata-0\" (UID: \"f1350889-6def-43ce-9827-761796bea156\") " pod="openstack/nova-metadata-0" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.900602 4802 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6242db88-a9c1-403e-8de6-d162448e7b81-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.901505 4802 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6242db88-a9c1-403e-8de6-d162448e7b81-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.901520 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ls7hb\" (UniqueName: \"kubernetes.io/projected/6242db88-a9c1-403e-8de6-d162448e7b81-kube-api-access-ls7hb\") on node \"crc\" DevicePath \"\"" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.901537 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6242db88-a9c1-403e-8de6-d162448e7b81-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.901725 4802 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6242db88-a9c1-403e-8de6-d162448e7b81-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.905813 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1350889-6def-43ce-9827-761796bea156-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f1350889-6def-43ce-9827-761796bea156\") " pod="openstack/nova-metadata-0" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.906168 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1350889-6def-43ce-9827-761796bea156-config-data\") pod \"nova-metadata-0\" (UID: \"f1350889-6def-43ce-9827-761796bea156\") " pod="openstack/nova-metadata-0" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.914135 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/723826fa-70b3-4be3-9380-42561ca2041b-config-data\") pod \"nova-scheduler-0\" (UID: \"723826fa-70b3-4be3-9380-42561ca2041b\") " pod="openstack/nova-scheduler-0" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.915087 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/723826fa-70b3-4be3-9380-42561ca2041b-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"723826fa-70b3-4be3-9380-42561ca2041b\") " pod="openstack/nova-scheduler-0" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.915809 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1350889-6def-43ce-9827-761796bea156-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f1350889-6def-43ce-9827-761796bea156\") " pod="openstack/nova-metadata-0" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.916099 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjlsq\" (UniqueName: \"kubernetes.io/projected/723826fa-70b3-4be3-9380-42561ca2041b-kube-api-access-mjlsq\") pod \"nova-scheduler-0\" (UID: \"723826fa-70b3-4be3-9380-42561ca2041b\") " pod="openstack/nova-scheduler-0" Dec 06 04:02:16 crc kubenswrapper[4802]: I1206 04:02:16.918828 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvckz\" (UniqueName: \"kubernetes.io/projected/f1350889-6def-43ce-9827-761796bea156-kube-api-access-fvckz\") pod \"nova-metadata-0\" (UID: \"f1350889-6def-43ce-9827-761796bea156\") " pod="openstack/nova-metadata-0" Dec 06 04:02:17 crc kubenswrapper[4802]: I1206 04:02:17.084325 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 04:02:17 crc kubenswrapper[4802]: I1206 04:02:17.189128 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 04:02:17 crc kubenswrapper[4802]: I1206 04:02:17.462865 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be" path="/var/lib/kubelet/pods/7655cfe4-be90-4fd1-8bdc-e9bd6aa0c2be/volumes" Dec 06 04:02:17 crc kubenswrapper[4802]: I1206 04:02:17.463641 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb0d4e0d-9954-49c8-b05a-118fb157c87e" path="/var/lib/kubelet/pods/cb0d4e0d-9954-49c8-b05a-118fb157c87e/volumes" Dec 06 04:02:17 crc kubenswrapper[4802]: I1206 04:02:17.580198 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 04:02:17 crc kubenswrapper[4802]: I1206 04:02:17.601007 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 06 04:02:17 crc kubenswrapper[4802]: I1206 04:02:17.612957 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 06 04:02:17 crc kubenswrapper[4802]: I1206 04:02:17.620925 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 06 04:02:17 crc kubenswrapper[4802]: I1206 04:02:17.622400 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 04:02:17 crc kubenswrapper[4802]: I1206 04:02:17.627499 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 06 04:02:17 crc kubenswrapper[4802]: I1206 04:02:17.627507 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 06 04:02:17 crc kubenswrapper[4802]: I1206 04:02:17.627719 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 06 04:02:17 crc kubenswrapper[4802]: I1206 04:02:17.641024 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 06 04:02:17 crc kubenswrapper[4802]: I1206 04:02:17.714895 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 04:02:17 crc kubenswrapper[4802]: W1206 04:02:17.715908 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1350889_6def_43ce_9827_761796bea156.slice/crio-8bc01bd81f0ae15c8181a56d561d702d955b5d6ef5b83d9afa9129df82d95c99 WatchSource:0}: Error finding container 8bc01bd81f0ae15c8181a56d561d702d955b5d6ef5b83d9afa9129df82d95c99: Status 404 returned error can't find the container with id 8bc01bd81f0ae15c8181a56d561d702d955b5d6ef5b83d9afa9129df82d95c99 Dec 06 04:02:17 crc kubenswrapper[4802]: I1206 04:02:17.716042 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/add9c7ee-92a7-4cd8-adbf-02ec103cb6dc-internal-tls-certs\") pod \"nova-api-0\" (UID: \"add9c7ee-92a7-4cd8-adbf-02ec103cb6dc\") " pod="openstack/nova-api-0" Dec 06 04:02:17 crc kubenswrapper[4802]: I1206 04:02:17.716126 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9nt2g\" (UniqueName: \"kubernetes.io/projected/add9c7ee-92a7-4cd8-adbf-02ec103cb6dc-kube-api-access-9nt2g\") pod \"nova-api-0\" (UID: \"add9c7ee-92a7-4cd8-adbf-02ec103cb6dc\") " pod="openstack/nova-api-0" Dec 06 04:02:17 crc kubenswrapper[4802]: I1206 04:02:17.716222 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/add9c7ee-92a7-4cd8-adbf-02ec103cb6dc-public-tls-certs\") pod \"nova-api-0\" (UID: \"add9c7ee-92a7-4cd8-adbf-02ec103cb6dc\") " pod="openstack/nova-api-0" Dec 06 04:02:17 crc kubenswrapper[4802]: I1206 04:02:17.716371 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/add9c7ee-92a7-4cd8-adbf-02ec103cb6dc-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"add9c7ee-92a7-4cd8-adbf-02ec103cb6dc\") " pod="openstack/nova-api-0" Dec 06 04:02:17 crc kubenswrapper[4802]: I1206 04:02:17.716436 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/add9c7ee-92a7-4cd8-adbf-02ec103cb6dc-config-data\") pod \"nova-api-0\" (UID: \"add9c7ee-92a7-4cd8-adbf-02ec103cb6dc\") " pod="openstack/nova-api-0" Dec 06 04:02:17 crc kubenswrapper[4802]: I1206 04:02:17.716543 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/add9c7ee-92a7-4cd8-adbf-02ec103cb6dc-logs\") pod \"nova-api-0\" (UID: \"add9c7ee-92a7-4cd8-adbf-02ec103cb6dc\") " pod="openstack/nova-api-0" Dec 06 04:02:17 crc kubenswrapper[4802]: I1206 04:02:17.836938 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/add9c7ee-92a7-4cd8-adbf-02ec103cb6dc-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"add9c7ee-92a7-4cd8-adbf-02ec103cb6dc\") " pod="openstack/nova-api-0" Dec 06 04:02:17 crc kubenswrapper[4802]: I1206 04:02:17.836986 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/add9c7ee-92a7-4cd8-adbf-02ec103cb6dc-config-data\") pod \"nova-api-0\" (UID: \"add9c7ee-92a7-4cd8-adbf-02ec103cb6dc\") " pod="openstack/nova-api-0" Dec 06 04:02:17 crc kubenswrapper[4802]: I1206 04:02:17.837036 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/add9c7ee-92a7-4cd8-adbf-02ec103cb6dc-logs\") pod \"nova-api-0\" (UID: \"add9c7ee-92a7-4cd8-adbf-02ec103cb6dc\") " pod="openstack/nova-api-0" Dec 06 04:02:17 crc kubenswrapper[4802]: I1206 04:02:17.837132 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/add9c7ee-92a7-4cd8-adbf-02ec103cb6dc-internal-tls-certs\") pod \"nova-api-0\" (UID: \"add9c7ee-92a7-4cd8-adbf-02ec103cb6dc\") " pod="openstack/nova-api-0" Dec 06 04:02:17 crc kubenswrapper[4802]: I1206 04:02:17.837185 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9nt2g\" (UniqueName: \"kubernetes.io/projected/add9c7ee-92a7-4cd8-adbf-02ec103cb6dc-kube-api-access-9nt2g\") pod \"nova-api-0\" (UID: \"add9c7ee-92a7-4cd8-adbf-02ec103cb6dc\") " pod="openstack/nova-api-0" Dec 06 04:02:17 crc kubenswrapper[4802]: I1206 04:02:17.837233 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/add9c7ee-92a7-4cd8-adbf-02ec103cb6dc-public-tls-certs\") pod \"nova-api-0\" (UID: \"add9c7ee-92a7-4cd8-adbf-02ec103cb6dc\") " pod="openstack/nova-api-0" Dec 06 04:02:17 crc kubenswrapper[4802]: I1206 04:02:17.837601 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/add9c7ee-92a7-4cd8-adbf-02ec103cb6dc-logs\") pod \"nova-api-0\" (UID: \"add9c7ee-92a7-4cd8-adbf-02ec103cb6dc\") " pod="openstack/nova-api-0" Dec 06 04:02:17 crc kubenswrapper[4802]: I1206 04:02:17.842829 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/add9c7ee-92a7-4cd8-adbf-02ec103cb6dc-config-data\") pod \"nova-api-0\" (UID: \"add9c7ee-92a7-4cd8-adbf-02ec103cb6dc\") " pod="openstack/nova-api-0" Dec 06 04:02:17 crc kubenswrapper[4802]: I1206 04:02:17.843032 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/add9c7ee-92a7-4cd8-adbf-02ec103cb6dc-internal-tls-certs\") pod \"nova-api-0\" (UID: \"add9c7ee-92a7-4cd8-adbf-02ec103cb6dc\") " pod="openstack/nova-api-0" Dec 06 04:02:17 crc kubenswrapper[4802]: I1206 04:02:17.845544 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/add9c7ee-92a7-4cd8-adbf-02ec103cb6dc-public-tls-certs\") pod \"nova-api-0\" (UID: \"add9c7ee-92a7-4cd8-adbf-02ec103cb6dc\") " pod="openstack/nova-api-0" Dec 06 04:02:17 crc kubenswrapper[4802]: I1206 04:02:17.846267 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/add9c7ee-92a7-4cd8-adbf-02ec103cb6dc-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"add9c7ee-92a7-4cd8-adbf-02ec103cb6dc\") " pod="openstack/nova-api-0" Dec 06 04:02:17 crc kubenswrapper[4802]: I1206 04:02:17.855294 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9nt2g\" (UniqueName: \"kubernetes.io/projected/add9c7ee-92a7-4cd8-adbf-02ec103cb6dc-kube-api-access-9nt2g\") pod \"nova-api-0\" (UID: \"add9c7ee-92a7-4cd8-adbf-02ec103cb6dc\") " pod="openstack/nova-api-0" Dec 06 04:02:17 crc kubenswrapper[4802]: I1206 04:02:17.880218 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 04:02:17 crc kubenswrapper[4802]: W1206 04:02:17.884362 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod723826fa_70b3_4be3_9380_42561ca2041b.slice/crio-e4b72a4aad1e949ca07bf8f7f338f97d0d0ae9d9bee69605e644fbb6ddd6eab7 WatchSource:0}: Error finding container e4b72a4aad1e949ca07bf8f7f338f97d0d0ae9d9bee69605e644fbb6ddd6eab7: Status 404 returned error can't find the container with id e4b72a4aad1e949ca07bf8f7f338f97d0d0ae9d9bee69605e644fbb6ddd6eab7 Dec 06 04:02:17 crc kubenswrapper[4802]: I1206 04:02:17.942361 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 04:02:18 crc kubenswrapper[4802]: I1206 04:02:18.425639 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 06 04:02:18 crc kubenswrapper[4802]: W1206 04:02:18.432599 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podadd9c7ee_92a7_4cd8_adbf_02ec103cb6dc.slice/crio-a856d4ca1cea51bd3213d1f012bae1a102e8736e2af3a6cf8bf29f9b47fe3ec1 WatchSource:0}: Error finding container a856d4ca1cea51bd3213d1f012bae1a102e8736e2af3a6cf8bf29f9b47fe3ec1: Status 404 returned error can't find the container with id a856d4ca1cea51bd3213d1f012bae1a102e8736e2af3a6cf8bf29f9b47fe3ec1 Dec 06 04:02:18 crc kubenswrapper[4802]: I1206 04:02:18.611736 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"723826fa-70b3-4be3-9380-42561ca2041b","Type":"ContainerStarted","Data":"18d3fa13943a4e87ee2210009e5f7eb062b4111aeea53af6337bcf08c0616f98"} Dec 06 04:02:18 crc kubenswrapper[4802]: I1206 04:02:18.612166 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"723826fa-70b3-4be3-9380-42561ca2041b","Type":"ContainerStarted","Data":"e4b72a4aad1e949ca07bf8f7f338f97d0d0ae9d9bee69605e644fbb6ddd6eab7"} Dec 06 04:02:18 crc kubenswrapper[4802]: I1206 04:02:18.614888 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"add9c7ee-92a7-4cd8-adbf-02ec103cb6dc","Type":"ContainerStarted","Data":"a856d4ca1cea51bd3213d1f012bae1a102e8736e2af3a6cf8bf29f9b47fe3ec1"} Dec 06 04:02:18 crc kubenswrapper[4802]: I1206 04:02:18.624306 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f1350889-6def-43ce-9827-761796bea156","Type":"ContainerStarted","Data":"c08c68018aca805619c4884818b36bf2d911bef532306be4d644de2a3015c5b1"} Dec 06 04:02:18 crc kubenswrapper[4802]: I1206 04:02:18.624392 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f1350889-6def-43ce-9827-761796bea156","Type":"ContainerStarted","Data":"fcce16b76eed8d003878ebd70a83f2f67a2ecf34ca13382f47e0829fe2c9ad10"} Dec 06 04:02:18 crc kubenswrapper[4802]: I1206 04:02:18.624408 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f1350889-6def-43ce-9827-761796bea156","Type":"ContainerStarted","Data":"8bc01bd81f0ae15c8181a56d561d702d955b5d6ef5b83d9afa9129df82d95c99"} Dec 06 04:02:18 crc kubenswrapper[4802]: I1206 04:02:18.639656 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.6396382210000002 podStartE2EDuration="2.639638221s" podCreationTimestamp="2025-12-06 04:02:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:02:18.63009331 +0000 UTC m=+1331.502002462" watchObservedRunningTime="2025-12-06 04:02:18.639638221 +0000 UTC m=+1331.511547383" Dec 06 04:02:18 crc kubenswrapper[4802]: I1206 04:02:18.687579 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.687555252 podStartE2EDuration="2.687555252s" podCreationTimestamp="2025-12-06 04:02:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:02:18.665211081 +0000 UTC m=+1331.537120243" watchObservedRunningTime="2025-12-06 04:02:18.687555252 +0000 UTC m=+1331.559464404" Dec 06 04:02:19 crc kubenswrapper[4802]: I1206 04:02:19.466111 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6242db88-a9c1-403e-8de6-d162448e7b81" path="/var/lib/kubelet/pods/6242db88-a9c1-403e-8de6-d162448e7b81/volumes" Dec 06 04:02:19 crc kubenswrapper[4802]: I1206 04:02:19.646608 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"add9c7ee-92a7-4cd8-adbf-02ec103cb6dc","Type":"ContainerStarted","Data":"4a6154a02a7a826d3b1763dba6756f9b60c9cfddf10c4e3eb80a39a66e1c8ef2"} Dec 06 04:02:19 crc kubenswrapper[4802]: I1206 04:02:19.646676 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"add9c7ee-92a7-4cd8-adbf-02ec103cb6dc","Type":"ContainerStarted","Data":"f182f7a30e6b01497eab1a55731dc76da09be037734f7f021b3fecd9e9c4c48b"} Dec 06 04:02:19 crc kubenswrapper[4802]: I1206 04:02:19.684497 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.684479755 podStartE2EDuration="2.684479755s" podCreationTimestamp="2025-12-06 04:02:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:02:19.680868436 +0000 UTC m=+1332.552777628" watchObservedRunningTime="2025-12-06 04:02:19.684479755 +0000 UTC m=+1332.556388917" Dec 06 04:02:22 crc kubenswrapper[4802]: I1206 04:02:22.085343 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 06 04:02:22 crc kubenswrapper[4802]: I1206 04:02:22.085847 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 06 04:02:22 crc kubenswrapper[4802]: I1206 04:02:22.190348 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 06 04:02:27 crc kubenswrapper[4802]: I1206 04:02:27.085999 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 06 04:02:27 crc kubenswrapper[4802]: I1206 04:02:27.086602 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 06 04:02:27 crc kubenswrapper[4802]: I1206 04:02:27.190249 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 06 04:02:27 crc kubenswrapper[4802]: I1206 04:02:27.220488 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 06 04:02:27 crc kubenswrapper[4802]: I1206 04:02:27.752042 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 06 04:02:27 crc kubenswrapper[4802]: I1206 04:02:27.943497 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 06 04:02:27 crc kubenswrapper[4802]: I1206 04:02:27.943541 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 06 04:02:28 crc kubenswrapper[4802]: I1206 04:02:28.104949 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="f1350889-6def-43ce-9827-761796bea156" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.211:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 04:02:28 crc kubenswrapper[4802]: I1206 04:02:28.104958 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="f1350889-6def-43ce-9827-761796bea156" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.211:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 04:02:28 crc kubenswrapper[4802]: I1206 04:02:28.956003 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="add9c7ee-92a7-4cd8-adbf-02ec103cb6dc" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.213:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 04:02:28 crc kubenswrapper[4802]: I1206 04:02:28.956077 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="add9c7ee-92a7-4cd8-adbf-02ec103cb6dc" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.213:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 04:02:30 crc kubenswrapper[4802]: I1206 04:02:30.112047 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 06 04:02:37 crc kubenswrapper[4802]: I1206 04:02:37.092143 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 06 04:02:37 crc kubenswrapper[4802]: I1206 04:02:37.092878 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 06 04:02:37 crc kubenswrapper[4802]: I1206 04:02:37.099768 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 06 04:02:37 crc kubenswrapper[4802]: I1206 04:02:37.099833 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 06 04:02:37 crc kubenswrapper[4802]: I1206 04:02:37.948686 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 06 04:02:37 crc kubenswrapper[4802]: I1206 04:02:37.949940 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 06 04:02:37 crc kubenswrapper[4802]: I1206 04:02:37.950773 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 06 04:02:37 crc kubenswrapper[4802]: I1206 04:02:37.959074 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 06 04:02:38 crc kubenswrapper[4802]: I1206 04:02:38.839760 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 06 04:02:38 crc kubenswrapper[4802]: I1206 04:02:38.846687 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 06 04:02:43 crc kubenswrapper[4802]: I1206 04:02:43.283913 4802 patch_prober.go:28] interesting pod/machine-config-daemon-zpxxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:02:43 crc kubenswrapper[4802]: I1206 04:02:43.284448 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:02:43 crc kubenswrapper[4802]: I1206 04:02:43.284495 4802 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" Dec 06 04:02:43 crc kubenswrapper[4802]: I1206 04:02:43.285263 4802 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"61a067de7e4bf972bb44f2b14a9f2f9bdc755648ec9a47e6260216803462a3b1"} pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 04:02:43 crc kubenswrapper[4802]: I1206 04:02:43.285317 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" containerID="cri-o://61a067de7e4bf972bb44f2b14a9f2f9bdc755648ec9a47e6260216803462a3b1" gracePeriod=600 Dec 06 04:02:43 crc kubenswrapper[4802]: I1206 04:02:43.892355 4802 generic.go:334] "Generic (PLEG): container finished" podID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerID="61a067de7e4bf972bb44f2b14a9f2f9bdc755648ec9a47e6260216803462a3b1" exitCode=0 Dec 06 04:02:43 crc kubenswrapper[4802]: I1206 04:02:43.892417 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" event={"ID":"6df38316-e0d3-4018-8d27-3620eba3a68d","Type":"ContainerDied","Data":"61a067de7e4bf972bb44f2b14a9f2f9bdc755648ec9a47e6260216803462a3b1"} Dec 06 04:02:43 crc kubenswrapper[4802]: I1206 04:02:43.892784 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" event={"ID":"6df38316-e0d3-4018-8d27-3620eba3a68d","Type":"ContainerStarted","Data":"3711adecb81191bc35ac469efa73faeb2346d30f4d5a4a148de8ac88071cdb00"} Dec 06 04:02:43 crc kubenswrapper[4802]: I1206 04:02:43.892809 4802 scope.go:117] "RemoveContainer" containerID="04d0056d567a84a69f17a8c41df89172171e22968697aae49e1b7d82534da79c" Dec 06 04:02:47 crc kubenswrapper[4802]: I1206 04:02:47.439161 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 04:02:48 crc kubenswrapper[4802]: I1206 04:02:48.394369 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 04:02:52 crc kubenswrapper[4802]: I1206 04:02:52.036483 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="7b4f22a1-41dc-4a75-8999-53df29d2fb5d" containerName="rabbitmq" containerID="cri-o://ed17f5b81a54486ceb0680c299b5f84abc02bd1e687d265819e37dc8b4570d6b" gracePeriod=604796 Dec 06 04:02:52 crc kubenswrapper[4802]: I1206 04:02:52.391456 4802 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="7b4f22a1-41dc-4a75-8999-53df29d2fb5d" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.98:5671: connect: connection refused" Dec 06 04:02:52 crc kubenswrapper[4802]: I1206 04:02:52.582940 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="8c8b6c49-6f71-403d-b639-4a17ae3bc802" containerName="rabbitmq" containerID="cri-o://098aed31247101a27a8ff5c8dd2bfd3731dd532b7504c6cf5431f78994869852" gracePeriod=604796 Dec 06 04:02:58 crc kubenswrapper[4802]: I1206 04:02:58.889970 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.023503 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-pod-info\") pod \"7b4f22a1-41dc-4a75-8999-53df29d2fb5d\" (UID: \"7b4f22a1-41dc-4a75-8999-53df29d2fb5d\") " Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.023567 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-rabbitmq-confd\") pod \"7b4f22a1-41dc-4a75-8999-53df29d2fb5d\" (UID: \"7b4f22a1-41dc-4a75-8999-53df29d2fb5d\") " Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.023634 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-config-data\") pod \"7b4f22a1-41dc-4a75-8999-53df29d2fb5d\" (UID: \"7b4f22a1-41dc-4a75-8999-53df29d2fb5d\") " Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.023705 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"7b4f22a1-41dc-4a75-8999-53df29d2fb5d\" (UID: \"7b4f22a1-41dc-4a75-8999-53df29d2fb5d\") " Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.023782 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cjz5x\" (UniqueName: \"kubernetes.io/projected/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-kube-api-access-cjz5x\") pod \"7b4f22a1-41dc-4a75-8999-53df29d2fb5d\" (UID: \"7b4f22a1-41dc-4a75-8999-53df29d2fb5d\") " Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.024629 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-rabbitmq-tls\") pod \"7b4f22a1-41dc-4a75-8999-53df29d2fb5d\" (UID: \"7b4f22a1-41dc-4a75-8999-53df29d2fb5d\") " Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.024675 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-rabbitmq-erlang-cookie\") pod \"7b4f22a1-41dc-4a75-8999-53df29d2fb5d\" (UID: \"7b4f22a1-41dc-4a75-8999-53df29d2fb5d\") " Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.024701 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-erlang-cookie-secret\") pod \"7b4f22a1-41dc-4a75-8999-53df29d2fb5d\" (UID: \"7b4f22a1-41dc-4a75-8999-53df29d2fb5d\") " Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.024804 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-plugins-conf\") pod \"7b4f22a1-41dc-4a75-8999-53df29d2fb5d\" (UID: \"7b4f22a1-41dc-4a75-8999-53df29d2fb5d\") " Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.024834 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-server-conf\") pod \"7b4f22a1-41dc-4a75-8999-53df29d2fb5d\" (UID: \"7b4f22a1-41dc-4a75-8999-53df29d2fb5d\") " Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.024865 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-rabbitmq-plugins\") pod \"7b4f22a1-41dc-4a75-8999-53df29d2fb5d\" (UID: \"7b4f22a1-41dc-4a75-8999-53df29d2fb5d\") " Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.025904 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "7b4f22a1-41dc-4a75-8999-53df29d2fb5d" (UID: "7b4f22a1-41dc-4a75-8999-53df29d2fb5d"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.026381 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "7b4f22a1-41dc-4a75-8999-53df29d2fb5d" (UID: "7b4f22a1-41dc-4a75-8999-53df29d2fb5d"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.026691 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "7b4f22a1-41dc-4a75-8999-53df29d2fb5d" (UID: "7b4f22a1-41dc-4a75-8999-53df29d2fb5d"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.030435 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "7b4f22a1-41dc-4a75-8999-53df29d2fb5d" (UID: "7b4f22a1-41dc-4a75-8999-53df29d2fb5d"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.032469 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "7b4f22a1-41dc-4a75-8999-53df29d2fb5d" (UID: "7b4f22a1-41dc-4a75-8999-53df29d2fb5d"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.032723 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "persistence") pod "7b4f22a1-41dc-4a75-8999-53df29d2fb5d" (UID: "7b4f22a1-41dc-4a75-8999-53df29d2fb5d"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.043375 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-pod-info" (OuterVolumeSpecName: "pod-info") pod "7b4f22a1-41dc-4a75-8999-53df29d2fb5d" (UID: "7b4f22a1-41dc-4a75-8999-53df29d2fb5d"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.078825 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-kube-api-access-cjz5x" (OuterVolumeSpecName: "kube-api-access-cjz5x") pod "7b4f22a1-41dc-4a75-8999-53df29d2fb5d" (UID: "7b4f22a1-41dc-4a75-8999-53df29d2fb5d"). InnerVolumeSpecName "kube-api-access-cjz5x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.079338 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-config-data" (OuterVolumeSpecName: "config-data") pod "7b4f22a1-41dc-4a75-8999-53df29d2fb5d" (UID: "7b4f22a1-41dc-4a75-8999-53df29d2fb5d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.107351 4802 generic.go:334] "Generic (PLEG): container finished" podID="8c8b6c49-6f71-403d-b639-4a17ae3bc802" containerID="098aed31247101a27a8ff5c8dd2bfd3731dd532b7504c6cf5431f78994869852" exitCode=0 Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.107473 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8c8b6c49-6f71-403d-b639-4a17ae3bc802","Type":"ContainerDied","Data":"098aed31247101a27a8ff5c8dd2bfd3731dd532b7504c6cf5431f78994869852"} Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.113436 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-server-conf" (OuterVolumeSpecName: "server-conf") pod "7b4f22a1-41dc-4a75-8999-53df29d2fb5d" (UID: "7b4f22a1-41dc-4a75-8999-53df29d2fb5d"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.115090 4802 generic.go:334] "Generic (PLEG): container finished" podID="7b4f22a1-41dc-4a75-8999-53df29d2fb5d" containerID="ed17f5b81a54486ceb0680c299b5f84abc02bd1e687d265819e37dc8b4570d6b" exitCode=0 Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.115129 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7b4f22a1-41dc-4a75-8999-53df29d2fb5d","Type":"ContainerDied","Data":"ed17f5b81a54486ceb0680c299b5f84abc02bd1e687d265819e37dc8b4570d6b"} Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.115155 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7b4f22a1-41dc-4a75-8999-53df29d2fb5d","Type":"ContainerDied","Data":"73c1739eed3a5ff0dcfc6eb179c448a0489e80c5312852db4d5873a762d3b055"} Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.115173 4802 scope.go:117] "RemoveContainer" containerID="ed17f5b81a54486ceb0680c299b5f84abc02bd1e687d265819e37dc8b4570d6b" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.115292 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.127909 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cjz5x\" (UniqueName: \"kubernetes.io/projected/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-kube-api-access-cjz5x\") on node \"crc\" DevicePath \"\"" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.127946 4802 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.127960 4802 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.127973 4802 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.127984 4802 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.127995 4802 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-server-conf\") on node \"crc\" DevicePath \"\"" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.128006 4802 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.128019 4802 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-pod-info\") on node \"crc\" DevicePath \"\"" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.128030 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.128058 4802 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.149304 4802 scope.go:117] "RemoveContainer" containerID="fd1082f2aaf9365d3d9655c6924fa59bdb4c429364af5f69505c71489894ed4f" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.163675 4802 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.170878 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.178098 4802 scope.go:117] "RemoveContainer" containerID="ed17f5b81a54486ceb0680c299b5f84abc02bd1e687d265819e37dc8b4570d6b" Dec 06 04:02:59 crc kubenswrapper[4802]: E1206 04:02:59.180353 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed17f5b81a54486ceb0680c299b5f84abc02bd1e687d265819e37dc8b4570d6b\": container with ID starting with ed17f5b81a54486ceb0680c299b5f84abc02bd1e687d265819e37dc8b4570d6b not found: ID does not exist" containerID="ed17f5b81a54486ceb0680c299b5f84abc02bd1e687d265819e37dc8b4570d6b" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.180395 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed17f5b81a54486ceb0680c299b5f84abc02bd1e687d265819e37dc8b4570d6b"} err="failed to get container status \"ed17f5b81a54486ceb0680c299b5f84abc02bd1e687d265819e37dc8b4570d6b\": rpc error: code = NotFound desc = could not find container \"ed17f5b81a54486ceb0680c299b5f84abc02bd1e687d265819e37dc8b4570d6b\": container with ID starting with ed17f5b81a54486ceb0680c299b5f84abc02bd1e687d265819e37dc8b4570d6b not found: ID does not exist" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.180416 4802 scope.go:117] "RemoveContainer" containerID="fd1082f2aaf9365d3d9655c6924fa59bdb4c429364af5f69505c71489894ed4f" Dec 06 04:02:59 crc kubenswrapper[4802]: E1206 04:02:59.180843 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd1082f2aaf9365d3d9655c6924fa59bdb4c429364af5f69505c71489894ed4f\": container with ID starting with fd1082f2aaf9365d3d9655c6924fa59bdb4c429364af5f69505c71489894ed4f not found: ID does not exist" containerID="fd1082f2aaf9365d3d9655c6924fa59bdb4c429364af5f69505c71489894ed4f" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.180871 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd1082f2aaf9365d3d9655c6924fa59bdb4c429364af5f69505c71489894ed4f"} err="failed to get container status \"fd1082f2aaf9365d3d9655c6924fa59bdb4c429364af5f69505c71489894ed4f\": rpc error: code = NotFound desc = could not find container \"fd1082f2aaf9365d3d9655c6924fa59bdb4c429364af5f69505c71489894ed4f\": container with ID starting with fd1082f2aaf9365d3d9655c6924fa59bdb4c429364af5f69505c71489894ed4f not found: ID does not exist" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.185153 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "7b4f22a1-41dc-4a75-8999-53df29d2fb5d" (UID: "7b4f22a1-41dc-4a75-8999-53df29d2fb5d"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.229893 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8c8b6c49-6f71-403d-b639-4a17ae3bc802-erlang-cookie-secret\") pod \"8c8b6c49-6f71-403d-b639-4a17ae3bc802\" (UID: \"8c8b6c49-6f71-403d-b639-4a17ae3bc802\") " Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.229970 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"8c8b6c49-6f71-403d-b639-4a17ae3bc802\" (UID: \"8c8b6c49-6f71-403d-b639-4a17ae3bc802\") " Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.230025 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8c8b6c49-6f71-403d-b639-4a17ae3bc802-rabbitmq-confd\") pod \"8c8b6c49-6f71-403d-b639-4a17ae3bc802\" (UID: \"8c8b6c49-6f71-403d-b639-4a17ae3bc802\") " Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.230059 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8c8b6c49-6f71-403d-b639-4a17ae3bc802-rabbitmq-erlang-cookie\") pod \"8c8b6c49-6f71-403d-b639-4a17ae3bc802\" (UID: \"8c8b6c49-6f71-403d-b639-4a17ae3bc802\") " Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.230122 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8c8b6c49-6f71-403d-b639-4a17ae3bc802-rabbitmq-tls\") pod \"8c8b6c49-6f71-403d-b639-4a17ae3bc802\" (UID: \"8c8b6c49-6f71-403d-b639-4a17ae3bc802\") " Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.230151 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8c8b6c49-6f71-403d-b639-4a17ae3bc802-plugins-conf\") pod \"8c8b6c49-6f71-403d-b639-4a17ae3bc802\" (UID: \"8c8b6c49-6f71-403d-b639-4a17ae3bc802\") " Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.230264 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8c8b6c49-6f71-403d-b639-4a17ae3bc802-config-data\") pod \"8c8b6c49-6f71-403d-b639-4a17ae3bc802\" (UID: \"8c8b6c49-6f71-403d-b639-4a17ae3bc802\") " Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.230463 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9lzq4\" (UniqueName: \"kubernetes.io/projected/8c8b6c49-6f71-403d-b639-4a17ae3bc802-kube-api-access-9lzq4\") pod \"8c8b6c49-6f71-403d-b639-4a17ae3bc802\" (UID: \"8c8b6c49-6f71-403d-b639-4a17ae3bc802\") " Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.230547 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8c8b6c49-6f71-403d-b639-4a17ae3bc802-server-conf\") pod \"8c8b6c49-6f71-403d-b639-4a17ae3bc802\" (UID: \"8c8b6c49-6f71-403d-b639-4a17ae3bc802\") " Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.230618 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8c8b6c49-6f71-403d-b639-4a17ae3bc802-rabbitmq-plugins\") pod \"8c8b6c49-6f71-403d-b639-4a17ae3bc802\" (UID: \"8c8b6c49-6f71-403d-b639-4a17ae3bc802\") " Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.230658 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8c8b6c49-6f71-403d-b639-4a17ae3bc802-pod-info\") pod \"8c8b6c49-6f71-403d-b639-4a17ae3bc802\" (UID: \"8c8b6c49-6f71-403d-b639-4a17ae3bc802\") " Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.231294 4802 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7b4f22a1-41dc-4a75-8999-53df29d2fb5d-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.231321 4802 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.231688 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c8b6c49-6f71-403d-b639-4a17ae3bc802-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "8c8b6c49-6f71-403d-b639-4a17ae3bc802" (UID: "8c8b6c49-6f71-403d-b639-4a17ae3bc802"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.232227 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c8b6c49-6f71-403d-b639-4a17ae3bc802-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "8c8b6c49-6f71-403d-b639-4a17ae3bc802" (UID: "8c8b6c49-6f71-403d-b639-4a17ae3bc802"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.234599 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c8b6c49-6f71-403d-b639-4a17ae3bc802-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "8c8b6c49-6f71-403d-b639-4a17ae3bc802" (UID: "8c8b6c49-6f71-403d-b639-4a17ae3bc802"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.235157 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/8c8b6c49-6f71-403d-b639-4a17ae3bc802-pod-info" (OuterVolumeSpecName: "pod-info") pod "8c8b6c49-6f71-403d-b639-4a17ae3bc802" (UID: "8c8b6c49-6f71-403d-b639-4a17ae3bc802"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.236101 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c8b6c49-6f71-403d-b639-4a17ae3bc802-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "8c8b6c49-6f71-403d-b639-4a17ae3bc802" (UID: "8c8b6c49-6f71-403d-b639-4a17ae3bc802"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.236240 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c8b6c49-6f71-403d-b639-4a17ae3bc802-kube-api-access-9lzq4" (OuterVolumeSpecName: "kube-api-access-9lzq4") pod "8c8b6c49-6f71-403d-b639-4a17ae3bc802" (UID: "8c8b6c49-6f71-403d-b639-4a17ae3bc802"). InnerVolumeSpecName "kube-api-access-9lzq4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.240329 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c8b6c49-6f71-403d-b639-4a17ae3bc802-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "8c8b6c49-6f71-403d-b639-4a17ae3bc802" (UID: "8c8b6c49-6f71-403d-b639-4a17ae3bc802"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.244922 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "persistence") pod "8c8b6c49-6f71-403d-b639-4a17ae3bc802" (UID: "8c8b6c49-6f71-403d-b639-4a17ae3bc802"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.275236 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c8b6c49-6f71-403d-b639-4a17ae3bc802-config-data" (OuterVolumeSpecName: "config-data") pod "8c8b6c49-6f71-403d-b639-4a17ae3bc802" (UID: "8c8b6c49-6f71-403d-b639-4a17ae3bc802"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.311295 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c8b6c49-6f71-403d-b639-4a17ae3bc802-server-conf" (OuterVolumeSpecName: "server-conf") pod "8c8b6c49-6f71-403d-b639-4a17ae3bc802" (UID: "8c8b6c49-6f71-403d-b639-4a17ae3bc802"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.333090 4802 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8c8b6c49-6f71-403d-b639-4a17ae3bc802-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.333119 4802 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8c8b6c49-6f71-403d-b639-4a17ae3bc802-pod-info\") on node \"crc\" DevicePath \"\"" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.333128 4802 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8c8b6c49-6f71-403d-b639-4a17ae3bc802-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.333154 4802 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.333165 4802 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8c8b6c49-6f71-403d-b639-4a17ae3bc802-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.333174 4802 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8c8b6c49-6f71-403d-b639-4a17ae3bc802-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.333186 4802 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8c8b6c49-6f71-403d-b639-4a17ae3bc802-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.333194 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8c8b6c49-6f71-403d-b639-4a17ae3bc802-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.333203 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9lzq4\" (UniqueName: \"kubernetes.io/projected/8c8b6c49-6f71-403d-b639-4a17ae3bc802-kube-api-access-9lzq4\") on node \"crc\" DevicePath \"\"" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.333211 4802 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8c8b6c49-6f71-403d-b639-4a17ae3bc802-server-conf\") on node \"crc\" DevicePath \"\"" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.388384 4802 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.439933 4802 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.458038 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c8b6c49-6f71-403d-b639-4a17ae3bc802-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "8c8b6c49-6f71-403d-b639-4a17ae3bc802" (UID: "8c8b6c49-6f71-403d-b639-4a17ae3bc802"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.536914 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.542052 4802 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8c8b6c49-6f71-403d-b639-4a17ae3bc802-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.571911 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.584504 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 04:02:59 crc kubenswrapper[4802]: E1206 04:02:59.584971 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c8b6c49-6f71-403d-b639-4a17ae3bc802" containerName="setup-container" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.584989 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c8b6c49-6f71-403d-b639-4a17ae3bc802" containerName="setup-container" Dec 06 04:02:59 crc kubenswrapper[4802]: E1206 04:02:59.585004 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b4f22a1-41dc-4a75-8999-53df29d2fb5d" containerName="setup-container" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.585011 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b4f22a1-41dc-4a75-8999-53df29d2fb5d" containerName="setup-container" Dec 06 04:02:59 crc kubenswrapper[4802]: E1206 04:02:59.585045 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b4f22a1-41dc-4a75-8999-53df29d2fb5d" containerName="rabbitmq" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.585051 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b4f22a1-41dc-4a75-8999-53df29d2fb5d" containerName="rabbitmq" Dec 06 04:02:59 crc kubenswrapper[4802]: E1206 04:02:59.585068 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c8b6c49-6f71-403d-b639-4a17ae3bc802" containerName="rabbitmq" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.585073 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c8b6c49-6f71-403d-b639-4a17ae3bc802" containerName="rabbitmq" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.585237 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b4f22a1-41dc-4a75-8999-53df29d2fb5d" containerName="rabbitmq" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.585252 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c8b6c49-6f71-403d-b639-4a17ae3bc802" containerName="rabbitmq" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.586213 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.588633 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.589267 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.589601 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-8mwrg" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.589742 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.589913 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.590346 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.590499 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.601244 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.751982 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/cdc37fb9-aeae-4fe0-b86e-e9d41e48314b-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"cdc37fb9-aeae-4fe0-b86e-e9d41e48314b\") " pod="openstack/rabbitmq-server-0" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.752051 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/cdc37fb9-aeae-4fe0-b86e-e9d41e48314b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"cdc37fb9-aeae-4fe0-b86e-e9d41e48314b\") " pod="openstack/rabbitmq-server-0" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.752077 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/cdc37fb9-aeae-4fe0-b86e-e9d41e48314b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"cdc37fb9-aeae-4fe0-b86e-e9d41e48314b\") " pod="openstack/rabbitmq-server-0" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.752116 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cdc37fb9-aeae-4fe0-b86e-e9d41e48314b-config-data\") pod \"rabbitmq-server-0\" (UID: \"cdc37fb9-aeae-4fe0-b86e-e9d41e48314b\") " pod="openstack/rabbitmq-server-0" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.752138 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/cdc37fb9-aeae-4fe0-b86e-e9d41e48314b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"cdc37fb9-aeae-4fe0-b86e-e9d41e48314b\") " pod="openstack/rabbitmq-server-0" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.752168 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/cdc37fb9-aeae-4fe0-b86e-e9d41e48314b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"cdc37fb9-aeae-4fe0-b86e-e9d41e48314b\") " pod="openstack/rabbitmq-server-0" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.752221 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/cdc37fb9-aeae-4fe0-b86e-e9d41e48314b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"cdc37fb9-aeae-4fe0-b86e-e9d41e48314b\") " pod="openstack/rabbitmq-server-0" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.752241 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkfxg\" (UniqueName: \"kubernetes.io/projected/cdc37fb9-aeae-4fe0-b86e-e9d41e48314b-kube-api-access-gkfxg\") pod \"rabbitmq-server-0\" (UID: \"cdc37fb9-aeae-4fe0-b86e-e9d41e48314b\") " pod="openstack/rabbitmq-server-0" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.752310 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/cdc37fb9-aeae-4fe0-b86e-e9d41e48314b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"cdc37fb9-aeae-4fe0-b86e-e9d41e48314b\") " pod="openstack/rabbitmq-server-0" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.752335 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/cdc37fb9-aeae-4fe0-b86e-e9d41e48314b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"cdc37fb9-aeae-4fe0-b86e-e9d41e48314b\") " pod="openstack/rabbitmq-server-0" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.752355 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-server-0\" (UID: \"cdc37fb9-aeae-4fe0-b86e-e9d41e48314b\") " pod="openstack/rabbitmq-server-0" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.853833 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/cdc37fb9-aeae-4fe0-b86e-e9d41e48314b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"cdc37fb9-aeae-4fe0-b86e-e9d41e48314b\") " pod="openstack/rabbitmq-server-0" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.853907 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/cdc37fb9-aeae-4fe0-b86e-e9d41e48314b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"cdc37fb9-aeae-4fe0-b86e-e9d41e48314b\") " pod="openstack/rabbitmq-server-0" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.853941 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/cdc37fb9-aeae-4fe0-b86e-e9d41e48314b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"cdc37fb9-aeae-4fe0-b86e-e9d41e48314b\") " pod="openstack/rabbitmq-server-0" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.853971 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkfxg\" (UniqueName: \"kubernetes.io/projected/cdc37fb9-aeae-4fe0-b86e-e9d41e48314b-kube-api-access-gkfxg\") pod \"rabbitmq-server-0\" (UID: \"cdc37fb9-aeae-4fe0-b86e-e9d41e48314b\") " pod="openstack/rabbitmq-server-0" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.854042 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/cdc37fb9-aeae-4fe0-b86e-e9d41e48314b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"cdc37fb9-aeae-4fe0-b86e-e9d41e48314b\") " pod="openstack/rabbitmq-server-0" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.854073 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/cdc37fb9-aeae-4fe0-b86e-e9d41e48314b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"cdc37fb9-aeae-4fe0-b86e-e9d41e48314b\") " pod="openstack/rabbitmq-server-0" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.854100 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-server-0\" (UID: \"cdc37fb9-aeae-4fe0-b86e-e9d41e48314b\") " pod="openstack/rabbitmq-server-0" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.854123 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/cdc37fb9-aeae-4fe0-b86e-e9d41e48314b-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"cdc37fb9-aeae-4fe0-b86e-e9d41e48314b\") " pod="openstack/rabbitmq-server-0" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.854165 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/cdc37fb9-aeae-4fe0-b86e-e9d41e48314b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"cdc37fb9-aeae-4fe0-b86e-e9d41e48314b\") " pod="openstack/rabbitmq-server-0" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.854189 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/cdc37fb9-aeae-4fe0-b86e-e9d41e48314b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"cdc37fb9-aeae-4fe0-b86e-e9d41e48314b\") " pod="openstack/rabbitmq-server-0" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.854229 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cdc37fb9-aeae-4fe0-b86e-e9d41e48314b-config-data\") pod \"rabbitmq-server-0\" (UID: \"cdc37fb9-aeae-4fe0-b86e-e9d41e48314b\") " pod="openstack/rabbitmq-server-0" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.854383 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/cdc37fb9-aeae-4fe0-b86e-e9d41e48314b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"cdc37fb9-aeae-4fe0-b86e-e9d41e48314b\") " pod="openstack/rabbitmq-server-0" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.854956 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/cdc37fb9-aeae-4fe0-b86e-e9d41e48314b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"cdc37fb9-aeae-4fe0-b86e-e9d41e48314b\") " pod="openstack/rabbitmq-server-0" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.855084 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cdc37fb9-aeae-4fe0-b86e-e9d41e48314b-config-data\") pod \"rabbitmq-server-0\" (UID: \"cdc37fb9-aeae-4fe0-b86e-e9d41e48314b\") " pod="openstack/rabbitmq-server-0" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.855685 4802 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-server-0\" (UID: \"cdc37fb9-aeae-4fe0-b86e-e9d41e48314b\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/rabbitmq-server-0" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.856605 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/cdc37fb9-aeae-4fe0-b86e-e9d41e48314b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"cdc37fb9-aeae-4fe0-b86e-e9d41e48314b\") " pod="openstack/rabbitmq-server-0" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.859442 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/cdc37fb9-aeae-4fe0-b86e-e9d41e48314b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"cdc37fb9-aeae-4fe0-b86e-e9d41e48314b\") " pod="openstack/rabbitmq-server-0" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.859781 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/cdc37fb9-aeae-4fe0-b86e-e9d41e48314b-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"cdc37fb9-aeae-4fe0-b86e-e9d41e48314b\") " pod="openstack/rabbitmq-server-0" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.859832 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/cdc37fb9-aeae-4fe0-b86e-e9d41e48314b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"cdc37fb9-aeae-4fe0-b86e-e9d41e48314b\") " pod="openstack/rabbitmq-server-0" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.864154 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/cdc37fb9-aeae-4fe0-b86e-e9d41e48314b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"cdc37fb9-aeae-4fe0-b86e-e9d41e48314b\") " pod="openstack/rabbitmq-server-0" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.864201 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/cdc37fb9-aeae-4fe0-b86e-e9d41e48314b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"cdc37fb9-aeae-4fe0-b86e-e9d41e48314b\") " pod="openstack/rabbitmq-server-0" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.873077 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkfxg\" (UniqueName: \"kubernetes.io/projected/cdc37fb9-aeae-4fe0-b86e-e9d41e48314b-kube-api-access-gkfxg\") pod \"rabbitmq-server-0\" (UID: \"cdc37fb9-aeae-4fe0-b86e-e9d41e48314b\") " pod="openstack/rabbitmq-server-0" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.893377 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-server-0\" (UID: \"cdc37fb9-aeae-4fe0-b86e-e9d41e48314b\") " pod="openstack/rabbitmq-server-0" Dec 06 04:02:59 crc kubenswrapper[4802]: I1206 04:02:59.924990 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 06 04:03:00 crc kubenswrapper[4802]: I1206 04:03:00.130403 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8c8b6c49-6f71-403d-b639-4a17ae3bc802","Type":"ContainerDied","Data":"9560d5c1751b3a124c55a3571e5ffd1725570c0666099243625f058a51958f87"} Dec 06 04:03:00 crc kubenswrapper[4802]: I1206 04:03:00.130669 4802 scope.go:117] "RemoveContainer" containerID="098aed31247101a27a8ff5c8dd2bfd3731dd532b7504c6cf5431f78994869852" Dec 06 04:03:00 crc kubenswrapper[4802]: I1206 04:03:00.130624 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 06 04:03:00 crc kubenswrapper[4802]: I1206 04:03:00.157112 4802 scope.go:117] "RemoveContainer" containerID="f470e3dda7735fd4bf057dbf62cde306583097f22e2d9c550167fc04c5e92f72" Dec 06 04:03:00 crc kubenswrapper[4802]: I1206 04:03:00.171741 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 04:03:00 crc kubenswrapper[4802]: I1206 04:03:00.183532 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 04:03:00 crc kubenswrapper[4802]: I1206 04:03:00.189926 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 04:03:00 crc kubenswrapper[4802]: I1206 04:03:00.195046 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 06 04:03:00 crc kubenswrapper[4802]: I1206 04:03:00.198742 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 06 04:03:00 crc kubenswrapper[4802]: I1206 04:03:00.199055 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 06 04:03:00 crc kubenswrapper[4802]: I1206 04:03:00.199371 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 06 04:03:00 crc kubenswrapper[4802]: I1206 04:03:00.199631 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 06 04:03:00 crc kubenswrapper[4802]: I1206 04:03:00.199795 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 06 04:03:00 crc kubenswrapper[4802]: I1206 04:03:00.199953 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-lf7lb" Dec 06 04:03:00 crc kubenswrapper[4802]: I1206 04:03:00.200568 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 06 04:03:00 crc kubenswrapper[4802]: I1206 04:03:00.210046 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 04:03:00 crc kubenswrapper[4802]: I1206 04:03:00.363510 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/cdcf8137-9bbe-467f-bbb3-e30f350760a6-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"cdcf8137-9bbe-467f-bbb3-e30f350760a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 04:03:00 crc kubenswrapper[4802]: I1206 04:03:00.364121 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/cdcf8137-9bbe-467f-bbb3-e30f350760a6-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"cdcf8137-9bbe-467f-bbb3-e30f350760a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 04:03:00 crc kubenswrapper[4802]: I1206 04:03:00.364295 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/cdcf8137-9bbe-467f-bbb3-e30f350760a6-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"cdcf8137-9bbe-467f-bbb3-e30f350760a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 04:03:00 crc kubenswrapper[4802]: I1206 04:03:00.364503 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/cdcf8137-9bbe-467f-bbb3-e30f350760a6-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"cdcf8137-9bbe-467f-bbb3-e30f350760a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 04:03:00 crc kubenswrapper[4802]: I1206 04:03:00.364683 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"cdcf8137-9bbe-467f-bbb3-e30f350760a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 04:03:00 crc kubenswrapper[4802]: I1206 04:03:00.364904 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cdcf8137-9bbe-467f-bbb3-e30f350760a6-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"cdcf8137-9bbe-467f-bbb3-e30f350760a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 04:03:00 crc kubenswrapper[4802]: I1206 04:03:00.365120 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/cdcf8137-9bbe-467f-bbb3-e30f350760a6-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"cdcf8137-9bbe-467f-bbb3-e30f350760a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 04:03:00 crc kubenswrapper[4802]: I1206 04:03:00.365314 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/cdcf8137-9bbe-467f-bbb3-e30f350760a6-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"cdcf8137-9bbe-467f-bbb3-e30f350760a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 04:03:00 crc kubenswrapper[4802]: I1206 04:03:00.365602 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/cdcf8137-9bbe-467f-bbb3-e30f350760a6-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"cdcf8137-9bbe-467f-bbb3-e30f350760a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 04:03:00 crc kubenswrapper[4802]: I1206 04:03:00.365792 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r474t\" (UniqueName: \"kubernetes.io/projected/cdcf8137-9bbe-467f-bbb3-e30f350760a6-kube-api-access-r474t\") pod \"rabbitmq-cell1-server-0\" (UID: \"cdcf8137-9bbe-467f-bbb3-e30f350760a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 04:03:00 crc kubenswrapper[4802]: I1206 04:03:00.366004 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/cdcf8137-9bbe-467f-bbb3-e30f350760a6-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"cdcf8137-9bbe-467f-bbb3-e30f350760a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 04:03:00 crc kubenswrapper[4802]: I1206 04:03:00.460122 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 04:03:00 crc kubenswrapper[4802]: I1206 04:03:00.467099 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"cdcf8137-9bbe-467f-bbb3-e30f350760a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 04:03:00 crc kubenswrapper[4802]: I1206 04:03:00.467778 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cdcf8137-9bbe-467f-bbb3-e30f350760a6-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"cdcf8137-9bbe-467f-bbb3-e30f350760a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 04:03:00 crc kubenswrapper[4802]: I1206 04:03:00.467807 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/cdcf8137-9bbe-467f-bbb3-e30f350760a6-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"cdcf8137-9bbe-467f-bbb3-e30f350760a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 04:03:00 crc kubenswrapper[4802]: I1206 04:03:00.467842 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/cdcf8137-9bbe-467f-bbb3-e30f350760a6-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"cdcf8137-9bbe-467f-bbb3-e30f350760a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 04:03:00 crc kubenswrapper[4802]: I1206 04:03:00.467902 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/cdcf8137-9bbe-467f-bbb3-e30f350760a6-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"cdcf8137-9bbe-467f-bbb3-e30f350760a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 04:03:00 crc kubenswrapper[4802]: I1206 04:03:00.467918 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r474t\" (UniqueName: \"kubernetes.io/projected/cdcf8137-9bbe-467f-bbb3-e30f350760a6-kube-api-access-r474t\") pod \"rabbitmq-cell1-server-0\" (UID: \"cdcf8137-9bbe-467f-bbb3-e30f350760a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 04:03:00 crc kubenswrapper[4802]: I1206 04:03:00.467942 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/cdcf8137-9bbe-467f-bbb3-e30f350760a6-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"cdcf8137-9bbe-467f-bbb3-e30f350760a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 04:03:00 crc kubenswrapper[4802]: I1206 04:03:00.467984 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/cdcf8137-9bbe-467f-bbb3-e30f350760a6-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"cdcf8137-9bbe-467f-bbb3-e30f350760a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 04:03:00 crc kubenswrapper[4802]: I1206 04:03:00.468003 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/cdcf8137-9bbe-467f-bbb3-e30f350760a6-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"cdcf8137-9bbe-467f-bbb3-e30f350760a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 04:03:00 crc kubenswrapper[4802]: I1206 04:03:00.468021 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/cdcf8137-9bbe-467f-bbb3-e30f350760a6-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"cdcf8137-9bbe-467f-bbb3-e30f350760a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 04:03:00 crc kubenswrapper[4802]: I1206 04:03:00.468055 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/cdcf8137-9bbe-467f-bbb3-e30f350760a6-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"cdcf8137-9bbe-467f-bbb3-e30f350760a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 04:03:00 crc kubenswrapper[4802]: I1206 04:03:00.467680 4802 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"cdcf8137-9bbe-467f-bbb3-e30f350760a6\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/rabbitmq-cell1-server-0" Dec 06 04:03:00 crc kubenswrapper[4802]: I1206 04:03:00.470132 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cdcf8137-9bbe-467f-bbb3-e30f350760a6-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"cdcf8137-9bbe-467f-bbb3-e30f350760a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 04:03:00 crc kubenswrapper[4802]: I1206 04:03:00.470594 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/cdcf8137-9bbe-467f-bbb3-e30f350760a6-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"cdcf8137-9bbe-467f-bbb3-e30f350760a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 04:03:00 crc kubenswrapper[4802]: I1206 04:03:00.471203 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/cdcf8137-9bbe-467f-bbb3-e30f350760a6-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"cdcf8137-9bbe-467f-bbb3-e30f350760a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 04:03:00 crc kubenswrapper[4802]: I1206 04:03:00.471598 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/cdcf8137-9bbe-467f-bbb3-e30f350760a6-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"cdcf8137-9bbe-467f-bbb3-e30f350760a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 04:03:00 crc kubenswrapper[4802]: I1206 04:03:00.471701 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/cdcf8137-9bbe-467f-bbb3-e30f350760a6-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"cdcf8137-9bbe-467f-bbb3-e30f350760a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 04:03:00 crc kubenswrapper[4802]: I1206 04:03:00.474383 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/cdcf8137-9bbe-467f-bbb3-e30f350760a6-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"cdcf8137-9bbe-467f-bbb3-e30f350760a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 04:03:00 crc kubenswrapper[4802]: I1206 04:03:00.476320 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/cdcf8137-9bbe-467f-bbb3-e30f350760a6-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"cdcf8137-9bbe-467f-bbb3-e30f350760a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 04:03:00 crc kubenswrapper[4802]: I1206 04:03:00.478825 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/cdcf8137-9bbe-467f-bbb3-e30f350760a6-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"cdcf8137-9bbe-467f-bbb3-e30f350760a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 04:03:00 crc kubenswrapper[4802]: I1206 04:03:00.490227 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/cdcf8137-9bbe-467f-bbb3-e30f350760a6-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"cdcf8137-9bbe-467f-bbb3-e30f350760a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 04:03:00 crc kubenswrapper[4802]: I1206 04:03:00.492302 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r474t\" (UniqueName: \"kubernetes.io/projected/cdcf8137-9bbe-467f-bbb3-e30f350760a6-kube-api-access-r474t\") pod \"rabbitmq-cell1-server-0\" (UID: \"cdcf8137-9bbe-467f-bbb3-e30f350760a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 04:03:00 crc kubenswrapper[4802]: I1206 04:03:00.533274 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"cdcf8137-9bbe-467f-bbb3-e30f350760a6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 04:03:00 crc kubenswrapper[4802]: I1206 04:03:00.825016 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 06 04:03:01 crc kubenswrapper[4802]: I1206 04:03:01.144870 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"cdc37fb9-aeae-4fe0-b86e-e9d41e48314b","Type":"ContainerStarted","Data":"256519d48090f3763a715dc3a1b3efa9637c6bf50f7313ecdc8a9e1080e0e93b"} Dec 06 04:03:01 crc kubenswrapper[4802]: W1206 04:03:01.335782 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcdcf8137_9bbe_467f_bbb3_e30f350760a6.slice/crio-cb4d6d36141b185c2b4472973cf9e79e8a2525618710f7813ddf4792862eb329 WatchSource:0}: Error finding container cb4d6d36141b185c2b4472973cf9e79e8a2525618710f7813ddf4792862eb329: Status 404 returned error can't find the container with id cb4d6d36141b185c2b4472973cf9e79e8a2525618710f7813ddf4792862eb329 Dec 06 04:03:01 crc kubenswrapper[4802]: I1206 04:03:01.336343 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 04:03:01 crc kubenswrapper[4802]: I1206 04:03:01.466463 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b4f22a1-41dc-4a75-8999-53df29d2fb5d" path="/var/lib/kubelet/pods/7b4f22a1-41dc-4a75-8999-53df29d2fb5d/volumes" Dec 06 04:03:01 crc kubenswrapper[4802]: I1206 04:03:01.468197 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c8b6c49-6f71-403d-b639-4a17ae3bc802" path="/var/lib/kubelet/pods/8c8b6c49-6f71-403d-b639-4a17ae3bc802/volumes" Dec 06 04:03:02 crc kubenswrapper[4802]: I1206 04:03:02.039281 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b75489c6f-btb4g"] Dec 06 04:03:02 crc kubenswrapper[4802]: I1206 04:03:02.041217 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b75489c6f-btb4g" Dec 06 04:03:02 crc kubenswrapper[4802]: I1206 04:03:02.043357 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Dec 06 04:03:02 crc kubenswrapper[4802]: I1206 04:03:02.085562 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b75489c6f-btb4g"] Dec 06 04:03:02 crc kubenswrapper[4802]: I1206 04:03:02.120879 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/55d5e710-5432-48a4-b0ac-a865f823aae9-ovsdbserver-nb\") pod \"dnsmasq-dns-5b75489c6f-btb4g\" (UID: \"55d5e710-5432-48a4-b0ac-a865f823aae9\") " pod="openstack/dnsmasq-dns-5b75489c6f-btb4g" Dec 06 04:03:02 crc kubenswrapper[4802]: I1206 04:03:02.120949 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvfmw\" (UniqueName: \"kubernetes.io/projected/55d5e710-5432-48a4-b0ac-a865f823aae9-kube-api-access-tvfmw\") pod \"dnsmasq-dns-5b75489c6f-btb4g\" (UID: \"55d5e710-5432-48a4-b0ac-a865f823aae9\") " pod="openstack/dnsmasq-dns-5b75489c6f-btb4g" Dec 06 04:03:02 crc kubenswrapper[4802]: I1206 04:03:02.120982 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/55d5e710-5432-48a4-b0ac-a865f823aae9-ovsdbserver-sb\") pod \"dnsmasq-dns-5b75489c6f-btb4g\" (UID: \"55d5e710-5432-48a4-b0ac-a865f823aae9\") " pod="openstack/dnsmasq-dns-5b75489c6f-btb4g" Dec 06 04:03:02 crc kubenswrapper[4802]: I1206 04:03:02.121099 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/55d5e710-5432-48a4-b0ac-a865f823aae9-dns-swift-storage-0\") pod \"dnsmasq-dns-5b75489c6f-btb4g\" (UID: \"55d5e710-5432-48a4-b0ac-a865f823aae9\") " pod="openstack/dnsmasq-dns-5b75489c6f-btb4g" Dec 06 04:03:02 crc kubenswrapper[4802]: I1206 04:03:02.121357 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/55d5e710-5432-48a4-b0ac-a865f823aae9-dns-svc\") pod \"dnsmasq-dns-5b75489c6f-btb4g\" (UID: \"55d5e710-5432-48a4-b0ac-a865f823aae9\") " pod="openstack/dnsmasq-dns-5b75489c6f-btb4g" Dec 06 04:03:02 crc kubenswrapper[4802]: I1206 04:03:02.121453 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55d5e710-5432-48a4-b0ac-a865f823aae9-config\") pod \"dnsmasq-dns-5b75489c6f-btb4g\" (UID: \"55d5e710-5432-48a4-b0ac-a865f823aae9\") " pod="openstack/dnsmasq-dns-5b75489c6f-btb4g" Dec 06 04:03:02 crc kubenswrapper[4802]: I1206 04:03:02.121583 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/55d5e710-5432-48a4-b0ac-a865f823aae9-openstack-edpm-ipam\") pod \"dnsmasq-dns-5b75489c6f-btb4g\" (UID: \"55d5e710-5432-48a4-b0ac-a865f823aae9\") " pod="openstack/dnsmasq-dns-5b75489c6f-btb4g" Dec 06 04:03:02 crc kubenswrapper[4802]: I1206 04:03:02.155703 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"cdcf8137-9bbe-467f-bbb3-e30f350760a6","Type":"ContainerStarted","Data":"cb4d6d36141b185c2b4472973cf9e79e8a2525618710f7813ddf4792862eb329"} Dec 06 04:03:02 crc kubenswrapper[4802]: I1206 04:03:02.223443 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/55d5e710-5432-48a4-b0ac-a865f823aae9-dns-svc\") pod \"dnsmasq-dns-5b75489c6f-btb4g\" (UID: \"55d5e710-5432-48a4-b0ac-a865f823aae9\") " pod="openstack/dnsmasq-dns-5b75489c6f-btb4g" Dec 06 04:03:02 crc kubenswrapper[4802]: I1206 04:03:02.223557 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55d5e710-5432-48a4-b0ac-a865f823aae9-config\") pod \"dnsmasq-dns-5b75489c6f-btb4g\" (UID: \"55d5e710-5432-48a4-b0ac-a865f823aae9\") " pod="openstack/dnsmasq-dns-5b75489c6f-btb4g" Dec 06 04:03:02 crc kubenswrapper[4802]: I1206 04:03:02.223632 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/55d5e710-5432-48a4-b0ac-a865f823aae9-openstack-edpm-ipam\") pod \"dnsmasq-dns-5b75489c6f-btb4g\" (UID: \"55d5e710-5432-48a4-b0ac-a865f823aae9\") " pod="openstack/dnsmasq-dns-5b75489c6f-btb4g" Dec 06 04:03:02 crc kubenswrapper[4802]: I1206 04:03:02.223705 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/55d5e710-5432-48a4-b0ac-a865f823aae9-ovsdbserver-nb\") pod \"dnsmasq-dns-5b75489c6f-btb4g\" (UID: \"55d5e710-5432-48a4-b0ac-a865f823aae9\") " pod="openstack/dnsmasq-dns-5b75489c6f-btb4g" Dec 06 04:03:02 crc kubenswrapper[4802]: I1206 04:03:02.223802 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvfmw\" (UniqueName: \"kubernetes.io/projected/55d5e710-5432-48a4-b0ac-a865f823aae9-kube-api-access-tvfmw\") pod \"dnsmasq-dns-5b75489c6f-btb4g\" (UID: \"55d5e710-5432-48a4-b0ac-a865f823aae9\") " pod="openstack/dnsmasq-dns-5b75489c6f-btb4g" Dec 06 04:03:02 crc kubenswrapper[4802]: I1206 04:03:02.223850 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/55d5e710-5432-48a4-b0ac-a865f823aae9-ovsdbserver-sb\") pod \"dnsmasq-dns-5b75489c6f-btb4g\" (UID: \"55d5e710-5432-48a4-b0ac-a865f823aae9\") " pod="openstack/dnsmasq-dns-5b75489c6f-btb4g" Dec 06 04:03:02 crc kubenswrapper[4802]: I1206 04:03:02.223901 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/55d5e710-5432-48a4-b0ac-a865f823aae9-dns-swift-storage-0\") pod \"dnsmasq-dns-5b75489c6f-btb4g\" (UID: \"55d5e710-5432-48a4-b0ac-a865f823aae9\") " pod="openstack/dnsmasq-dns-5b75489c6f-btb4g" Dec 06 04:03:02 crc kubenswrapper[4802]: I1206 04:03:02.224618 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/55d5e710-5432-48a4-b0ac-a865f823aae9-dns-svc\") pod \"dnsmasq-dns-5b75489c6f-btb4g\" (UID: \"55d5e710-5432-48a4-b0ac-a865f823aae9\") " pod="openstack/dnsmasq-dns-5b75489c6f-btb4g" Dec 06 04:03:02 crc kubenswrapper[4802]: I1206 04:03:02.224968 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/55d5e710-5432-48a4-b0ac-a865f823aae9-openstack-edpm-ipam\") pod \"dnsmasq-dns-5b75489c6f-btb4g\" (UID: \"55d5e710-5432-48a4-b0ac-a865f823aae9\") " pod="openstack/dnsmasq-dns-5b75489c6f-btb4g" Dec 06 04:03:02 crc kubenswrapper[4802]: I1206 04:03:02.225126 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55d5e710-5432-48a4-b0ac-a865f823aae9-config\") pod \"dnsmasq-dns-5b75489c6f-btb4g\" (UID: \"55d5e710-5432-48a4-b0ac-a865f823aae9\") " pod="openstack/dnsmasq-dns-5b75489c6f-btb4g" Dec 06 04:03:02 crc kubenswrapper[4802]: I1206 04:03:02.225242 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/55d5e710-5432-48a4-b0ac-a865f823aae9-ovsdbserver-sb\") pod \"dnsmasq-dns-5b75489c6f-btb4g\" (UID: \"55d5e710-5432-48a4-b0ac-a865f823aae9\") " pod="openstack/dnsmasq-dns-5b75489c6f-btb4g" Dec 06 04:03:02 crc kubenswrapper[4802]: I1206 04:03:02.225427 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/55d5e710-5432-48a4-b0ac-a865f823aae9-dns-swift-storage-0\") pod \"dnsmasq-dns-5b75489c6f-btb4g\" (UID: \"55d5e710-5432-48a4-b0ac-a865f823aae9\") " pod="openstack/dnsmasq-dns-5b75489c6f-btb4g" Dec 06 04:03:02 crc kubenswrapper[4802]: I1206 04:03:02.225767 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/55d5e710-5432-48a4-b0ac-a865f823aae9-ovsdbserver-nb\") pod \"dnsmasq-dns-5b75489c6f-btb4g\" (UID: \"55d5e710-5432-48a4-b0ac-a865f823aae9\") " pod="openstack/dnsmasq-dns-5b75489c6f-btb4g" Dec 06 04:03:02 crc kubenswrapper[4802]: I1206 04:03:02.253720 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvfmw\" (UniqueName: \"kubernetes.io/projected/55d5e710-5432-48a4-b0ac-a865f823aae9-kube-api-access-tvfmw\") pod \"dnsmasq-dns-5b75489c6f-btb4g\" (UID: \"55d5e710-5432-48a4-b0ac-a865f823aae9\") " pod="openstack/dnsmasq-dns-5b75489c6f-btb4g" Dec 06 04:03:02 crc kubenswrapper[4802]: I1206 04:03:02.368440 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b75489c6f-btb4g" Dec 06 04:03:03 crc kubenswrapper[4802]: I1206 04:03:03.166369 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"cdcf8137-9bbe-467f-bbb3-e30f350760a6","Type":"ContainerStarted","Data":"c545f42c645714e766b7fbdf96d6f9640ad5674d757b708e42efa28409b54a2f"} Dec 06 04:03:03 crc kubenswrapper[4802]: I1206 04:03:03.168089 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"cdc37fb9-aeae-4fe0-b86e-e9d41e48314b","Type":"ContainerStarted","Data":"d4d728141f248f1490ca02ca5946c044789016b61141bf7f54a0361a42bdabc1"} Dec 06 04:03:04 crc kubenswrapper[4802]: I1206 04:03:04.217910 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b75489c6f-btb4g"] Dec 06 04:03:05 crc kubenswrapper[4802]: I1206 04:03:05.191641 4802 generic.go:334] "Generic (PLEG): container finished" podID="55d5e710-5432-48a4-b0ac-a865f823aae9" containerID="80ea7b9c201fc2ea83d842b152b891c82abd46609ef9d2f8a4fba6037028a9a6" exitCode=0 Dec 06 04:03:05 crc kubenswrapper[4802]: I1206 04:03:05.191743 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b75489c6f-btb4g" event={"ID":"55d5e710-5432-48a4-b0ac-a865f823aae9","Type":"ContainerDied","Data":"80ea7b9c201fc2ea83d842b152b891c82abd46609ef9d2f8a4fba6037028a9a6"} Dec 06 04:03:05 crc kubenswrapper[4802]: I1206 04:03:05.192026 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b75489c6f-btb4g" event={"ID":"55d5e710-5432-48a4-b0ac-a865f823aae9","Type":"ContainerStarted","Data":"67dd379b6525a695691d3b15a9077c6829267853326587caaa36e51c2a5c57f1"} Dec 06 04:03:06 crc kubenswrapper[4802]: I1206 04:03:06.221302 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b75489c6f-btb4g" event={"ID":"55d5e710-5432-48a4-b0ac-a865f823aae9","Type":"ContainerStarted","Data":"dcbb966c79b7cd8abcd4594881d4f3e90e758a68637c9c14a701c02f6e5301e0"} Dec 06 04:03:06 crc kubenswrapper[4802]: I1206 04:03:06.221874 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b75489c6f-btb4g" Dec 06 04:03:06 crc kubenswrapper[4802]: I1206 04:03:06.247369 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5b75489c6f-btb4g" podStartSLOduration=4.247348845 podStartE2EDuration="4.247348845s" podCreationTimestamp="2025-12-06 04:03:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:03:06.241442024 +0000 UTC m=+1379.113351186" watchObservedRunningTime="2025-12-06 04:03:06.247348845 +0000 UTC m=+1379.119257997" Dec 06 04:03:12 crc kubenswrapper[4802]: I1206 04:03:12.370072 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5b75489c6f-btb4g" Dec 06 04:03:12 crc kubenswrapper[4802]: I1206 04:03:12.450993 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f84f9ccf-fc42d"] Dec 06 04:03:12 crc kubenswrapper[4802]: I1206 04:03:12.451272 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-f84f9ccf-fc42d" podUID="d01257b6-a8d4-41e1-b0a0-894cba15e527" containerName="dnsmasq-dns" containerID="cri-o://3b8fd0883832b00c8e9451fb374667099110f5304d39b4dcaeac3002f8ef78c9" gracePeriod=10 Dec 06 04:03:12 crc kubenswrapper[4802]: I1206 04:03:12.636768 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5d75f767dc-75bb8"] Dec 06 04:03:12 crc kubenswrapper[4802]: I1206 04:03:12.648031 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d75f767dc-75bb8" Dec 06 04:03:12 crc kubenswrapper[4802]: I1206 04:03:12.648392 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d75f767dc-75bb8"] Dec 06 04:03:12 crc kubenswrapper[4802]: I1206 04:03:12.761008 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvd6x\" (UniqueName: \"kubernetes.io/projected/ad17cfaf-5662-41ae-9987-766441edb7a4-kube-api-access-lvd6x\") pod \"dnsmasq-dns-5d75f767dc-75bb8\" (UID: \"ad17cfaf-5662-41ae-9987-766441edb7a4\") " pod="openstack/dnsmasq-dns-5d75f767dc-75bb8" Dec 06 04:03:12 crc kubenswrapper[4802]: I1206 04:03:12.761054 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad17cfaf-5662-41ae-9987-766441edb7a4-config\") pod \"dnsmasq-dns-5d75f767dc-75bb8\" (UID: \"ad17cfaf-5662-41ae-9987-766441edb7a4\") " pod="openstack/dnsmasq-dns-5d75f767dc-75bb8" Dec 06 04:03:12 crc kubenswrapper[4802]: I1206 04:03:12.761081 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ad17cfaf-5662-41ae-9987-766441edb7a4-ovsdbserver-sb\") pod \"dnsmasq-dns-5d75f767dc-75bb8\" (UID: \"ad17cfaf-5662-41ae-9987-766441edb7a4\") " pod="openstack/dnsmasq-dns-5d75f767dc-75bb8" Dec 06 04:03:12 crc kubenswrapper[4802]: I1206 04:03:12.761112 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ad17cfaf-5662-41ae-9987-766441edb7a4-openstack-edpm-ipam\") pod \"dnsmasq-dns-5d75f767dc-75bb8\" (UID: \"ad17cfaf-5662-41ae-9987-766441edb7a4\") " pod="openstack/dnsmasq-dns-5d75f767dc-75bb8" Dec 06 04:03:12 crc kubenswrapper[4802]: I1206 04:03:12.761143 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ad17cfaf-5662-41ae-9987-766441edb7a4-dns-swift-storage-0\") pod \"dnsmasq-dns-5d75f767dc-75bb8\" (UID: \"ad17cfaf-5662-41ae-9987-766441edb7a4\") " pod="openstack/dnsmasq-dns-5d75f767dc-75bb8" Dec 06 04:03:12 crc kubenswrapper[4802]: I1206 04:03:12.761219 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ad17cfaf-5662-41ae-9987-766441edb7a4-dns-svc\") pod \"dnsmasq-dns-5d75f767dc-75bb8\" (UID: \"ad17cfaf-5662-41ae-9987-766441edb7a4\") " pod="openstack/dnsmasq-dns-5d75f767dc-75bb8" Dec 06 04:03:12 crc kubenswrapper[4802]: I1206 04:03:12.761348 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ad17cfaf-5662-41ae-9987-766441edb7a4-ovsdbserver-nb\") pod \"dnsmasq-dns-5d75f767dc-75bb8\" (UID: \"ad17cfaf-5662-41ae-9987-766441edb7a4\") " pod="openstack/dnsmasq-dns-5d75f767dc-75bb8" Dec 06 04:03:12 crc kubenswrapper[4802]: I1206 04:03:12.863325 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ad17cfaf-5662-41ae-9987-766441edb7a4-ovsdbserver-nb\") pod \"dnsmasq-dns-5d75f767dc-75bb8\" (UID: \"ad17cfaf-5662-41ae-9987-766441edb7a4\") " pod="openstack/dnsmasq-dns-5d75f767dc-75bb8" Dec 06 04:03:12 crc kubenswrapper[4802]: I1206 04:03:12.863601 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvd6x\" (UniqueName: \"kubernetes.io/projected/ad17cfaf-5662-41ae-9987-766441edb7a4-kube-api-access-lvd6x\") pod \"dnsmasq-dns-5d75f767dc-75bb8\" (UID: \"ad17cfaf-5662-41ae-9987-766441edb7a4\") " pod="openstack/dnsmasq-dns-5d75f767dc-75bb8" Dec 06 04:03:12 crc kubenswrapper[4802]: I1206 04:03:12.863619 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad17cfaf-5662-41ae-9987-766441edb7a4-config\") pod \"dnsmasq-dns-5d75f767dc-75bb8\" (UID: \"ad17cfaf-5662-41ae-9987-766441edb7a4\") " pod="openstack/dnsmasq-dns-5d75f767dc-75bb8" Dec 06 04:03:12 crc kubenswrapper[4802]: I1206 04:03:12.863636 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ad17cfaf-5662-41ae-9987-766441edb7a4-ovsdbserver-sb\") pod \"dnsmasq-dns-5d75f767dc-75bb8\" (UID: \"ad17cfaf-5662-41ae-9987-766441edb7a4\") " pod="openstack/dnsmasq-dns-5d75f767dc-75bb8" Dec 06 04:03:12 crc kubenswrapper[4802]: I1206 04:03:12.863653 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ad17cfaf-5662-41ae-9987-766441edb7a4-openstack-edpm-ipam\") pod \"dnsmasq-dns-5d75f767dc-75bb8\" (UID: \"ad17cfaf-5662-41ae-9987-766441edb7a4\") " pod="openstack/dnsmasq-dns-5d75f767dc-75bb8" Dec 06 04:03:12 crc kubenswrapper[4802]: I1206 04:03:12.863676 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ad17cfaf-5662-41ae-9987-766441edb7a4-dns-swift-storage-0\") pod \"dnsmasq-dns-5d75f767dc-75bb8\" (UID: \"ad17cfaf-5662-41ae-9987-766441edb7a4\") " pod="openstack/dnsmasq-dns-5d75f767dc-75bb8" Dec 06 04:03:12 crc kubenswrapper[4802]: I1206 04:03:12.863724 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ad17cfaf-5662-41ae-9987-766441edb7a4-dns-svc\") pod \"dnsmasq-dns-5d75f767dc-75bb8\" (UID: \"ad17cfaf-5662-41ae-9987-766441edb7a4\") " pod="openstack/dnsmasq-dns-5d75f767dc-75bb8" Dec 06 04:03:12 crc kubenswrapper[4802]: I1206 04:03:12.864562 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ad17cfaf-5662-41ae-9987-766441edb7a4-dns-svc\") pod \"dnsmasq-dns-5d75f767dc-75bb8\" (UID: \"ad17cfaf-5662-41ae-9987-766441edb7a4\") " pod="openstack/dnsmasq-dns-5d75f767dc-75bb8" Dec 06 04:03:12 crc kubenswrapper[4802]: I1206 04:03:12.865226 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ad17cfaf-5662-41ae-9987-766441edb7a4-ovsdbserver-nb\") pod \"dnsmasq-dns-5d75f767dc-75bb8\" (UID: \"ad17cfaf-5662-41ae-9987-766441edb7a4\") " pod="openstack/dnsmasq-dns-5d75f767dc-75bb8" Dec 06 04:03:12 crc kubenswrapper[4802]: I1206 04:03:12.866258 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad17cfaf-5662-41ae-9987-766441edb7a4-config\") pod \"dnsmasq-dns-5d75f767dc-75bb8\" (UID: \"ad17cfaf-5662-41ae-9987-766441edb7a4\") " pod="openstack/dnsmasq-dns-5d75f767dc-75bb8" Dec 06 04:03:12 crc kubenswrapper[4802]: I1206 04:03:12.867024 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ad17cfaf-5662-41ae-9987-766441edb7a4-openstack-edpm-ipam\") pod \"dnsmasq-dns-5d75f767dc-75bb8\" (UID: \"ad17cfaf-5662-41ae-9987-766441edb7a4\") " pod="openstack/dnsmasq-dns-5d75f767dc-75bb8" Dec 06 04:03:12 crc kubenswrapper[4802]: I1206 04:03:12.867540 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ad17cfaf-5662-41ae-9987-766441edb7a4-dns-swift-storage-0\") pod \"dnsmasq-dns-5d75f767dc-75bb8\" (UID: \"ad17cfaf-5662-41ae-9987-766441edb7a4\") " pod="openstack/dnsmasq-dns-5d75f767dc-75bb8" Dec 06 04:03:12 crc kubenswrapper[4802]: I1206 04:03:12.868546 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ad17cfaf-5662-41ae-9987-766441edb7a4-ovsdbserver-sb\") pod \"dnsmasq-dns-5d75f767dc-75bb8\" (UID: \"ad17cfaf-5662-41ae-9987-766441edb7a4\") " pod="openstack/dnsmasq-dns-5d75f767dc-75bb8" Dec 06 04:03:12 crc kubenswrapper[4802]: I1206 04:03:12.883584 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvd6x\" (UniqueName: \"kubernetes.io/projected/ad17cfaf-5662-41ae-9987-766441edb7a4-kube-api-access-lvd6x\") pod \"dnsmasq-dns-5d75f767dc-75bb8\" (UID: \"ad17cfaf-5662-41ae-9987-766441edb7a4\") " pod="openstack/dnsmasq-dns-5d75f767dc-75bb8" Dec 06 04:03:12 crc kubenswrapper[4802]: I1206 04:03:12.953292 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f84f9ccf-fc42d" Dec 06 04:03:12 crc kubenswrapper[4802]: I1206 04:03:12.968230 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d75f767dc-75bb8" Dec 06 04:03:13 crc kubenswrapper[4802]: I1206 04:03:13.066674 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d01257b6-a8d4-41e1-b0a0-894cba15e527-ovsdbserver-sb\") pod \"d01257b6-a8d4-41e1-b0a0-894cba15e527\" (UID: \"d01257b6-a8d4-41e1-b0a0-894cba15e527\") " Dec 06 04:03:13 crc kubenswrapper[4802]: I1206 04:03:13.066806 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d01257b6-a8d4-41e1-b0a0-894cba15e527-dns-svc\") pod \"d01257b6-a8d4-41e1-b0a0-894cba15e527\" (UID: \"d01257b6-a8d4-41e1-b0a0-894cba15e527\") " Dec 06 04:03:13 crc kubenswrapper[4802]: I1206 04:03:13.066888 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d01257b6-a8d4-41e1-b0a0-894cba15e527-config\") pod \"d01257b6-a8d4-41e1-b0a0-894cba15e527\" (UID: \"d01257b6-a8d4-41e1-b0a0-894cba15e527\") " Dec 06 04:03:13 crc kubenswrapper[4802]: I1206 04:03:13.066957 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jqzk7\" (UniqueName: \"kubernetes.io/projected/d01257b6-a8d4-41e1-b0a0-894cba15e527-kube-api-access-jqzk7\") pod \"d01257b6-a8d4-41e1-b0a0-894cba15e527\" (UID: \"d01257b6-a8d4-41e1-b0a0-894cba15e527\") " Dec 06 04:03:13 crc kubenswrapper[4802]: I1206 04:03:13.067022 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d01257b6-a8d4-41e1-b0a0-894cba15e527-ovsdbserver-nb\") pod \"d01257b6-a8d4-41e1-b0a0-894cba15e527\" (UID: \"d01257b6-a8d4-41e1-b0a0-894cba15e527\") " Dec 06 04:03:13 crc kubenswrapper[4802]: I1206 04:03:13.067169 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d01257b6-a8d4-41e1-b0a0-894cba15e527-dns-swift-storage-0\") pod \"d01257b6-a8d4-41e1-b0a0-894cba15e527\" (UID: \"d01257b6-a8d4-41e1-b0a0-894cba15e527\") " Dec 06 04:03:13 crc kubenswrapper[4802]: I1206 04:03:13.073941 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d01257b6-a8d4-41e1-b0a0-894cba15e527-kube-api-access-jqzk7" (OuterVolumeSpecName: "kube-api-access-jqzk7") pod "d01257b6-a8d4-41e1-b0a0-894cba15e527" (UID: "d01257b6-a8d4-41e1-b0a0-894cba15e527"). InnerVolumeSpecName "kube-api-access-jqzk7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:03:13 crc kubenswrapper[4802]: I1206 04:03:13.124185 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d01257b6-a8d4-41e1-b0a0-894cba15e527-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d01257b6-a8d4-41e1-b0a0-894cba15e527" (UID: "d01257b6-a8d4-41e1-b0a0-894cba15e527"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:03:13 crc kubenswrapper[4802]: I1206 04:03:13.137981 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d01257b6-a8d4-41e1-b0a0-894cba15e527-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d01257b6-a8d4-41e1-b0a0-894cba15e527" (UID: "d01257b6-a8d4-41e1-b0a0-894cba15e527"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:03:13 crc kubenswrapper[4802]: I1206 04:03:13.142072 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d01257b6-a8d4-41e1-b0a0-894cba15e527-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d01257b6-a8d4-41e1-b0a0-894cba15e527" (UID: "d01257b6-a8d4-41e1-b0a0-894cba15e527"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:03:13 crc kubenswrapper[4802]: I1206 04:03:13.149438 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d01257b6-a8d4-41e1-b0a0-894cba15e527-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d01257b6-a8d4-41e1-b0a0-894cba15e527" (UID: "d01257b6-a8d4-41e1-b0a0-894cba15e527"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:03:13 crc kubenswrapper[4802]: I1206 04:03:13.158763 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d01257b6-a8d4-41e1-b0a0-894cba15e527-config" (OuterVolumeSpecName: "config") pod "d01257b6-a8d4-41e1-b0a0-894cba15e527" (UID: "d01257b6-a8d4-41e1-b0a0-894cba15e527"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:03:13 crc kubenswrapper[4802]: I1206 04:03:13.170504 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jqzk7\" (UniqueName: \"kubernetes.io/projected/d01257b6-a8d4-41e1-b0a0-894cba15e527-kube-api-access-jqzk7\") on node \"crc\" DevicePath \"\"" Dec 06 04:03:13 crc kubenswrapper[4802]: I1206 04:03:13.170526 4802 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d01257b6-a8d4-41e1-b0a0-894cba15e527-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 04:03:13 crc kubenswrapper[4802]: I1206 04:03:13.170535 4802 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d01257b6-a8d4-41e1-b0a0-894cba15e527-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 04:03:13 crc kubenswrapper[4802]: I1206 04:03:13.170543 4802 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d01257b6-a8d4-41e1-b0a0-894cba15e527-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 04:03:13 crc kubenswrapper[4802]: I1206 04:03:13.170551 4802 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d01257b6-a8d4-41e1-b0a0-894cba15e527-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 04:03:13 crc kubenswrapper[4802]: I1206 04:03:13.170559 4802 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d01257b6-a8d4-41e1-b0a0-894cba15e527-config\") on node \"crc\" DevicePath \"\"" Dec 06 04:03:13 crc kubenswrapper[4802]: I1206 04:03:13.298420 4802 generic.go:334] "Generic (PLEG): container finished" podID="d01257b6-a8d4-41e1-b0a0-894cba15e527" containerID="3b8fd0883832b00c8e9451fb374667099110f5304d39b4dcaeac3002f8ef78c9" exitCode=0 Dec 06 04:03:13 crc kubenswrapper[4802]: I1206 04:03:13.298478 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f84f9ccf-fc42d" event={"ID":"d01257b6-a8d4-41e1-b0a0-894cba15e527","Type":"ContainerDied","Data":"3b8fd0883832b00c8e9451fb374667099110f5304d39b4dcaeac3002f8ef78c9"} Dec 06 04:03:13 crc kubenswrapper[4802]: I1206 04:03:13.298527 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f84f9ccf-fc42d" event={"ID":"d01257b6-a8d4-41e1-b0a0-894cba15e527","Type":"ContainerDied","Data":"2640aee878c0a7c70a9250585c44ea065e7ca7fdb683890442847ce021c6de29"} Dec 06 04:03:13 crc kubenswrapper[4802]: I1206 04:03:13.298544 4802 scope.go:117] "RemoveContainer" containerID="3b8fd0883832b00c8e9451fb374667099110f5304d39b4dcaeac3002f8ef78c9" Dec 06 04:03:13 crc kubenswrapper[4802]: I1206 04:03:13.298803 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f84f9ccf-fc42d" Dec 06 04:03:13 crc kubenswrapper[4802]: I1206 04:03:13.334008 4802 scope.go:117] "RemoveContainer" containerID="e407afae41d5750349a249f62b1c86ddc5f47cb7d5a7e26a5747658c452b8613" Dec 06 04:03:13 crc kubenswrapper[4802]: I1206 04:03:13.339280 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f84f9ccf-fc42d"] Dec 06 04:03:13 crc kubenswrapper[4802]: I1206 04:03:13.354501 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-f84f9ccf-fc42d"] Dec 06 04:03:13 crc kubenswrapper[4802]: I1206 04:03:13.368339 4802 scope.go:117] "RemoveContainer" containerID="3b8fd0883832b00c8e9451fb374667099110f5304d39b4dcaeac3002f8ef78c9" Dec 06 04:03:13 crc kubenswrapper[4802]: E1206 04:03:13.368744 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b8fd0883832b00c8e9451fb374667099110f5304d39b4dcaeac3002f8ef78c9\": container with ID starting with 3b8fd0883832b00c8e9451fb374667099110f5304d39b4dcaeac3002f8ef78c9 not found: ID does not exist" containerID="3b8fd0883832b00c8e9451fb374667099110f5304d39b4dcaeac3002f8ef78c9" Dec 06 04:03:13 crc kubenswrapper[4802]: I1206 04:03:13.368853 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b8fd0883832b00c8e9451fb374667099110f5304d39b4dcaeac3002f8ef78c9"} err="failed to get container status \"3b8fd0883832b00c8e9451fb374667099110f5304d39b4dcaeac3002f8ef78c9\": rpc error: code = NotFound desc = could not find container \"3b8fd0883832b00c8e9451fb374667099110f5304d39b4dcaeac3002f8ef78c9\": container with ID starting with 3b8fd0883832b00c8e9451fb374667099110f5304d39b4dcaeac3002f8ef78c9 not found: ID does not exist" Dec 06 04:03:13 crc kubenswrapper[4802]: I1206 04:03:13.368919 4802 scope.go:117] "RemoveContainer" containerID="e407afae41d5750349a249f62b1c86ddc5f47cb7d5a7e26a5747658c452b8613" Dec 06 04:03:13 crc kubenswrapper[4802]: E1206 04:03:13.371270 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e407afae41d5750349a249f62b1c86ddc5f47cb7d5a7e26a5747658c452b8613\": container with ID starting with e407afae41d5750349a249f62b1c86ddc5f47cb7d5a7e26a5747658c452b8613 not found: ID does not exist" containerID="e407afae41d5750349a249f62b1c86ddc5f47cb7d5a7e26a5747658c452b8613" Dec 06 04:03:13 crc kubenswrapper[4802]: I1206 04:03:13.371321 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e407afae41d5750349a249f62b1c86ddc5f47cb7d5a7e26a5747658c452b8613"} err="failed to get container status \"e407afae41d5750349a249f62b1c86ddc5f47cb7d5a7e26a5747658c452b8613\": rpc error: code = NotFound desc = could not find container \"e407afae41d5750349a249f62b1c86ddc5f47cb7d5a7e26a5747658c452b8613\": container with ID starting with e407afae41d5750349a249f62b1c86ddc5f47cb7d5a7e26a5747658c452b8613 not found: ID does not exist" Dec 06 04:03:13 crc kubenswrapper[4802]: E1206 04:03:13.376064 4802 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd01257b6_a8d4_41e1_b0a0_894cba15e527.slice/crio-2640aee878c0a7c70a9250585c44ea065e7ca7fdb683890442847ce021c6de29\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd01257b6_a8d4_41e1_b0a0_894cba15e527.slice\": RecentStats: unable to find data in memory cache]" Dec 06 04:03:13 crc kubenswrapper[4802]: I1206 04:03:13.439025 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d75f767dc-75bb8"] Dec 06 04:03:13 crc kubenswrapper[4802]: I1206 04:03:13.460693 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d01257b6-a8d4-41e1-b0a0-894cba15e527" path="/var/lib/kubelet/pods/d01257b6-a8d4-41e1-b0a0-894cba15e527/volumes" Dec 06 04:03:14 crc kubenswrapper[4802]: I1206 04:03:14.313347 4802 generic.go:334] "Generic (PLEG): container finished" podID="ad17cfaf-5662-41ae-9987-766441edb7a4" containerID="20b58346cbd432b5027447f6a75ae41a7331f6f2d8ce7bbf844e0fb7ccaa5d0e" exitCode=0 Dec 06 04:03:14 crc kubenswrapper[4802]: I1206 04:03:14.313408 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d75f767dc-75bb8" event={"ID":"ad17cfaf-5662-41ae-9987-766441edb7a4","Type":"ContainerDied","Data":"20b58346cbd432b5027447f6a75ae41a7331f6f2d8ce7bbf844e0fb7ccaa5d0e"} Dec 06 04:03:14 crc kubenswrapper[4802]: I1206 04:03:14.313699 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d75f767dc-75bb8" event={"ID":"ad17cfaf-5662-41ae-9987-766441edb7a4","Type":"ContainerStarted","Data":"c97a30f9a3ccb3922809fc57506ebb6c4c3fddaa8b6ef5b145baa94e49cd54b2"} Dec 06 04:03:15 crc kubenswrapper[4802]: I1206 04:03:15.327526 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d75f767dc-75bb8" event={"ID":"ad17cfaf-5662-41ae-9987-766441edb7a4","Type":"ContainerStarted","Data":"b18b083735b9c881163401587f7c808740f4a319270206782f0f50dd9ea6bab1"} Dec 06 04:03:15 crc kubenswrapper[4802]: I1206 04:03:15.327983 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5d75f767dc-75bb8" Dec 06 04:03:15 crc kubenswrapper[4802]: I1206 04:03:15.356205 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5d75f767dc-75bb8" podStartSLOduration=3.356185544 podStartE2EDuration="3.356185544s" podCreationTimestamp="2025-12-06 04:03:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:03:15.349865721 +0000 UTC m=+1388.221774873" watchObservedRunningTime="2025-12-06 04:03:15.356185544 +0000 UTC m=+1388.228094696" Dec 06 04:03:17 crc kubenswrapper[4802]: I1206 04:03:17.719331 4802 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-f84f9ccf-fc42d" podUID="d01257b6-a8d4-41e1-b0a0-894cba15e527" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.207:5353: i/o timeout" Dec 06 04:03:22 crc kubenswrapper[4802]: I1206 04:03:22.970038 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5d75f767dc-75bb8" Dec 06 04:03:23 crc kubenswrapper[4802]: I1206 04:03:23.051219 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b75489c6f-btb4g"] Dec 06 04:03:23 crc kubenswrapper[4802]: I1206 04:03:23.051511 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5b75489c6f-btb4g" podUID="55d5e710-5432-48a4-b0ac-a865f823aae9" containerName="dnsmasq-dns" containerID="cri-o://dcbb966c79b7cd8abcd4594881d4f3e90e758a68637c9c14a701c02f6e5301e0" gracePeriod=10 Dec 06 04:03:23 crc kubenswrapper[4802]: I1206 04:03:23.426335 4802 generic.go:334] "Generic (PLEG): container finished" podID="55d5e710-5432-48a4-b0ac-a865f823aae9" containerID="dcbb966c79b7cd8abcd4594881d4f3e90e758a68637c9c14a701c02f6e5301e0" exitCode=0 Dec 06 04:03:23 crc kubenswrapper[4802]: I1206 04:03:23.426648 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b75489c6f-btb4g" event={"ID":"55d5e710-5432-48a4-b0ac-a865f823aae9","Type":"ContainerDied","Data":"dcbb966c79b7cd8abcd4594881d4f3e90e758a68637c9c14a701c02f6e5301e0"} Dec 06 04:03:23 crc kubenswrapper[4802]: I1206 04:03:23.578219 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b75489c6f-btb4g" Dec 06 04:03:23 crc kubenswrapper[4802]: I1206 04:03:23.681364 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tvfmw\" (UniqueName: \"kubernetes.io/projected/55d5e710-5432-48a4-b0ac-a865f823aae9-kube-api-access-tvfmw\") pod \"55d5e710-5432-48a4-b0ac-a865f823aae9\" (UID: \"55d5e710-5432-48a4-b0ac-a865f823aae9\") " Dec 06 04:03:23 crc kubenswrapper[4802]: I1206 04:03:23.681406 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/55d5e710-5432-48a4-b0ac-a865f823aae9-ovsdbserver-nb\") pod \"55d5e710-5432-48a4-b0ac-a865f823aae9\" (UID: \"55d5e710-5432-48a4-b0ac-a865f823aae9\") " Dec 06 04:03:23 crc kubenswrapper[4802]: I1206 04:03:23.681457 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/55d5e710-5432-48a4-b0ac-a865f823aae9-dns-swift-storage-0\") pod \"55d5e710-5432-48a4-b0ac-a865f823aae9\" (UID: \"55d5e710-5432-48a4-b0ac-a865f823aae9\") " Dec 06 04:03:23 crc kubenswrapper[4802]: I1206 04:03:23.681484 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55d5e710-5432-48a4-b0ac-a865f823aae9-config\") pod \"55d5e710-5432-48a4-b0ac-a865f823aae9\" (UID: \"55d5e710-5432-48a4-b0ac-a865f823aae9\") " Dec 06 04:03:23 crc kubenswrapper[4802]: I1206 04:03:23.681522 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/55d5e710-5432-48a4-b0ac-a865f823aae9-openstack-edpm-ipam\") pod \"55d5e710-5432-48a4-b0ac-a865f823aae9\" (UID: \"55d5e710-5432-48a4-b0ac-a865f823aae9\") " Dec 06 04:03:23 crc kubenswrapper[4802]: I1206 04:03:23.681593 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/55d5e710-5432-48a4-b0ac-a865f823aae9-ovsdbserver-sb\") pod \"55d5e710-5432-48a4-b0ac-a865f823aae9\" (UID: \"55d5e710-5432-48a4-b0ac-a865f823aae9\") " Dec 06 04:03:23 crc kubenswrapper[4802]: I1206 04:03:23.681657 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/55d5e710-5432-48a4-b0ac-a865f823aae9-dns-svc\") pod \"55d5e710-5432-48a4-b0ac-a865f823aae9\" (UID: \"55d5e710-5432-48a4-b0ac-a865f823aae9\") " Dec 06 04:03:23 crc kubenswrapper[4802]: I1206 04:03:23.688640 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55d5e710-5432-48a4-b0ac-a865f823aae9-kube-api-access-tvfmw" (OuterVolumeSpecName: "kube-api-access-tvfmw") pod "55d5e710-5432-48a4-b0ac-a865f823aae9" (UID: "55d5e710-5432-48a4-b0ac-a865f823aae9"). InnerVolumeSpecName "kube-api-access-tvfmw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:03:23 crc kubenswrapper[4802]: I1206 04:03:23.729377 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55d5e710-5432-48a4-b0ac-a865f823aae9-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "55d5e710-5432-48a4-b0ac-a865f823aae9" (UID: "55d5e710-5432-48a4-b0ac-a865f823aae9"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:03:23 crc kubenswrapper[4802]: I1206 04:03:23.746363 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55d5e710-5432-48a4-b0ac-a865f823aae9-config" (OuterVolumeSpecName: "config") pod "55d5e710-5432-48a4-b0ac-a865f823aae9" (UID: "55d5e710-5432-48a4-b0ac-a865f823aae9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:03:23 crc kubenswrapper[4802]: I1206 04:03:23.749712 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55d5e710-5432-48a4-b0ac-a865f823aae9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "55d5e710-5432-48a4-b0ac-a865f823aae9" (UID: "55d5e710-5432-48a4-b0ac-a865f823aae9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:03:23 crc kubenswrapper[4802]: I1206 04:03:23.751393 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55d5e710-5432-48a4-b0ac-a865f823aae9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "55d5e710-5432-48a4-b0ac-a865f823aae9" (UID: "55d5e710-5432-48a4-b0ac-a865f823aae9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:03:23 crc kubenswrapper[4802]: I1206 04:03:23.753073 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55d5e710-5432-48a4-b0ac-a865f823aae9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "55d5e710-5432-48a4-b0ac-a865f823aae9" (UID: "55d5e710-5432-48a4-b0ac-a865f823aae9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:03:23 crc kubenswrapper[4802]: I1206 04:03:23.759928 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55d5e710-5432-48a4-b0ac-a865f823aae9-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "55d5e710-5432-48a4-b0ac-a865f823aae9" (UID: "55d5e710-5432-48a4-b0ac-a865f823aae9"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:03:23 crc kubenswrapper[4802]: I1206 04:03:23.783404 4802 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/55d5e710-5432-48a4-b0ac-a865f823aae9-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 04:03:23 crc kubenswrapper[4802]: I1206 04:03:23.783440 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tvfmw\" (UniqueName: \"kubernetes.io/projected/55d5e710-5432-48a4-b0ac-a865f823aae9-kube-api-access-tvfmw\") on node \"crc\" DevicePath \"\"" Dec 06 04:03:23 crc kubenswrapper[4802]: I1206 04:03:23.783452 4802 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/55d5e710-5432-48a4-b0ac-a865f823aae9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 04:03:23 crc kubenswrapper[4802]: I1206 04:03:23.783461 4802 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/55d5e710-5432-48a4-b0ac-a865f823aae9-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 04:03:23 crc kubenswrapper[4802]: I1206 04:03:23.783469 4802 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55d5e710-5432-48a4-b0ac-a865f823aae9-config\") on node \"crc\" DevicePath \"\"" Dec 06 04:03:23 crc kubenswrapper[4802]: I1206 04:03:23.783478 4802 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/55d5e710-5432-48a4-b0ac-a865f823aae9-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 06 04:03:23 crc kubenswrapper[4802]: I1206 04:03:23.783487 4802 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/55d5e710-5432-48a4-b0ac-a865f823aae9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 04:03:24 crc kubenswrapper[4802]: I1206 04:03:24.435813 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b75489c6f-btb4g" event={"ID":"55d5e710-5432-48a4-b0ac-a865f823aae9","Type":"ContainerDied","Data":"67dd379b6525a695691d3b15a9077c6829267853326587caaa36e51c2a5c57f1"} Dec 06 04:03:24 crc kubenswrapper[4802]: I1206 04:03:24.435864 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b75489c6f-btb4g" Dec 06 04:03:24 crc kubenswrapper[4802]: I1206 04:03:24.435869 4802 scope.go:117] "RemoveContainer" containerID="dcbb966c79b7cd8abcd4594881d4f3e90e758a68637c9c14a701c02f6e5301e0" Dec 06 04:03:24 crc kubenswrapper[4802]: I1206 04:03:24.472971 4802 scope.go:117] "RemoveContainer" containerID="80ea7b9c201fc2ea83d842b152b891c82abd46609ef9d2f8a4fba6037028a9a6" Dec 06 04:03:24 crc kubenswrapper[4802]: I1206 04:03:24.478976 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b75489c6f-btb4g"] Dec 06 04:03:24 crc kubenswrapper[4802]: I1206 04:03:24.487326 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b75489c6f-btb4g"] Dec 06 04:03:25 crc kubenswrapper[4802]: I1206 04:03:25.459861 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55d5e710-5432-48a4-b0ac-a865f823aae9" path="/var/lib/kubelet/pods/55d5e710-5432-48a4-b0ac-a865f823aae9/volumes" Dec 06 04:03:35 crc kubenswrapper[4802]: I1206 04:03:35.548032 4802 generic.go:334] "Generic (PLEG): container finished" podID="cdc37fb9-aeae-4fe0-b86e-e9d41e48314b" containerID="d4d728141f248f1490ca02ca5946c044789016b61141bf7f54a0361a42bdabc1" exitCode=0 Dec 06 04:03:35 crc kubenswrapper[4802]: I1206 04:03:35.548340 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"cdc37fb9-aeae-4fe0-b86e-e9d41e48314b","Type":"ContainerDied","Data":"d4d728141f248f1490ca02ca5946c044789016b61141bf7f54a0361a42bdabc1"} Dec 06 04:03:35 crc kubenswrapper[4802]: I1206 04:03:35.556720 4802 generic.go:334] "Generic (PLEG): container finished" podID="cdcf8137-9bbe-467f-bbb3-e30f350760a6" containerID="c545f42c645714e766b7fbdf96d6f9640ad5674d757b708e42efa28409b54a2f" exitCode=0 Dec 06 04:03:35 crc kubenswrapper[4802]: I1206 04:03:35.556785 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"cdcf8137-9bbe-467f-bbb3-e30f350760a6","Type":"ContainerDied","Data":"c545f42c645714e766b7fbdf96d6f9640ad5674d757b708e42efa28409b54a2f"} Dec 06 04:03:36 crc kubenswrapper[4802]: I1206 04:03:36.162251 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jrmv7"] Dec 06 04:03:36 crc kubenswrapper[4802]: E1206 04:03:36.163182 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d01257b6-a8d4-41e1-b0a0-894cba15e527" containerName="dnsmasq-dns" Dec 06 04:03:36 crc kubenswrapper[4802]: I1206 04:03:36.163247 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="d01257b6-a8d4-41e1-b0a0-894cba15e527" containerName="dnsmasq-dns" Dec 06 04:03:36 crc kubenswrapper[4802]: E1206 04:03:36.163315 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d01257b6-a8d4-41e1-b0a0-894cba15e527" containerName="init" Dec 06 04:03:36 crc kubenswrapper[4802]: I1206 04:03:36.163399 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="d01257b6-a8d4-41e1-b0a0-894cba15e527" containerName="init" Dec 06 04:03:36 crc kubenswrapper[4802]: E1206 04:03:36.163469 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55d5e710-5432-48a4-b0ac-a865f823aae9" containerName="dnsmasq-dns" Dec 06 04:03:36 crc kubenswrapper[4802]: I1206 04:03:36.163528 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="55d5e710-5432-48a4-b0ac-a865f823aae9" containerName="dnsmasq-dns" Dec 06 04:03:36 crc kubenswrapper[4802]: E1206 04:03:36.163599 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55d5e710-5432-48a4-b0ac-a865f823aae9" containerName="init" Dec 06 04:03:36 crc kubenswrapper[4802]: I1206 04:03:36.163649 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="55d5e710-5432-48a4-b0ac-a865f823aae9" containerName="init" Dec 06 04:03:36 crc kubenswrapper[4802]: I1206 04:03:36.163887 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="55d5e710-5432-48a4-b0ac-a865f823aae9" containerName="dnsmasq-dns" Dec 06 04:03:36 crc kubenswrapper[4802]: I1206 04:03:36.163953 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="d01257b6-a8d4-41e1-b0a0-894cba15e527" containerName="dnsmasq-dns" Dec 06 04:03:36 crc kubenswrapper[4802]: I1206 04:03:36.164628 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jrmv7" Dec 06 04:03:36 crc kubenswrapper[4802]: I1206 04:03:36.170085 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 04:03:36 crc kubenswrapper[4802]: I1206 04:03:36.170644 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-rgvjc" Dec 06 04:03:36 crc kubenswrapper[4802]: I1206 04:03:36.170894 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 04:03:36 crc kubenswrapper[4802]: I1206 04:03:36.171179 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 04:03:36 crc kubenswrapper[4802]: I1206 04:03:36.188856 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jrmv7"] Dec 06 04:03:36 crc kubenswrapper[4802]: I1206 04:03:36.263355 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/44aaebb1-a5a3-42e6-aacd-04292e80f33d-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jrmv7\" (UID: \"44aaebb1-a5a3-42e6-aacd-04292e80f33d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jrmv7" Dec 06 04:03:36 crc kubenswrapper[4802]: I1206 04:03:36.263507 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44aaebb1-a5a3-42e6-aacd-04292e80f33d-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jrmv7\" (UID: \"44aaebb1-a5a3-42e6-aacd-04292e80f33d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jrmv7" Dec 06 04:03:36 crc kubenswrapper[4802]: I1206 04:03:36.263542 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/44aaebb1-a5a3-42e6-aacd-04292e80f33d-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jrmv7\" (UID: \"44aaebb1-a5a3-42e6-aacd-04292e80f33d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jrmv7" Dec 06 04:03:36 crc kubenswrapper[4802]: I1206 04:03:36.263570 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2cqqx\" (UniqueName: \"kubernetes.io/projected/44aaebb1-a5a3-42e6-aacd-04292e80f33d-kube-api-access-2cqqx\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jrmv7\" (UID: \"44aaebb1-a5a3-42e6-aacd-04292e80f33d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jrmv7" Dec 06 04:03:36 crc kubenswrapper[4802]: I1206 04:03:36.364930 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44aaebb1-a5a3-42e6-aacd-04292e80f33d-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jrmv7\" (UID: \"44aaebb1-a5a3-42e6-aacd-04292e80f33d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jrmv7" Dec 06 04:03:36 crc kubenswrapper[4802]: I1206 04:03:36.364982 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/44aaebb1-a5a3-42e6-aacd-04292e80f33d-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jrmv7\" (UID: \"44aaebb1-a5a3-42e6-aacd-04292e80f33d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jrmv7" Dec 06 04:03:36 crc kubenswrapper[4802]: I1206 04:03:36.365004 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2cqqx\" (UniqueName: \"kubernetes.io/projected/44aaebb1-a5a3-42e6-aacd-04292e80f33d-kube-api-access-2cqqx\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jrmv7\" (UID: \"44aaebb1-a5a3-42e6-aacd-04292e80f33d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jrmv7" Dec 06 04:03:36 crc kubenswrapper[4802]: I1206 04:03:36.365097 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/44aaebb1-a5a3-42e6-aacd-04292e80f33d-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jrmv7\" (UID: \"44aaebb1-a5a3-42e6-aacd-04292e80f33d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jrmv7" Dec 06 04:03:36 crc kubenswrapper[4802]: I1206 04:03:36.370928 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/44aaebb1-a5a3-42e6-aacd-04292e80f33d-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jrmv7\" (UID: \"44aaebb1-a5a3-42e6-aacd-04292e80f33d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jrmv7" Dec 06 04:03:36 crc kubenswrapper[4802]: I1206 04:03:36.371378 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/44aaebb1-a5a3-42e6-aacd-04292e80f33d-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jrmv7\" (UID: \"44aaebb1-a5a3-42e6-aacd-04292e80f33d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jrmv7" Dec 06 04:03:36 crc kubenswrapper[4802]: I1206 04:03:36.384348 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44aaebb1-a5a3-42e6-aacd-04292e80f33d-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jrmv7\" (UID: \"44aaebb1-a5a3-42e6-aacd-04292e80f33d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jrmv7" Dec 06 04:03:36 crc kubenswrapper[4802]: I1206 04:03:36.397775 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2cqqx\" (UniqueName: \"kubernetes.io/projected/44aaebb1-a5a3-42e6-aacd-04292e80f33d-kube-api-access-2cqqx\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jrmv7\" (UID: \"44aaebb1-a5a3-42e6-aacd-04292e80f33d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jrmv7" Dec 06 04:03:36 crc kubenswrapper[4802]: I1206 04:03:36.489541 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jrmv7" Dec 06 04:03:36 crc kubenswrapper[4802]: I1206 04:03:36.572562 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"cdcf8137-9bbe-467f-bbb3-e30f350760a6","Type":"ContainerStarted","Data":"6664cc514e9a51f19dbec9f3f7934f923747197a2df4e8489f3dd8e208506d35"} Dec 06 04:03:36 crc kubenswrapper[4802]: I1206 04:03:36.574090 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 06 04:03:36 crc kubenswrapper[4802]: I1206 04:03:36.576264 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"cdc37fb9-aeae-4fe0-b86e-e9d41e48314b","Type":"ContainerStarted","Data":"50279ba3df354cd579506acf4064b918eac188c7d8b21ddc264bfd122416451a"} Dec 06 04:03:36 crc kubenswrapper[4802]: I1206 04:03:36.576772 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 06 04:03:36 crc kubenswrapper[4802]: I1206 04:03:36.639494 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.639474654 podStartE2EDuration="36.639474654s" podCreationTimestamp="2025-12-06 04:03:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:03:36.61217094 +0000 UTC m=+1409.484080102" watchObservedRunningTime="2025-12-06 04:03:36.639474654 +0000 UTC m=+1409.511383806" Dec 06 04:03:36 crc kubenswrapper[4802]: I1206 04:03:36.640961 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.640953095 podStartE2EDuration="37.640953095s" podCreationTimestamp="2025-12-06 04:02:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:03:36.637800569 +0000 UTC m=+1409.509709731" watchObservedRunningTime="2025-12-06 04:03:36.640953095 +0000 UTC m=+1409.512862247" Dec 06 04:03:37 crc kubenswrapper[4802]: I1206 04:03:37.112468 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jrmv7"] Dec 06 04:03:37 crc kubenswrapper[4802]: W1206 04:03:37.122906 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod44aaebb1_a5a3_42e6_aacd_04292e80f33d.slice/crio-8968b097f18b0527a0584a1eda16dfe8d50389bf421c23342cb736c7b9112c97 WatchSource:0}: Error finding container 8968b097f18b0527a0584a1eda16dfe8d50389bf421c23342cb736c7b9112c97: Status 404 returned error can't find the container with id 8968b097f18b0527a0584a1eda16dfe8d50389bf421c23342cb736c7b9112c97 Dec 06 04:03:37 crc kubenswrapper[4802]: I1206 04:03:37.584792 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jrmv7" event={"ID":"44aaebb1-a5a3-42e6-aacd-04292e80f33d","Type":"ContainerStarted","Data":"8968b097f18b0527a0584a1eda16dfe8d50389bf421c23342cb736c7b9112c97"} Dec 06 04:03:49 crc kubenswrapper[4802]: I1206 04:03:49.928120 4802 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="cdc37fb9-aeae-4fe0-b86e-e9d41e48314b" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.214:5671: connect: connection refused" Dec 06 04:03:50 crc kubenswrapper[4802]: I1206 04:03:50.828899 4802 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="cdcf8137-9bbe-467f-bbb3-e30f350760a6" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.215:5671: connect: connection refused" Dec 06 04:03:53 crc kubenswrapper[4802]: I1206 04:03:53.754280 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jrmv7" event={"ID":"44aaebb1-a5a3-42e6-aacd-04292e80f33d","Type":"ContainerStarted","Data":"0bc7052eac0ba466f3aace6c538930ee1556fb63b2a04a294b37ed7cfc687655"} Dec 06 04:03:53 crc kubenswrapper[4802]: I1206 04:03:53.784585 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jrmv7" podStartSLOduration=2.297994893 podStartE2EDuration="17.78456665s" podCreationTimestamp="2025-12-06 04:03:36 +0000 UTC" firstStartedPulling="2025-12-06 04:03:37.124558527 +0000 UTC m=+1409.996467699" lastFinishedPulling="2025-12-06 04:03:52.611130304 +0000 UTC m=+1425.483039456" observedRunningTime="2025-12-06 04:03:53.776674244 +0000 UTC m=+1426.648583416" watchObservedRunningTime="2025-12-06 04:03:53.78456665 +0000 UTC m=+1426.656475802" Dec 06 04:03:59 crc kubenswrapper[4802]: I1206 04:03:59.928164 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 06 04:04:00 crc kubenswrapper[4802]: I1206 04:04:00.826956 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 06 04:04:04 crc kubenswrapper[4802]: I1206 04:04:04.872688 4802 generic.go:334] "Generic (PLEG): container finished" podID="44aaebb1-a5a3-42e6-aacd-04292e80f33d" containerID="0bc7052eac0ba466f3aace6c538930ee1556fb63b2a04a294b37ed7cfc687655" exitCode=0 Dec 06 04:04:04 crc kubenswrapper[4802]: I1206 04:04:04.872845 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jrmv7" event={"ID":"44aaebb1-a5a3-42e6-aacd-04292e80f33d","Type":"ContainerDied","Data":"0bc7052eac0ba466f3aace6c538930ee1556fb63b2a04a294b37ed7cfc687655"} Dec 06 04:04:06 crc kubenswrapper[4802]: I1206 04:04:06.363911 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jrmv7" Dec 06 04:04:06 crc kubenswrapper[4802]: I1206 04:04:06.413124 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/44aaebb1-a5a3-42e6-aacd-04292e80f33d-ssh-key\") pod \"44aaebb1-a5a3-42e6-aacd-04292e80f33d\" (UID: \"44aaebb1-a5a3-42e6-aacd-04292e80f33d\") " Dec 06 04:04:06 crc kubenswrapper[4802]: I1206 04:04:06.413370 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44aaebb1-a5a3-42e6-aacd-04292e80f33d-repo-setup-combined-ca-bundle\") pod \"44aaebb1-a5a3-42e6-aacd-04292e80f33d\" (UID: \"44aaebb1-a5a3-42e6-aacd-04292e80f33d\") " Dec 06 04:04:06 crc kubenswrapper[4802]: I1206 04:04:06.413431 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2cqqx\" (UniqueName: \"kubernetes.io/projected/44aaebb1-a5a3-42e6-aacd-04292e80f33d-kube-api-access-2cqqx\") pod \"44aaebb1-a5a3-42e6-aacd-04292e80f33d\" (UID: \"44aaebb1-a5a3-42e6-aacd-04292e80f33d\") " Dec 06 04:04:06 crc kubenswrapper[4802]: I1206 04:04:06.413467 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/44aaebb1-a5a3-42e6-aacd-04292e80f33d-inventory\") pod \"44aaebb1-a5a3-42e6-aacd-04292e80f33d\" (UID: \"44aaebb1-a5a3-42e6-aacd-04292e80f33d\") " Dec 06 04:04:06 crc kubenswrapper[4802]: I1206 04:04:06.419041 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44aaebb1-a5a3-42e6-aacd-04292e80f33d-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "44aaebb1-a5a3-42e6-aacd-04292e80f33d" (UID: "44aaebb1-a5a3-42e6-aacd-04292e80f33d"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:04:06 crc kubenswrapper[4802]: I1206 04:04:06.419120 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44aaebb1-a5a3-42e6-aacd-04292e80f33d-kube-api-access-2cqqx" (OuterVolumeSpecName: "kube-api-access-2cqqx") pod "44aaebb1-a5a3-42e6-aacd-04292e80f33d" (UID: "44aaebb1-a5a3-42e6-aacd-04292e80f33d"). InnerVolumeSpecName "kube-api-access-2cqqx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:04:06 crc kubenswrapper[4802]: I1206 04:04:06.445035 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44aaebb1-a5a3-42e6-aacd-04292e80f33d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "44aaebb1-a5a3-42e6-aacd-04292e80f33d" (UID: "44aaebb1-a5a3-42e6-aacd-04292e80f33d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:04:06 crc kubenswrapper[4802]: I1206 04:04:06.446651 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44aaebb1-a5a3-42e6-aacd-04292e80f33d-inventory" (OuterVolumeSpecName: "inventory") pod "44aaebb1-a5a3-42e6-aacd-04292e80f33d" (UID: "44aaebb1-a5a3-42e6-aacd-04292e80f33d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:04:06 crc kubenswrapper[4802]: I1206 04:04:06.516067 4802 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/44aaebb1-a5a3-42e6-aacd-04292e80f33d-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 04:04:06 crc kubenswrapper[4802]: I1206 04:04:06.516107 4802 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44aaebb1-a5a3-42e6-aacd-04292e80f33d-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:04:06 crc kubenswrapper[4802]: I1206 04:04:06.516118 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2cqqx\" (UniqueName: \"kubernetes.io/projected/44aaebb1-a5a3-42e6-aacd-04292e80f33d-kube-api-access-2cqqx\") on node \"crc\" DevicePath \"\"" Dec 06 04:04:06 crc kubenswrapper[4802]: I1206 04:04:06.516128 4802 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/44aaebb1-a5a3-42e6-aacd-04292e80f33d-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 04:04:06 crc kubenswrapper[4802]: I1206 04:04:06.915876 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jrmv7" event={"ID":"44aaebb1-a5a3-42e6-aacd-04292e80f33d","Type":"ContainerDied","Data":"8968b097f18b0527a0584a1eda16dfe8d50389bf421c23342cb736c7b9112c97"} Dec 06 04:04:06 crc kubenswrapper[4802]: I1206 04:04:06.915913 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jrmv7" Dec 06 04:04:06 crc kubenswrapper[4802]: I1206 04:04:06.915925 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8968b097f18b0527a0584a1eda16dfe8d50389bf421c23342cb736c7b9112c97" Dec 06 04:04:06 crc kubenswrapper[4802]: I1206 04:04:06.984320 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-tgsk9"] Dec 06 04:04:06 crc kubenswrapper[4802]: E1206 04:04:06.984725 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44aaebb1-a5a3-42e6-aacd-04292e80f33d" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 06 04:04:06 crc kubenswrapper[4802]: I1206 04:04:06.984743 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="44aaebb1-a5a3-42e6-aacd-04292e80f33d" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 06 04:04:06 crc kubenswrapper[4802]: I1206 04:04:06.984941 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="44aaebb1-a5a3-42e6-aacd-04292e80f33d" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 06 04:04:06 crc kubenswrapper[4802]: I1206 04:04:06.985569 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tgsk9" Dec 06 04:04:06 crc kubenswrapper[4802]: I1206 04:04:06.988502 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 04:04:06 crc kubenswrapper[4802]: I1206 04:04:06.988912 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-rgvjc" Dec 06 04:04:06 crc kubenswrapper[4802]: I1206 04:04:06.989087 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 04:04:06 crc kubenswrapper[4802]: I1206 04:04:06.989254 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 04:04:07 crc kubenswrapper[4802]: I1206 04:04:07.002805 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-tgsk9"] Dec 06 04:04:07 crc kubenswrapper[4802]: I1206 04:04:07.127316 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/63d8761a-c5bb-41d5-a830-f58cc99b0837-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-tgsk9\" (UID: \"63d8761a-c5bb-41d5-a830-f58cc99b0837\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tgsk9" Dec 06 04:04:07 crc kubenswrapper[4802]: I1206 04:04:07.127489 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/63d8761a-c5bb-41d5-a830-f58cc99b0837-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-tgsk9\" (UID: \"63d8761a-c5bb-41d5-a830-f58cc99b0837\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tgsk9" Dec 06 04:04:07 crc kubenswrapper[4802]: I1206 04:04:07.127675 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2zt5\" (UniqueName: \"kubernetes.io/projected/63d8761a-c5bb-41d5-a830-f58cc99b0837-kube-api-access-z2zt5\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-tgsk9\" (UID: \"63d8761a-c5bb-41d5-a830-f58cc99b0837\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tgsk9" Dec 06 04:04:07 crc kubenswrapper[4802]: I1206 04:04:07.229719 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/63d8761a-c5bb-41d5-a830-f58cc99b0837-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-tgsk9\" (UID: \"63d8761a-c5bb-41d5-a830-f58cc99b0837\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tgsk9" Dec 06 04:04:07 crc kubenswrapper[4802]: I1206 04:04:07.230041 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/63d8761a-c5bb-41d5-a830-f58cc99b0837-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-tgsk9\" (UID: \"63d8761a-c5bb-41d5-a830-f58cc99b0837\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tgsk9" Dec 06 04:04:07 crc kubenswrapper[4802]: I1206 04:04:07.230106 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2zt5\" (UniqueName: \"kubernetes.io/projected/63d8761a-c5bb-41d5-a830-f58cc99b0837-kube-api-access-z2zt5\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-tgsk9\" (UID: \"63d8761a-c5bb-41d5-a830-f58cc99b0837\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tgsk9" Dec 06 04:04:07 crc kubenswrapper[4802]: I1206 04:04:07.240696 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/63d8761a-c5bb-41d5-a830-f58cc99b0837-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-tgsk9\" (UID: \"63d8761a-c5bb-41d5-a830-f58cc99b0837\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tgsk9" Dec 06 04:04:07 crc kubenswrapper[4802]: I1206 04:04:07.240765 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/63d8761a-c5bb-41d5-a830-f58cc99b0837-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-tgsk9\" (UID: \"63d8761a-c5bb-41d5-a830-f58cc99b0837\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tgsk9" Dec 06 04:04:07 crc kubenswrapper[4802]: I1206 04:04:07.263276 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2zt5\" (UniqueName: \"kubernetes.io/projected/63d8761a-c5bb-41d5-a830-f58cc99b0837-kube-api-access-z2zt5\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-tgsk9\" (UID: \"63d8761a-c5bb-41d5-a830-f58cc99b0837\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tgsk9" Dec 06 04:04:07 crc kubenswrapper[4802]: I1206 04:04:07.304737 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tgsk9" Dec 06 04:04:07 crc kubenswrapper[4802]: I1206 04:04:07.889557 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-tgsk9"] Dec 06 04:04:07 crc kubenswrapper[4802]: I1206 04:04:07.926934 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tgsk9" event={"ID":"63d8761a-c5bb-41d5-a830-f58cc99b0837","Type":"ContainerStarted","Data":"61f2b9b00c3bfed16afedaf96ca44e2b15a54b1eed3abee31cf066d9a07fa06a"} Dec 06 04:04:08 crc kubenswrapper[4802]: I1206 04:04:08.610703 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 04:04:09 crc kubenswrapper[4802]: I1206 04:04:09.947521 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tgsk9" event={"ID":"63d8761a-c5bb-41d5-a830-f58cc99b0837","Type":"ContainerStarted","Data":"b36709b1b0047615bbbf4822b4b4a70c34dd57a2ce5a62dbe9e5f4ae966a5ddf"} Dec 06 04:04:09 crc kubenswrapper[4802]: I1206 04:04:09.972975 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tgsk9" podStartSLOduration=3.262006206 podStartE2EDuration="3.972959906s" podCreationTimestamp="2025-12-06 04:04:06 +0000 UTC" firstStartedPulling="2025-12-06 04:04:07.896808065 +0000 UTC m=+1440.768717217" lastFinishedPulling="2025-12-06 04:04:08.607761765 +0000 UTC m=+1441.479670917" observedRunningTime="2025-12-06 04:04:09.966483649 +0000 UTC m=+1442.838392801" watchObservedRunningTime="2025-12-06 04:04:09.972959906 +0000 UTC m=+1442.844869058" Dec 06 04:04:11 crc kubenswrapper[4802]: I1206 04:04:11.970078 4802 generic.go:334] "Generic (PLEG): container finished" podID="63d8761a-c5bb-41d5-a830-f58cc99b0837" containerID="b36709b1b0047615bbbf4822b4b4a70c34dd57a2ce5a62dbe9e5f4ae966a5ddf" exitCode=0 Dec 06 04:04:11 crc kubenswrapper[4802]: I1206 04:04:11.970140 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tgsk9" event={"ID":"63d8761a-c5bb-41d5-a830-f58cc99b0837","Type":"ContainerDied","Data":"b36709b1b0047615bbbf4822b4b4a70c34dd57a2ce5a62dbe9e5f4ae966a5ddf"} Dec 06 04:04:13 crc kubenswrapper[4802]: I1206 04:04:13.491629 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tgsk9" Dec 06 04:04:13 crc kubenswrapper[4802]: I1206 04:04:13.577586 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z2zt5\" (UniqueName: \"kubernetes.io/projected/63d8761a-c5bb-41d5-a830-f58cc99b0837-kube-api-access-z2zt5\") pod \"63d8761a-c5bb-41d5-a830-f58cc99b0837\" (UID: \"63d8761a-c5bb-41d5-a830-f58cc99b0837\") " Dec 06 04:04:13 crc kubenswrapper[4802]: I1206 04:04:13.577711 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/63d8761a-c5bb-41d5-a830-f58cc99b0837-inventory\") pod \"63d8761a-c5bb-41d5-a830-f58cc99b0837\" (UID: \"63d8761a-c5bb-41d5-a830-f58cc99b0837\") " Dec 06 04:04:13 crc kubenswrapper[4802]: I1206 04:04:13.577766 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/63d8761a-c5bb-41d5-a830-f58cc99b0837-ssh-key\") pod \"63d8761a-c5bb-41d5-a830-f58cc99b0837\" (UID: \"63d8761a-c5bb-41d5-a830-f58cc99b0837\") " Dec 06 04:04:13 crc kubenswrapper[4802]: I1206 04:04:13.583640 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63d8761a-c5bb-41d5-a830-f58cc99b0837-kube-api-access-z2zt5" (OuterVolumeSpecName: "kube-api-access-z2zt5") pod "63d8761a-c5bb-41d5-a830-f58cc99b0837" (UID: "63d8761a-c5bb-41d5-a830-f58cc99b0837"). InnerVolumeSpecName "kube-api-access-z2zt5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:04:13 crc kubenswrapper[4802]: I1206 04:04:13.610952 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63d8761a-c5bb-41d5-a830-f58cc99b0837-inventory" (OuterVolumeSpecName: "inventory") pod "63d8761a-c5bb-41d5-a830-f58cc99b0837" (UID: "63d8761a-c5bb-41d5-a830-f58cc99b0837"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:04:13 crc kubenswrapper[4802]: I1206 04:04:13.637269 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63d8761a-c5bb-41d5-a830-f58cc99b0837-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "63d8761a-c5bb-41d5-a830-f58cc99b0837" (UID: "63d8761a-c5bb-41d5-a830-f58cc99b0837"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:04:13 crc kubenswrapper[4802]: I1206 04:04:13.679930 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z2zt5\" (UniqueName: \"kubernetes.io/projected/63d8761a-c5bb-41d5-a830-f58cc99b0837-kube-api-access-z2zt5\") on node \"crc\" DevicePath \"\"" Dec 06 04:04:13 crc kubenswrapper[4802]: I1206 04:04:13.679960 4802 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/63d8761a-c5bb-41d5-a830-f58cc99b0837-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 04:04:13 crc kubenswrapper[4802]: I1206 04:04:13.679970 4802 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/63d8761a-c5bb-41d5-a830-f58cc99b0837-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 04:04:14 crc kubenswrapper[4802]: I1206 04:04:14.000506 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tgsk9" event={"ID":"63d8761a-c5bb-41d5-a830-f58cc99b0837","Type":"ContainerDied","Data":"61f2b9b00c3bfed16afedaf96ca44e2b15a54b1eed3abee31cf066d9a07fa06a"} Dec 06 04:04:14 crc kubenswrapper[4802]: I1206 04:04:14.000557 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="61f2b9b00c3bfed16afedaf96ca44e2b15a54b1eed3abee31cf066d9a07fa06a" Dec 06 04:04:14 crc kubenswrapper[4802]: I1206 04:04:14.001101 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tgsk9" Dec 06 04:04:14 crc kubenswrapper[4802]: I1206 04:04:14.056547 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-n5lkw"] Dec 06 04:04:14 crc kubenswrapper[4802]: E1206 04:04:14.057239 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63d8761a-c5bb-41d5-a830-f58cc99b0837" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 06 04:04:14 crc kubenswrapper[4802]: I1206 04:04:14.057433 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="63d8761a-c5bb-41d5-a830-f58cc99b0837" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 06 04:04:14 crc kubenswrapper[4802]: I1206 04:04:14.057719 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="63d8761a-c5bb-41d5-a830-f58cc99b0837" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 06 04:04:14 crc kubenswrapper[4802]: I1206 04:04:14.058722 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-n5lkw" Dec 06 04:04:14 crc kubenswrapper[4802]: I1206 04:04:14.062803 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-rgvjc" Dec 06 04:04:14 crc kubenswrapper[4802]: I1206 04:04:14.062982 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 04:04:14 crc kubenswrapper[4802]: I1206 04:04:14.063088 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 04:04:14 crc kubenswrapper[4802]: I1206 04:04:14.064439 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 04:04:14 crc kubenswrapper[4802]: I1206 04:04:14.073853 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-n5lkw"] Dec 06 04:04:14 crc kubenswrapper[4802]: I1206 04:04:14.190062 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a07f3a26-501d-49cb-8dbf-648f1cac66da-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-n5lkw\" (UID: \"a07f3a26-501d-49cb-8dbf-648f1cac66da\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-n5lkw" Dec 06 04:04:14 crc kubenswrapper[4802]: I1206 04:04:14.190145 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a07f3a26-501d-49cb-8dbf-648f1cac66da-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-n5lkw\" (UID: \"a07f3a26-501d-49cb-8dbf-648f1cac66da\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-n5lkw" Dec 06 04:04:14 crc kubenswrapper[4802]: I1206 04:04:14.190211 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnbxt\" (UniqueName: \"kubernetes.io/projected/a07f3a26-501d-49cb-8dbf-648f1cac66da-kube-api-access-rnbxt\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-n5lkw\" (UID: \"a07f3a26-501d-49cb-8dbf-648f1cac66da\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-n5lkw" Dec 06 04:04:14 crc kubenswrapper[4802]: I1206 04:04:14.190241 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a07f3a26-501d-49cb-8dbf-648f1cac66da-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-n5lkw\" (UID: \"a07f3a26-501d-49cb-8dbf-648f1cac66da\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-n5lkw" Dec 06 04:04:14 crc kubenswrapper[4802]: I1206 04:04:14.291653 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a07f3a26-501d-49cb-8dbf-648f1cac66da-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-n5lkw\" (UID: \"a07f3a26-501d-49cb-8dbf-648f1cac66da\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-n5lkw" Dec 06 04:04:14 crc kubenswrapper[4802]: I1206 04:04:14.291834 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a07f3a26-501d-49cb-8dbf-648f1cac66da-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-n5lkw\" (UID: \"a07f3a26-501d-49cb-8dbf-648f1cac66da\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-n5lkw" Dec 06 04:04:14 crc kubenswrapper[4802]: I1206 04:04:14.291884 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a07f3a26-501d-49cb-8dbf-648f1cac66da-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-n5lkw\" (UID: \"a07f3a26-501d-49cb-8dbf-648f1cac66da\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-n5lkw" Dec 06 04:04:14 crc kubenswrapper[4802]: I1206 04:04:14.291925 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnbxt\" (UniqueName: \"kubernetes.io/projected/a07f3a26-501d-49cb-8dbf-648f1cac66da-kube-api-access-rnbxt\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-n5lkw\" (UID: \"a07f3a26-501d-49cb-8dbf-648f1cac66da\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-n5lkw" Dec 06 04:04:14 crc kubenswrapper[4802]: I1206 04:04:14.296054 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a07f3a26-501d-49cb-8dbf-648f1cac66da-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-n5lkw\" (UID: \"a07f3a26-501d-49cb-8dbf-648f1cac66da\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-n5lkw" Dec 06 04:04:14 crc kubenswrapper[4802]: I1206 04:04:14.297210 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a07f3a26-501d-49cb-8dbf-648f1cac66da-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-n5lkw\" (UID: \"a07f3a26-501d-49cb-8dbf-648f1cac66da\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-n5lkw" Dec 06 04:04:14 crc kubenswrapper[4802]: I1206 04:04:14.297418 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a07f3a26-501d-49cb-8dbf-648f1cac66da-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-n5lkw\" (UID: \"a07f3a26-501d-49cb-8dbf-648f1cac66da\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-n5lkw" Dec 06 04:04:14 crc kubenswrapper[4802]: I1206 04:04:14.308027 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnbxt\" (UniqueName: \"kubernetes.io/projected/a07f3a26-501d-49cb-8dbf-648f1cac66da-kube-api-access-rnbxt\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-n5lkw\" (UID: \"a07f3a26-501d-49cb-8dbf-648f1cac66da\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-n5lkw" Dec 06 04:04:14 crc kubenswrapper[4802]: I1206 04:04:14.395788 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-n5lkw" Dec 06 04:04:14 crc kubenswrapper[4802]: I1206 04:04:14.945172 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-n5lkw"] Dec 06 04:04:15 crc kubenswrapper[4802]: I1206 04:04:15.016424 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-n5lkw" event={"ID":"a07f3a26-501d-49cb-8dbf-648f1cac66da","Type":"ContainerStarted","Data":"c696b288132f6ecc51dd2dc8ff43b41280a9bbbc41f1ec8132ab7cefcbcd47fd"} Dec 06 04:04:17 crc kubenswrapper[4802]: I1206 04:04:17.045872 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-n5lkw" event={"ID":"a07f3a26-501d-49cb-8dbf-648f1cac66da","Type":"ContainerStarted","Data":"984316e8889de93db585481aa2ce837159302ff37f09380b6a58357681518637"} Dec 06 04:04:17 crc kubenswrapper[4802]: I1206 04:04:17.084396 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-n5lkw" podStartSLOduration=2.225478295 podStartE2EDuration="3.084362833s" podCreationTimestamp="2025-12-06 04:04:14 +0000 UTC" firstStartedPulling="2025-12-06 04:04:14.948914113 +0000 UTC m=+1447.820823265" lastFinishedPulling="2025-12-06 04:04:15.807798641 +0000 UTC m=+1448.679707803" observedRunningTime="2025-12-06 04:04:17.074292718 +0000 UTC m=+1449.946201920" watchObservedRunningTime="2025-12-06 04:04:17.084362833 +0000 UTC m=+1449.956272025" Dec 06 04:04:19 crc kubenswrapper[4802]: I1206 04:04:19.641794 4802 scope.go:117] "RemoveContainer" containerID="bbc75983c8257f385dfab8690e21029a136f6c9ea7a2b913140132166d7871b0" Dec 06 04:04:43 crc kubenswrapper[4802]: I1206 04:04:43.283203 4802 patch_prober.go:28] interesting pod/machine-config-daemon-zpxxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:04:43 crc kubenswrapper[4802]: I1206 04:04:43.283981 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:05:13 crc kubenswrapper[4802]: I1206 04:05:13.283334 4802 patch_prober.go:28] interesting pod/machine-config-daemon-zpxxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:05:13 crc kubenswrapper[4802]: I1206 04:05:13.283931 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:05:19 crc kubenswrapper[4802]: I1206 04:05:19.758117 4802 scope.go:117] "RemoveContainer" containerID="0f19de9d3ffb77620b7ecce801747092e84300ff9f114eff03946ce3f3ce3405" Dec 06 04:05:26 crc kubenswrapper[4802]: I1206 04:05:26.119131 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-9x45c"] Dec 06 04:05:26 crc kubenswrapper[4802]: I1206 04:05:26.121712 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9x45c" Dec 06 04:05:26 crc kubenswrapper[4802]: I1206 04:05:26.172923 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9x45c"] Dec 06 04:05:26 crc kubenswrapper[4802]: I1206 04:05:26.280197 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9ea1538-a9d0-4ca5-b76f-d041473ce005-utilities\") pod \"redhat-marketplace-9x45c\" (UID: \"d9ea1538-a9d0-4ca5-b76f-d041473ce005\") " pod="openshift-marketplace/redhat-marketplace-9x45c" Dec 06 04:05:26 crc kubenswrapper[4802]: I1206 04:05:26.280567 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9ea1538-a9d0-4ca5-b76f-d041473ce005-catalog-content\") pod \"redhat-marketplace-9x45c\" (UID: \"d9ea1538-a9d0-4ca5-b76f-d041473ce005\") " pod="openshift-marketplace/redhat-marketplace-9x45c" Dec 06 04:05:26 crc kubenswrapper[4802]: I1206 04:05:26.280645 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htnbt\" (UniqueName: \"kubernetes.io/projected/d9ea1538-a9d0-4ca5-b76f-d041473ce005-kube-api-access-htnbt\") pod \"redhat-marketplace-9x45c\" (UID: \"d9ea1538-a9d0-4ca5-b76f-d041473ce005\") " pod="openshift-marketplace/redhat-marketplace-9x45c" Dec 06 04:05:26 crc kubenswrapper[4802]: I1206 04:05:26.382511 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9ea1538-a9d0-4ca5-b76f-d041473ce005-utilities\") pod \"redhat-marketplace-9x45c\" (UID: \"d9ea1538-a9d0-4ca5-b76f-d041473ce005\") " pod="openshift-marketplace/redhat-marketplace-9x45c" Dec 06 04:05:26 crc kubenswrapper[4802]: I1206 04:05:26.382586 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9ea1538-a9d0-4ca5-b76f-d041473ce005-catalog-content\") pod \"redhat-marketplace-9x45c\" (UID: \"d9ea1538-a9d0-4ca5-b76f-d041473ce005\") " pod="openshift-marketplace/redhat-marketplace-9x45c" Dec 06 04:05:26 crc kubenswrapper[4802]: I1206 04:05:26.382645 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htnbt\" (UniqueName: \"kubernetes.io/projected/d9ea1538-a9d0-4ca5-b76f-d041473ce005-kube-api-access-htnbt\") pod \"redhat-marketplace-9x45c\" (UID: \"d9ea1538-a9d0-4ca5-b76f-d041473ce005\") " pod="openshift-marketplace/redhat-marketplace-9x45c" Dec 06 04:05:26 crc kubenswrapper[4802]: I1206 04:05:26.383004 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9ea1538-a9d0-4ca5-b76f-d041473ce005-utilities\") pod \"redhat-marketplace-9x45c\" (UID: \"d9ea1538-a9d0-4ca5-b76f-d041473ce005\") " pod="openshift-marketplace/redhat-marketplace-9x45c" Dec 06 04:05:26 crc kubenswrapper[4802]: I1206 04:05:26.383353 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9ea1538-a9d0-4ca5-b76f-d041473ce005-catalog-content\") pod \"redhat-marketplace-9x45c\" (UID: \"d9ea1538-a9d0-4ca5-b76f-d041473ce005\") " pod="openshift-marketplace/redhat-marketplace-9x45c" Dec 06 04:05:26 crc kubenswrapper[4802]: I1206 04:05:26.404128 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htnbt\" (UniqueName: \"kubernetes.io/projected/d9ea1538-a9d0-4ca5-b76f-d041473ce005-kube-api-access-htnbt\") pod \"redhat-marketplace-9x45c\" (UID: \"d9ea1538-a9d0-4ca5-b76f-d041473ce005\") " pod="openshift-marketplace/redhat-marketplace-9x45c" Dec 06 04:05:26 crc kubenswrapper[4802]: I1206 04:05:26.502032 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9x45c" Dec 06 04:05:26 crc kubenswrapper[4802]: I1206 04:05:26.979104 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9x45c"] Dec 06 04:05:26 crc kubenswrapper[4802]: W1206 04:05:26.988583 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd9ea1538_a9d0_4ca5_b76f_d041473ce005.slice/crio-a75d8f17b927850185e75dc4668841dab1ed57451b22e1803cc6c8d2c33031db WatchSource:0}: Error finding container a75d8f17b927850185e75dc4668841dab1ed57451b22e1803cc6c8d2c33031db: Status 404 returned error can't find the container with id a75d8f17b927850185e75dc4668841dab1ed57451b22e1803cc6c8d2c33031db Dec 06 04:05:27 crc kubenswrapper[4802]: I1206 04:05:27.839278 4802 generic.go:334] "Generic (PLEG): container finished" podID="d9ea1538-a9d0-4ca5-b76f-d041473ce005" containerID="54f86b8ac9a969a7420186e0a34af66744ab99efbee3a77e1bee19399aafa117" exitCode=0 Dec 06 04:05:27 crc kubenswrapper[4802]: I1206 04:05:27.839322 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9x45c" event={"ID":"d9ea1538-a9d0-4ca5-b76f-d041473ce005","Type":"ContainerDied","Data":"54f86b8ac9a969a7420186e0a34af66744ab99efbee3a77e1bee19399aafa117"} Dec 06 04:05:27 crc kubenswrapper[4802]: I1206 04:05:27.839350 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9x45c" event={"ID":"d9ea1538-a9d0-4ca5-b76f-d041473ce005","Type":"ContainerStarted","Data":"a75d8f17b927850185e75dc4668841dab1ed57451b22e1803cc6c8d2c33031db"} Dec 06 04:05:29 crc kubenswrapper[4802]: I1206 04:05:29.861400 4802 generic.go:334] "Generic (PLEG): container finished" podID="d9ea1538-a9d0-4ca5-b76f-d041473ce005" containerID="29dfb28d716cec62e93af5f1d95f5b0412e8e1cb43d055b26269805d456f808c" exitCode=0 Dec 06 04:05:29 crc kubenswrapper[4802]: I1206 04:05:29.861490 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9x45c" event={"ID":"d9ea1538-a9d0-4ca5-b76f-d041473ce005","Type":"ContainerDied","Data":"29dfb28d716cec62e93af5f1d95f5b0412e8e1cb43d055b26269805d456f808c"} Dec 06 04:05:31 crc kubenswrapper[4802]: I1206 04:05:31.883699 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9x45c" event={"ID":"d9ea1538-a9d0-4ca5-b76f-d041473ce005","Type":"ContainerStarted","Data":"5f622b674707ce8a16ccc5b808e6aaf1cdc79b449134ef889f9c6dda34193d15"} Dec 06 04:05:31 crc kubenswrapper[4802]: I1206 04:05:31.905982 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-9x45c" podStartSLOduration=3.097262333 podStartE2EDuration="5.905963253s" podCreationTimestamp="2025-12-06 04:05:26 +0000 UTC" firstStartedPulling="2025-12-06 04:05:27.84372342 +0000 UTC m=+1520.715632572" lastFinishedPulling="2025-12-06 04:05:30.65242434 +0000 UTC m=+1523.524333492" observedRunningTime="2025-12-06 04:05:31.898788407 +0000 UTC m=+1524.770697559" watchObservedRunningTime="2025-12-06 04:05:31.905963253 +0000 UTC m=+1524.777872405" Dec 06 04:05:36 crc kubenswrapper[4802]: I1206 04:05:36.502802 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-9x45c" Dec 06 04:05:36 crc kubenswrapper[4802]: I1206 04:05:36.503346 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-9x45c" Dec 06 04:05:36 crc kubenswrapper[4802]: I1206 04:05:36.566264 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-9x45c" Dec 06 04:05:36 crc kubenswrapper[4802]: I1206 04:05:36.994480 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-9x45c" Dec 06 04:05:37 crc kubenswrapper[4802]: I1206 04:05:37.802783 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9x45c"] Dec 06 04:05:38 crc kubenswrapper[4802]: I1206 04:05:38.967107 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-9x45c" podUID="d9ea1538-a9d0-4ca5-b76f-d041473ce005" containerName="registry-server" containerID="cri-o://5f622b674707ce8a16ccc5b808e6aaf1cdc79b449134ef889f9c6dda34193d15" gracePeriod=2 Dec 06 04:05:39 crc kubenswrapper[4802]: I1206 04:05:39.639685 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9x45c" Dec 06 04:05:39 crc kubenswrapper[4802]: I1206 04:05:39.772039 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9ea1538-a9d0-4ca5-b76f-d041473ce005-catalog-content\") pod \"d9ea1538-a9d0-4ca5-b76f-d041473ce005\" (UID: \"d9ea1538-a9d0-4ca5-b76f-d041473ce005\") " Dec 06 04:05:39 crc kubenswrapper[4802]: I1206 04:05:39.772466 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htnbt\" (UniqueName: \"kubernetes.io/projected/d9ea1538-a9d0-4ca5-b76f-d041473ce005-kube-api-access-htnbt\") pod \"d9ea1538-a9d0-4ca5-b76f-d041473ce005\" (UID: \"d9ea1538-a9d0-4ca5-b76f-d041473ce005\") " Dec 06 04:05:39 crc kubenswrapper[4802]: I1206 04:05:39.772535 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9ea1538-a9d0-4ca5-b76f-d041473ce005-utilities\") pod \"d9ea1538-a9d0-4ca5-b76f-d041473ce005\" (UID: \"d9ea1538-a9d0-4ca5-b76f-d041473ce005\") " Dec 06 04:05:39 crc kubenswrapper[4802]: I1206 04:05:39.774190 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9ea1538-a9d0-4ca5-b76f-d041473ce005-utilities" (OuterVolumeSpecName: "utilities") pod "d9ea1538-a9d0-4ca5-b76f-d041473ce005" (UID: "d9ea1538-a9d0-4ca5-b76f-d041473ce005"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:05:39 crc kubenswrapper[4802]: I1206 04:05:39.783776 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9ea1538-a9d0-4ca5-b76f-d041473ce005-kube-api-access-htnbt" (OuterVolumeSpecName: "kube-api-access-htnbt") pod "d9ea1538-a9d0-4ca5-b76f-d041473ce005" (UID: "d9ea1538-a9d0-4ca5-b76f-d041473ce005"). InnerVolumeSpecName "kube-api-access-htnbt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:05:39 crc kubenswrapper[4802]: I1206 04:05:39.793873 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9ea1538-a9d0-4ca5-b76f-d041473ce005-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d9ea1538-a9d0-4ca5-b76f-d041473ce005" (UID: "d9ea1538-a9d0-4ca5-b76f-d041473ce005"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:05:39 crc kubenswrapper[4802]: I1206 04:05:39.874662 4802 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9ea1538-a9d0-4ca5-b76f-d041473ce005-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 04:05:39 crc kubenswrapper[4802]: I1206 04:05:39.874705 4802 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9ea1538-a9d0-4ca5-b76f-d041473ce005-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 04:05:39 crc kubenswrapper[4802]: I1206 04:05:39.874723 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htnbt\" (UniqueName: \"kubernetes.io/projected/d9ea1538-a9d0-4ca5-b76f-d041473ce005-kube-api-access-htnbt\") on node \"crc\" DevicePath \"\"" Dec 06 04:05:39 crc kubenswrapper[4802]: I1206 04:05:39.982106 4802 generic.go:334] "Generic (PLEG): container finished" podID="d9ea1538-a9d0-4ca5-b76f-d041473ce005" containerID="5f622b674707ce8a16ccc5b808e6aaf1cdc79b449134ef889f9c6dda34193d15" exitCode=0 Dec 06 04:05:39 crc kubenswrapper[4802]: I1206 04:05:39.982164 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9x45c" event={"ID":"d9ea1538-a9d0-4ca5-b76f-d041473ce005","Type":"ContainerDied","Data":"5f622b674707ce8a16ccc5b808e6aaf1cdc79b449134ef889f9c6dda34193d15"} Dec 06 04:05:39 crc kubenswrapper[4802]: I1206 04:05:39.982189 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9x45c" Dec 06 04:05:39 crc kubenswrapper[4802]: I1206 04:05:39.982221 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9x45c" event={"ID":"d9ea1538-a9d0-4ca5-b76f-d041473ce005","Type":"ContainerDied","Data":"a75d8f17b927850185e75dc4668841dab1ed57451b22e1803cc6c8d2c33031db"} Dec 06 04:05:39 crc kubenswrapper[4802]: I1206 04:05:39.982248 4802 scope.go:117] "RemoveContainer" containerID="5f622b674707ce8a16ccc5b808e6aaf1cdc79b449134ef889f9c6dda34193d15" Dec 06 04:05:40 crc kubenswrapper[4802]: I1206 04:05:40.016070 4802 scope.go:117] "RemoveContainer" containerID="29dfb28d716cec62e93af5f1d95f5b0412e8e1cb43d055b26269805d456f808c" Dec 06 04:05:40 crc kubenswrapper[4802]: I1206 04:05:40.030196 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9x45c"] Dec 06 04:05:40 crc kubenswrapper[4802]: I1206 04:05:40.039270 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-9x45c"] Dec 06 04:05:40 crc kubenswrapper[4802]: I1206 04:05:40.055600 4802 scope.go:117] "RemoveContainer" containerID="54f86b8ac9a969a7420186e0a34af66744ab99efbee3a77e1bee19399aafa117" Dec 06 04:05:40 crc kubenswrapper[4802]: I1206 04:05:40.084036 4802 scope.go:117] "RemoveContainer" containerID="5f622b674707ce8a16ccc5b808e6aaf1cdc79b449134ef889f9c6dda34193d15" Dec 06 04:05:40 crc kubenswrapper[4802]: E1206 04:05:40.084518 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f622b674707ce8a16ccc5b808e6aaf1cdc79b449134ef889f9c6dda34193d15\": container with ID starting with 5f622b674707ce8a16ccc5b808e6aaf1cdc79b449134ef889f9c6dda34193d15 not found: ID does not exist" containerID="5f622b674707ce8a16ccc5b808e6aaf1cdc79b449134ef889f9c6dda34193d15" Dec 06 04:05:40 crc kubenswrapper[4802]: I1206 04:05:40.084568 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f622b674707ce8a16ccc5b808e6aaf1cdc79b449134ef889f9c6dda34193d15"} err="failed to get container status \"5f622b674707ce8a16ccc5b808e6aaf1cdc79b449134ef889f9c6dda34193d15\": rpc error: code = NotFound desc = could not find container \"5f622b674707ce8a16ccc5b808e6aaf1cdc79b449134ef889f9c6dda34193d15\": container with ID starting with 5f622b674707ce8a16ccc5b808e6aaf1cdc79b449134ef889f9c6dda34193d15 not found: ID does not exist" Dec 06 04:05:40 crc kubenswrapper[4802]: I1206 04:05:40.084602 4802 scope.go:117] "RemoveContainer" containerID="29dfb28d716cec62e93af5f1d95f5b0412e8e1cb43d055b26269805d456f808c" Dec 06 04:05:40 crc kubenswrapper[4802]: E1206 04:05:40.085051 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29dfb28d716cec62e93af5f1d95f5b0412e8e1cb43d055b26269805d456f808c\": container with ID starting with 29dfb28d716cec62e93af5f1d95f5b0412e8e1cb43d055b26269805d456f808c not found: ID does not exist" containerID="29dfb28d716cec62e93af5f1d95f5b0412e8e1cb43d055b26269805d456f808c" Dec 06 04:05:40 crc kubenswrapper[4802]: I1206 04:05:40.085091 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29dfb28d716cec62e93af5f1d95f5b0412e8e1cb43d055b26269805d456f808c"} err="failed to get container status \"29dfb28d716cec62e93af5f1d95f5b0412e8e1cb43d055b26269805d456f808c\": rpc error: code = NotFound desc = could not find container \"29dfb28d716cec62e93af5f1d95f5b0412e8e1cb43d055b26269805d456f808c\": container with ID starting with 29dfb28d716cec62e93af5f1d95f5b0412e8e1cb43d055b26269805d456f808c not found: ID does not exist" Dec 06 04:05:40 crc kubenswrapper[4802]: I1206 04:05:40.085145 4802 scope.go:117] "RemoveContainer" containerID="54f86b8ac9a969a7420186e0a34af66744ab99efbee3a77e1bee19399aafa117" Dec 06 04:05:40 crc kubenswrapper[4802]: E1206 04:05:40.085385 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54f86b8ac9a969a7420186e0a34af66744ab99efbee3a77e1bee19399aafa117\": container with ID starting with 54f86b8ac9a969a7420186e0a34af66744ab99efbee3a77e1bee19399aafa117 not found: ID does not exist" containerID="54f86b8ac9a969a7420186e0a34af66744ab99efbee3a77e1bee19399aafa117" Dec 06 04:05:40 crc kubenswrapper[4802]: I1206 04:05:40.085428 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54f86b8ac9a969a7420186e0a34af66744ab99efbee3a77e1bee19399aafa117"} err="failed to get container status \"54f86b8ac9a969a7420186e0a34af66744ab99efbee3a77e1bee19399aafa117\": rpc error: code = NotFound desc = could not find container \"54f86b8ac9a969a7420186e0a34af66744ab99efbee3a77e1bee19399aafa117\": container with ID starting with 54f86b8ac9a969a7420186e0a34af66744ab99efbee3a77e1bee19399aafa117 not found: ID does not exist" Dec 06 04:05:41 crc kubenswrapper[4802]: I1206 04:05:41.460968 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9ea1538-a9d0-4ca5-b76f-d041473ce005" path="/var/lib/kubelet/pods/d9ea1538-a9d0-4ca5-b76f-d041473ce005/volumes" Dec 06 04:05:43 crc kubenswrapper[4802]: I1206 04:05:43.283651 4802 patch_prober.go:28] interesting pod/machine-config-daemon-zpxxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:05:43 crc kubenswrapper[4802]: I1206 04:05:43.284028 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:05:43 crc kubenswrapper[4802]: I1206 04:05:43.284087 4802 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" Dec 06 04:05:43 crc kubenswrapper[4802]: I1206 04:05:43.284869 4802 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3711adecb81191bc35ac469efa73faeb2346d30f4d5a4a148de8ac88071cdb00"} pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 04:05:43 crc kubenswrapper[4802]: I1206 04:05:43.284924 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" containerID="cri-o://3711adecb81191bc35ac469efa73faeb2346d30f4d5a4a148de8ac88071cdb00" gracePeriod=600 Dec 06 04:05:43 crc kubenswrapper[4802]: E1206 04:05:43.910909 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:05:44 crc kubenswrapper[4802]: I1206 04:05:44.024186 4802 generic.go:334] "Generic (PLEG): container finished" podID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerID="3711adecb81191bc35ac469efa73faeb2346d30f4d5a4a148de8ac88071cdb00" exitCode=0 Dec 06 04:05:44 crc kubenswrapper[4802]: I1206 04:05:44.024234 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" event={"ID":"6df38316-e0d3-4018-8d27-3620eba3a68d","Type":"ContainerDied","Data":"3711adecb81191bc35ac469efa73faeb2346d30f4d5a4a148de8ac88071cdb00"} Dec 06 04:05:44 crc kubenswrapper[4802]: I1206 04:05:44.024272 4802 scope.go:117] "RemoveContainer" containerID="61a067de7e4bf972bb44f2b14a9f2f9bdc755648ec9a47e6260216803462a3b1" Dec 06 04:05:44 crc kubenswrapper[4802]: I1206 04:05:44.025048 4802 scope.go:117] "RemoveContainer" containerID="3711adecb81191bc35ac469efa73faeb2346d30f4d5a4a148de8ac88071cdb00" Dec 06 04:05:44 crc kubenswrapper[4802]: E1206 04:05:44.025374 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:05:50 crc kubenswrapper[4802]: I1206 04:05:50.630358 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dqb9s"] Dec 06 04:05:50 crc kubenswrapper[4802]: E1206 04:05:50.632231 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9ea1538-a9d0-4ca5-b76f-d041473ce005" containerName="extract-utilities" Dec 06 04:05:50 crc kubenswrapper[4802]: I1206 04:05:50.632392 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9ea1538-a9d0-4ca5-b76f-d041473ce005" containerName="extract-utilities" Dec 06 04:05:50 crc kubenswrapper[4802]: E1206 04:05:50.632463 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9ea1538-a9d0-4ca5-b76f-d041473ce005" containerName="registry-server" Dec 06 04:05:50 crc kubenswrapper[4802]: I1206 04:05:50.632518 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9ea1538-a9d0-4ca5-b76f-d041473ce005" containerName="registry-server" Dec 06 04:05:50 crc kubenswrapper[4802]: E1206 04:05:50.632588 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9ea1538-a9d0-4ca5-b76f-d041473ce005" containerName="extract-content" Dec 06 04:05:50 crc kubenswrapper[4802]: I1206 04:05:50.632658 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9ea1538-a9d0-4ca5-b76f-d041473ce005" containerName="extract-content" Dec 06 04:05:50 crc kubenswrapper[4802]: I1206 04:05:50.632910 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9ea1538-a9d0-4ca5-b76f-d041473ce005" containerName="registry-server" Dec 06 04:05:50 crc kubenswrapper[4802]: I1206 04:05:50.634299 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dqb9s" Dec 06 04:05:50 crc kubenswrapper[4802]: I1206 04:05:50.645517 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dqb9s"] Dec 06 04:05:50 crc kubenswrapper[4802]: I1206 04:05:50.691689 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b79eed0-a53d-4e9f-913b-3a5b8567a320-catalog-content\") pod \"redhat-operators-dqb9s\" (UID: \"3b79eed0-a53d-4e9f-913b-3a5b8567a320\") " pod="openshift-marketplace/redhat-operators-dqb9s" Dec 06 04:05:50 crc kubenswrapper[4802]: I1206 04:05:50.691736 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b79eed0-a53d-4e9f-913b-3a5b8567a320-utilities\") pod \"redhat-operators-dqb9s\" (UID: \"3b79eed0-a53d-4e9f-913b-3a5b8567a320\") " pod="openshift-marketplace/redhat-operators-dqb9s" Dec 06 04:05:50 crc kubenswrapper[4802]: I1206 04:05:50.691799 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhwtj\" (UniqueName: \"kubernetes.io/projected/3b79eed0-a53d-4e9f-913b-3a5b8567a320-kube-api-access-nhwtj\") pod \"redhat-operators-dqb9s\" (UID: \"3b79eed0-a53d-4e9f-913b-3a5b8567a320\") " pod="openshift-marketplace/redhat-operators-dqb9s" Dec 06 04:05:50 crc kubenswrapper[4802]: I1206 04:05:50.794569 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b79eed0-a53d-4e9f-913b-3a5b8567a320-catalog-content\") pod \"redhat-operators-dqb9s\" (UID: \"3b79eed0-a53d-4e9f-913b-3a5b8567a320\") " pod="openshift-marketplace/redhat-operators-dqb9s" Dec 06 04:05:50 crc kubenswrapper[4802]: I1206 04:05:50.795169 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b79eed0-a53d-4e9f-913b-3a5b8567a320-utilities\") pod \"redhat-operators-dqb9s\" (UID: \"3b79eed0-a53d-4e9f-913b-3a5b8567a320\") " pod="openshift-marketplace/redhat-operators-dqb9s" Dec 06 04:05:50 crc kubenswrapper[4802]: I1206 04:05:50.795351 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhwtj\" (UniqueName: \"kubernetes.io/projected/3b79eed0-a53d-4e9f-913b-3a5b8567a320-kube-api-access-nhwtj\") pod \"redhat-operators-dqb9s\" (UID: \"3b79eed0-a53d-4e9f-913b-3a5b8567a320\") " pod="openshift-marketplace/redhat-operators-dqb9s" Dec 06 04:05:50 crc kubenswrapper[4802]: I1206 04:05:50.795171 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b79eed0-a53d-4e9f-913b-3a5b8567a320-catalog-content\") pod \"redhat-operators-dqb9s\" (UID: \"3b79eed0-a53d-4e9f-913b-3a5b8567a320\") " pod="openshift-marketplace/redhat-operators-dqb9s" Dec 06 04:05:50 crc kubenswrapper[4802]: I1206 04:05:50.795577 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b79eed0-a53d-4e9f-913b-3a5b8567a320-utilities\") pod \"redhat-operators-dqb9s\" (UID: \"3b79eed0-a53d-4e9f-913b-3a5b8567a320\") " pod="openshift-marketplace/redhat-operators-dqb9s" Dec 06 04:05:50 crc kubenswrapper[4802]: I1206 04:05:50.813975 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhwtj\" (UniqueName: \"kubernetes.io/projected/3b79eed0-a53d-4e9f-913b-3a5b8567a320-kube-api-access-nhwtj\") pod \"redhat-operators-dqb9s\" (UID: \"3b79eed0-a53d-4e9f-913b-3a5b8567a320\") " pod="openshift-marketplace/redhat-operators-dqb9s" Dec 06 04:05:50 crc kubenswrapper[4802]: I1206 04:05:50.964899 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dqb9s" Dec 06 04:05:51 crc kubenswrapper[4802]: I1206 04:05:51.498861 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dqb9s"] Dec 06 04:05:52 crc kubenswrapper[4802]: I1206 04:05:52.116408 4802 generic.go:334] "Generic (PLEG): container finished" podID="3b79eed0-a53d-4e9f-913b-3a5b8567a320" containerID="6a75c76bdc6401abd9c0f168420a6c01478287df47bad61e0595f5108dcb9ca7" exitCode=0 Dec 06 04:05:52 crc kubenswrapper[4802]: I1206 04:05:52.116454 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dqb9s" event={"ID":"3b79eed0-a53d-4e9f-913b-3a5b8567a320","Type":"ContainerDied","Data":"6a75c76bdc6401abd9c0f168420a6c01478287df47bad61e0595f5108dcb9ca7"} Dec 06 04:05:52 crc kubenswrapper[4802]: I1206 04:05:52.116676 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dqb9s" event={"ID":"3b79eed0-a53d-4e9f-913b-3a5b8567a320","Type":"ContainerStarted","Data":"df06f13944c9bd9377c2ad84c15e298c96c40c35b83b8df180b6a4d167958162"} Dec 06 04:05:52 crc kubenswrapper[4802]: I1206 04:05:52.118123 4802 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 04:05:53 crc kubenswrapper[4802]: I1206 04:05:53.127799 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dqb9s" event={"ID":"3b79eed0-a53d-4e9f-913b-3a5b8567a320","Type":"ContainerStarted","Data":"21c05cef38ea96e6552b863932b8ea5ef2d621362938b88ea59b9a33d3617e7b"} Dec 06 04:05:54 crc kubenswrapper[4802]: I1206 04:05:54.138056 4802 generic.go:334] "Generic (PLEG): container finished" podID="3b79eed0-a53d-4e9f-913b-3a5b8567a320" containerID="21c05cef38ea96e6552b863932b8ea5ef2d621362938b88ea59b9a33d3617e7b" exitCode=0 Dec 06 04:05:54 crc kubenswrapper[4802]: I1206 04:05:54.138097 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dqb9s" event={"ID":"3b79eed0-a53d-4e9f-913b-3a5b8567a320","Type":"ContainerDied","Data":"21c05cef38ea96e6552b863932b8ea5ef2d621362938b88ea59b9a33d3617e7b"} Dec 06 04:05:54 crc kubenswrapper[4802]: I1206 04:05:54.450368 4802 scope.go:117] "RemoveContainer" containerID="3711adecb81191bc35ac469efa73faeb2346d30f4d5a4a148de8ac88071cdb00" Dec 06 04:05:54 crc kubenswrapper[4802]: E1206 04:05:54.450616 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:05:56 crc kubenswrapper[4802]: I1206 04:05:56.160708 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dqb9s" event={"ID":"3b79eed0-a53d-4e9f-913b-3a5b8567a320","Type":"ContainerStarted","Data":"d36b804a834f1acaf2dbace93837f76899ee1b025231697cc9f8f8a105c1f326"} Dec 06 04:05:56 crc kubenswrapper[4802]: I1206 04:05:56.186856 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dqb9s" podStartSLOduration=3.754350163 podStartE2EDuration="6.186837291s" podCreationTimestamp="2025-12-06 04:05:50 +0000 UTC" firstStartedPulling="2025-12-06 04:05:52.117882586 +0000 UTC m=+1544.989791738" lastFinishedPulling="2025-12-06 04:05:54.550369714 +0000 UTC m=+1547.422278866" observedRunningTime="2025-12-06 04:05:56.181360921 +0000 UTC m=+1549.053270073" watchObservedRunningTime="2025-12-06 04:05:56.186837291 +0000 UTC m=+1549.058746443" Dec 06 04:06:00 crc kubenswrapper[4802]: I1206 04:06:00.966497 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dqb9s" Dec 06 04:06:00 crc kubenswrapper[4802]: I1206 04:06:00.967052 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dqb9s" Dec 06 04:06:02 crc kubenswrapper[4802]: I1206 04:06:02.017605 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dqb9s" podUID="3b79eed0-a53d-4e9f-913b-3a5b8567a320" containerName="registry-server" probeResult="failure" output=< Dec 06 04:06:02 crc kubenswrapper[4802]: timeout: failed to connect service ":50051" within 1s Dec 06 04:06:02 crc kubenswrapper[4802]: > Dec 06 04:06:09 crc kubenswrapper[4802]: I1206 04:06:09.450547 4802 scope.go:117] "RemoveContainer" containerID="3711adecb81191bc35ac469efa73faeb2346d30f4d5a4a148de8ac88071cdb00" Dec 06 04:06:09 crc kubenswrapper[4802]: E1206 04:06:09.451868 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:06:11 crc kubenswrapper[4802]: I1206 04:06:11.013622 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dqb9s" Dec 06 04:06:11 crc kubenswrapper[4802]: I1206 04:06:11.067408 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dqb9s" Dec 06 04:06:11 crc kubenswrapper[4802]: I1206 04:06:11.250980 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dqb9s"] Dec 06 04:06:12 crc kubenswrapper[4802]: I1206 04:06:12.319640 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-dqb9s" podUID="3b79eed0-a53d-4e9f-913b-3a5b8567a320" containerName="registry-server" containerID="cri-o://d36b804a834f1acaf2dbace93837f76899ee1b025231697cc9f8f8a105c1f326" gracePeriod=2 Dec 06 04:06:13 crc kubenswrapper[4802]: I1206 04:06:13.293272 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dqb9s" Dec 06 04:06:13 crc kubenswrapper[4802]: I1206 04:06:13.332821 4802 generic.go:334] "Generic (PLEG): container finished" podID="3b79eed0-a53d-4e9f-913b-3a5b8567a320" containerID="d36b804a834f1acaf2dbace93837f76899ee1b025231697cc9f8f8a105c1f326" exitCode=0 Dec 06 04:06:13 crc kubenswrapper[4802]: I1206 04:06:13.332876 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dqb9s" event={"ID":"3b79eed0-a53d-4e9f-913b-3a5b8567a320","Type":"ContainerDied","Data":"d36b804a834f1acaf2dbace93837f76899ee1b025231697cc9f8f8a105c1f326"} Dec 06 04:06:13 crc kubenswrapper[4802]: I1206 04:06:13.332909 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dqb9s" event={"ID":"3b79eed0-a53d-4e9f-913b-3a5b8567a320","Type":"ContainerDied","Data":"df06f13944c9bd9377c2ad84c15e298c96c40c35b83b8df180b6a4d167958162"} Dec 06 04:06:13 crc kubenswrapper[4802]: I1206 04:06:13.332928 4802 scope.go:117] "RemoveContainer" containerID="d36b804a834f1acaf2dbace93837f76899ee1b025231697cc9f8f8a105c1f326" Dec 06 04:06:13 crc kubenswrapper[4802]: I1206 04:06:13.333098 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dqb9s" Dec 06 04:06:13 crc kubenswrapper[4802]: I1206 04:06:13.334066 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b79eed0-a53d-4e9f-913b-3a5b8567a320-utilities\") pod \"3b79eed0-a53d-4e9f-913b-3a5b8567a320\" (UID: \"3b79eed0-a53d-4e9f-913b-3a5b8567a320\") " Dec 06 04:06:13 crc kubenswrapper[4802]: I1206 04:06:13.335681 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b79eed0-a53d-4e9f-913b-3a5b8567a320-utilities" (OuterVolumeSpecName: "utilities") pod "3b79eed0-a53d-4e9f-913b-3a5b8567a320" (UID: "3b79eed0-a53d-4e9f-913b-3a5b8567a320"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:06:13 crc kubenswrapper[4802]: I1206 04:06:13.335924 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nhwtj\" (UniqueName: \"kubernetes.io/projected/3b79eed0-a53d-4e9f-913b-3a5b8567a320-kube-api-access-nhwtj\") pod \"3b79eed0-a53d-4e9f-913b-3a5b8567a320\" (UID: \"3b79eed0-a53d-4e9f-913b-3a5b8567a320\") " Dec 06 04:06:13 crc kubenswrapper[4802]: I1206 04:06:13.336016 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b79eed0-a53d-4e9f-913b-3a5b8567a320-catalog-content\") pod \"3b79eed0-a53d-4e9f-913b-3a5b8567a320\" (UID: \"3b79eed0-a53d-4e9f-913b-3a5b8567a320\") " Dec 06 04:06:13 crc kubenswrapper[4802]: I1206 04:06:13.336623 4802 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b79eed0-a53d-4e9f-913b-3a5b8567a320-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 04:06:13 crc kubenswrapper[4802]: I1206 04:06:13.346984 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b79eed0-a53d-4e9f-913b-3a5b8567a320-kube-api-access-nhwtj" (OuterVolumeSpecName: "kube-api-access-nhwtj") pod "3b79eed0-a53d-4e9f-913b-3a5b8567a320" (UID: "3b79eed0-a53d-4e9f-913b-3a5b8567a320"). InnerVolumeSpecName "kube-api-access-nhwtj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:06:13 crc kubenswrapper[4802]: I1206 04:06:13.361046 4802 scope.go:117] "RemoveContainer" containerID="21c05cef38ea96e6552b863932b8ea5ef2d621362938b88ea59b9a33d3617e7b" Dec 06 04:06:13 crc kubenswrapper[4802]: I1206 04:06:13.417181 4802 scope.go:117] "RemoveContainer" containerID="6a75c76bdc6401abd9c0f168420a6c01478287df47bad61e0595f5108dcb9ca7" Dec 06 04:06:13 crc kubenswrapper[4802]: I1206 04:06:13.440557 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nhwtj\" (UniqueName: \"kubernetes.io/projected/3b79eed0-a53d-4e9f-913b-3a5b8567a320-kube-api-access-nhwtj\") on node \"crc\" DevicePath \"\"" Dec 06 04:06:13 crc kubenswrapper[4802]: I1206 04:06:13.459319 4802 scope.go:117] "RemoveContainer" containerID="d36b804a834f1acaf2dbace93837f76899ee1b025231697cc9f8f8a105c1f326" Dec 06 04:06:13 crc kubenswrapper[4802]: I1206 04:06:13.461440 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b79eed0-a53d-4e9f-913b-3a5b8567a320-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3b79eed0-a53d-4e9f-913b-3a5b8567a320" (UID: "3b79eed0-a53d-4e9f-913b-3a5b8567a320"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:06:13 crc kubenswrapper[4802]: E1206 04:06:13.461606 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d36b804a834f1acaf2dbace93837f76899ee1b025231697cc9f8f8a105c1f326\": container with ID starting with d36b804a834f1acaf2dbace93837f76899ee1b025231697cc9f8f8a105c1f326 not found: ID does not exist" containerID="d36b804a834f1acaf2dbace93837f76899ee1b025231697cc9f8f8a105c1f326" Dec 06 04:06:13 crc kubenswrapper[4802]: I1206 04:06:13.461650 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d36b804a834f1acaf2dbace93837f76899ee1b025231697cc9f8f8a105c1f326"} err="failed to get container status \"d36b804a834f1acaf2dbace93837f76899ee1b025231697cc9f8f8a105c1f326\": rpc error: code = NotFound desc = could not find container \"d36b804a834f1acaf2dbace93837f76899ee1b025231697cc9f8f8a105c1f326\": container with ID starting with d36b804a834f1acaf2dbace93837f76899ee1b025231697cc9f8f8a105c1f326 not found: ID does not exist" Dec 06 04:06:13 crc kubenswrapper[4802]: I1206 04:06:13.461674 4802 scope.go:117] "RemoveContainer" containerID="21c05cef38ea96e6552b863932b8ea5ef2d621362938b88ea59b9a33d3617e7b" Dec 06 04:06:13 crc kubenswrapper[4802]: E1206 04:06:13.461994 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21c05cef38ea96e6552b863932b8ea5ef2d621362938b88ea59b9a33d3617e7b\": container with ID starting with 21c05cef38ea96e6552b863932b8ea5ef2d621362938b88ea59b9a33d3617e7b not found: ID does not exist" containerID="21c05cef38ea96e6552b863932b8ea5ef2d621362938b88ea59b9a33d3617e7b" Dec 06 04:06:13 crc kubenswrapper[4802]: I1206 04:06:13.462021 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21c05cef38ea96e6552b863932b8ea5ef2d621362938b88ea59b9a33d3617e7b"} err="failed to get container status \"21c05cef38ea96e6552b863932b8ea5ef2d621362938b88ea59b9a33d3617e7b\": rpc error: code = NotFound desc = could not find container \"21c05cef38ea96e6552b863932b8ea5ef2d621362938b88ea59b9a33d3617e7b\": container with ID starting with 21c05cef38ea96e6552b863932b8ea5ef2d621362938b88ea59b9a33d3617e7b not found: ID does not exist" Dec 06 04:06:13 crc kubenswrapper[4802]: I1206 04:06:13.462035 4802 scope.go:117] "RemoveContainer" containerID="6a75c76bdc6401abd9c0f168420a6c01478287df47bad61e0595f5108dcb9ca7" Dec 06 04:06:13 crc kubenswrapper[4802]: E1206 04:06:13.462267 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a75c76bdc6401abd9c0f168420a6c01478287df47bad61e0595f5108dcb9ca7\": container with ID starting with 6a75c76bdc6401abd9c0f168420a6c01478287df47bad61e0595f5108dcb9ca7 not found: ID does not exist" containerID="6a75c76bdc6401abd9c0f168420a6c01478287df47bad61e0595f5108dcb9ca7" Dec 06 04:06:13 crc kubenswrapper[4802]: I1206 04:06:13.462299 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a75c76bdc6401abd9c0f168420a6c01478287df47bad61e0595f5108dcb9ca7"} err="failed to get container status \"6a75c76bdc6401abd9c0f168420a6c01478287df47bad61e0595f5108dcb9ca7\": rpc error: code = NotFound desc = could not find container \"6a75c76bdc6401abd9c0f168420a6c01478287df47bad61e0595f5108dcb9ca7\": container with ID starting with 6a75c76bdc6401abd9c0f168420a6c01478287df47bad61e0595f5108dcb9ca7 not found: ID does not exist" Dec 06 04:06:13 crc kubenswrapper[4802]: I1206 04:06:13.544337 4802 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b79eed0-a53d-4e9f-913b-3a5b8567a320-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 04:06:13 crc kubenswrapper[4802]: I1206 04:06:13.677022 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dqb9s"] Dec 06 04:06:13 crc kubenswrapper[4802]: I1206 04:06:13.688392 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-dqb9s"] Dec 06 04:06:15 crc kubenswrapper[4802]: I1206 04:06:15.461010 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b79eed0-a53d-4e9f-913b-3a5b8567a320" path="/var/lib/kubelet/pods/3b79eed0-a53d-4e9f-913b-3a5b8567a320/volumes" Dec 06 04:06:24 crc kubenswrapper[4802]: I1206 04:06:24.449927 4802 scope.go:117] "RemoveContainer" containerID="3711adecb81191bc35ac469efa73faeb2346d30f4d5a4a148de8ac88071cdb00" Dec 06 04:06:24 crc kubenswrapper[4802]: E1206 04:06:24.450813 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:06:39 crc kubenswrapper[4802]: I1206 04:06:39.450122 4802 scope.go:117] "RemoveContainer" containerID="3711adecb81191bc35ac469efa73faeb2346d30f4d5a4a148de8ac88071cdb00" Dec 06 04:06:39 crc kubenswrapper[4802]: E1206 04:06:39.451121 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:06:54 crc kubenswrapper[4802]: I1206 04:06:54.453706 4802 scope.go:117] "RemoveContainer" containerID="3711adecb81191bc35ac469efa73faeb2346d30f4d5a4a148de8ac88071cdb00" Dec 06 04:06:54 crc kubenswrapper[4802]: E1206 04:06:54.454747 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:07:08 crc kubenswrapper[4802]: I1206 04:07:08.450946 4802 scope.go:117] "RemoveContainer" containerID="3711adecb81191bc35ac469efa73faeb2346d30f4d5a4a148de8ac88071cdb00" Dec 06 04:07:08 crc kubenswrapper[4802]: E1206 04:07:08.453258 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:07:12 crc kubenswrapper[4802]: I1206 04:07:12.277165 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mgf8b"] Dec 06 04:07:12 crc kubenswrapper[4802]: E1206 04:07:12.279247 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b79eed0-a53d-4e9f-913b-3a5b8567a320" containerName="extract-content" Dec 06 04:07:12 crc kubenswrapper[4802]: I1206 04:07:12.279577 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b79eed0-a53d-4e9f-913b-3a5b8567a320" containerName="extract-content" Dec 06 04:07:12 crc kubenswrapper[4802]: E1206 04:07:12.279671 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b79eed0-a53d-4e9f-913b-3a5b8567a320" containerName="registry-server" Dec 06 04:07:12 crc kubenswrapper[4802]: I1206 04:07:12.279769 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b79eed0-a53d-4e9f-913b-3a5b8567a320" containerName="registry-server" Dec 06 04:07:12 crc kubenswrapper[4802]: E1206 04:07:12.279884 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b79eed0-a53d-4e9f-913b-3a5b8567a320" containerName="extract-utilities" Dec 06 04:07:12 crc kubenswrapper[4802]: I1206 04:07:12.279955 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b79eed0-a53d-4e9f-913b-3a5b8567a320" containerName="extract-utilities" Dec 06 04:07:12 crc kubenswrapper[4802]: I1206 04:07:12.280256 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b79eed0-a53d-4e9f-913b-3a5b8567a320" containerName="registry-server" Dec 06 04:07:12 crc kubenswrapper[4802]: I1206 04:07:12.282127 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mgf8b" Dec 06 04:07:12 crc kubenswrapper[4802]: I1206 04:07:12.290648 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mgf8b"] Dec 06 04:07:12 crc kubenswrapper[4802]: I1206 04:07:12.336349 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvk9p\" (UniqueName: \"kubernetes.io/projected/3c5d99f5-e46d-4af3-be5f-48b5f27d2beb-kube-api-access-qvk9p\") pod \"community-operators-mgf8b\" (UID: \"3c5d99f5-e46d-4af3-be5f-48b5f27d2beb\") " pod="openshift-marketplace/community-operators-mgf8b" Dec 06 04:07:12 crc kubenswrapper[4802]: I1206 04:07:12.336672 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c5d99f5-e46d-4af3-be5f-48b5f27d2beb-catalog-content\") pod \"community-operators-mgf8b\" (UID: \"3c5d99f5-e46d-4af3-be5f-48b5f27d2beb\") " pod="openshift-marketplace/community-operators-mgf8b" Dec 06 04:07:12 crc kubenswrapper[4802]: I1206 04:07:12.336802 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c5d99f5-e46d-4af3-be5f-48b5f27d2beb-utilities\") pod \"community-operators-mgf8b\" (UID: \"3c5d99f5-e46d-4af3-be5f-48b5f27d2beb\") " pod="openshift-marketplace/community-operators-mgf8b" Dec 06 04:07:12 crc kubenswrapper[4802]: I1206 04:07:12.438385 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c5d99f5-e46d-4af3-be5f-48b5f27d2beb-catalog-content\") pod \"community-operators-mgf8b\" (UID: \"3c5d99f5-e46d-4af3-be5f-48b5f27d2beb\") " pod="openshift-marketplace/community-operators-mgf8b" Dec 06 04:07:12 crc kubenswrapper[4802]: I1206 04:07:12.438457 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c5d99f5-e46d-4af3-be5f-48b5f27d2beb-utilities\") pod \"community-operators-mgf8b\" (UID: \"3c5d99f5-e46d-4af3-be5f-48b5f27d2beb\") " pod="openshift-marketplace/community-operators-mgf8b" Dec 06 04:07:12 crc kubenswrapper[4802]: I1206 04:07:12.438582 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvk9p\" (UniqueName: \"kubernetes.io/projected/3c5d99f5-e46d-4af3-be5f-48b5f27d2beb-kube-api-access-qvk9p\") pod \"community-operators-mgf8b\" (UID: \"3c5d99f5-e46d-4af3-be5f-48b5f27d2beb\") " pod="openshift-marketplace/community-operators-mgf8b" Dec 06 04:07:12 crc kubenswrapper[4802]: I1206 04:07:12.439073 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c5d99f5-e46d-4af3-be5f-48b5f27d2beb-catalog-content\") pod \"community-operators-mgf8b\" (UID: \"3c5d99f5-e46d-4af3-be5f-48b5f27d2beb\") " pod="openshift-marketplace/community-operators-mgf8b" Dec 06 04:07:12 crc kubenswrapper[4802]: I1206 04:07:12.439313 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c5d99f5-e46d-4af3-be5f-48b5f27d2beb-utilities\") pod \"community-operators-mgf8b\" (UID: \"3c5d99f5-e46d-4af3-be5f-48b5f27d2beb\") " pod="openshift-marketplace/community-operators-mgf8b" Dec 06 04:07:12 crc kubenswrapper[4802]: I1206 04:07:12.460733 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvk9p\" (UniqueName: \"kubernetes.io/projected/3c5d99f5-e46d-4af3-be5f-48b5f27d2beb-kube-api-access-qvk9p\") pod \"community-operators-mgf8b\" (UID: \"3c5d99f5-e46d-4af3-be5f-48b5f27d2beb\") " pod="openshift-marketplace/community-operators-mgf8b" Dec 06 04:07:12 crc kubenswrapper[4802]: I1206 04:07:12.616355 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mgf8b" Dec 06 04:07:13 crc kubenswrapper[4802]: I1206 04:07:13.177304 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mgf8b"] Dec 06 04:07:13 crc kubenswrapper[4802]: W1206 04:07:13.190623 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3c5d99f5_e46d_4af3_be5f_48b5f27d2beb.slice/crio-266da08c17f05546f178fe82f46a760bd72211d0562898aa24716cd142202d61 WatchSource:0}: Error finding container 266da08c17f05546f178fe82f46a760bd72211d0562898aa24716cd142202d61: Status 404 returned error can't find the container with id 266da08c17f05546f178fe82f46a760bd72211d0562898aa24716cd142202d61 Dec 06 04:07:13 crc kubenswrapper[4802]: I1206 04:07:13.967775 4802 generic.go:334] "Generic (PLEG): container finished" podID="3c5d99f5-e46d-4af3-be5f-48b5f27d2beb" containerID="a059274502d40dae0339bac34e0b1ccb53455c518b2f73ba6d1c91b0c3f3d5ce" exitCode=0 Dec 06 04:07:13 crc kubenswrapper[4802]: I1206 04:07:13.967829 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mgf8b" event={"ID":"3c5d99f5-e46d-4af3-be5f-48b5f27d2beb","Type":"ContainerDied","Data":"a059274502d40dae0339bac34e0b1ccb53455c518b2f73ba6d1c91b0c3f3d5ce"} Dec 06 04:07:13 crc kubenswrapper[4802]: I1206 04:07:13.968097 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mgf8b" event={"ID":"3c5d99f5-e46d-4af3-be5f-48b5f27d2beb","Type":"ContainerStarted","Data":"266da08c17f05546f178fe82f46a760bd72211d0562898aa24716cd142202d61"} Dec 06 04:07:15 crc kubenswrapper[4802]: I1206 04:07:15.993911 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mgf8b" event={"ID":"3c5d99f5-e46d-4af3-be5f-48b5f27d2beb","Type":"ContainerStarted","Data":"07beb1ae30b408369972b47089ef37c80389e5083d6f045e22931c08d9a85853"} Dec 06 04:07:17 crc kubenswrapper[4802]: I1206 04:07:17.006193 4802 generic.go:334] "Generic (PLEG): container finished" podID="3c5d99f5-e46d-4af3-be5f-48b5f27d2beb" containerID="07beb1ae30b408369972b47089ef37c80389e5083d6f045e22931c08d9a85853" exitCode=0 Dec 06 04:07:17 crc kubenswrapper[4802]: I1206 04:07:17.006242 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mgf8b" event={"ID":"3c5d99f5-e46d-4af3-be5f-48b5f27d2beb","Type":"ContainerDied","Data":"07beb1ae30b408369972b47089ef37c80389e5083d6f045e22931c08d9a85853"} Dec 06 04:07:19 crc kubenswrapper[4802]: I1206 04:07:19.025576 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mgf8b" event={"ID":"3c5d99f5-e46d-4af3-be5f-48b5f27d2beb","Type":"ContainerStarted","Data":"0207bd4517637824b47e22124079be19eba6ed41da0497779b10784e9a3fabc6"} Dec 06 04:07:19 crc kubenswrapper[4802]: I1206 04:07:19.050956 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mgf8b" podStartSLOduration=3.050082118 podStartE2EDuration="7.050938621s" podCreationTimestamp="2025-12-06 04:07:12 +0000 UTC" firstStartedPulling="2025-12-06 04:07:13.970569336 +0000 UTC m=+1626.842478498" lastFinishedPulling="2025-12-06 04:07:17.971425819 +0000 UTC m=+1630.843335001" observedRunningTime="2025-12-06 04:07:19.042318408 +0000 UTC m=+1631.914227580" watchObservedRunningTime="2025-12-06 04:07:19.050938621 +0000 UTC m=+1631.922847773" Dec 06 04:07:19 crc kubenswrapper[4802]: I1206 04:07:19.926689 4802 scope.go:117] "RemoveContainer" containerID="75b98ba7c36b2ecb2ceea7d9979de123c001191248f80acbc51471e2e6cc98d8" Dec 06 04:07:19 crc kubenswrapper[4802]: I1206 04:07:19.947087 4802 scope.go:117] "RemoveContainer" containerID="d9b46c170bb4e57da13c1765cb60284b4284b5215be0ba5b64dcc3e4557574eb" Dec 06 04:07:19 crc kubenswrapper[4802]: I1206 04:07:19.972378 4802 scope.go:117] "RemoveContainer" containerID="3ab1775540cb21c804ef34451616e8edc0e7a1f035e391f33df05e801c645e4d" Dec 06 04:07:19 crc kubenswrapper[4802]: I1206 04:07:19.998914 4802 scope.go:117] "RemoveContainer" containerID="cb11fba1b171ed3e1a85133de81039e2e67c9cff821abaecbe190a62b89fa7c2" Dec 06 04:07:20 crc kubenswrapper[4802]: I1206 04:07:20.449988 4802 scope.go:117] "RemoveContainer" containerID="3711adecb81191bc35ac469efa73faeb2346d30f4d5a4a148de8ac88071cdb00" Dec 06 04:07:20 crc kubenswrapper[4802]: E1206 04:07:20.450234 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:07:22 crc kubenswrapper[4802]: I1206 04:07:22.616823 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mgf8b" Dec 06 04:07:22 crc kubenswrapper[4802]: I1206 04:07:22.617182 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mgf8b" Dec 06 04:07:22 crc kubenswrapper[4802]: I1206 04:07:22.700690 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mgf8b" Dec 06 04:07:23 crc kubenswrapper[4802]: I1206 04:07:23.138803 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mgf8b" Dec 06 04:07:23 crc kubenswrapper[4802]: I1206 04:07:23.201734 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mgf8b"] Dec 06 04:07:25 crc kubenswrapper[4802]: I1206 04:07:25.104601 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-mgf8b" podUID="3c5d99f5-e46d-4af3-be5f-48b5f27d2beb" containerName="registry-server" containerID="cri-o://0207bd4517637824b47e22124079be19eba6ed41da0497779b10784e9a3fabc6" gracePeriod=2 Dec 06 04:07:26 crc kubenswrapper[4802]: I1206 04:07:26.115623 4802 generic.go:334] "Generic (PLEG): container finished" podID="3c5d99f5-e46d-4af3-be5f-48b5f27d2beb" containerID="0207bd4517637824b47e22124079be19eba6ed41da0497779b10784e9a3fabc6" exitCode=0 Dec 06 04:07:26 crc kubenswrapper[4802]: I1206 04:07:26.115890 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mgf8b" event={"ID":"3c5d99f5-e46d-4af3-be5f-48b5f27d2beb","Type":"ContainerDied","Data":"0207bd4517637824b47e22124079be19eba6ed41da0497779b10784e9a3fabc6"} Dec 06 04:07:26 crc kubenswrapper[4802]: I1206 04:07:26.199197 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mgf8b" Dec 06 04:07:26 crc kubenswrapper[4802]: I1206 04:07:26.216517 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c5d99f5-e46d-4af3-be5f-48b5f27d2beb-catalog-content\") pod \"3c5d99f5-e46d-4af3-be5f-48b5f27d2beb\" (UID: \"3c5d99f5-e46d-4af3-be5f-48b5f27d2beb\") " Dec 06 04:07:26 crc kubenswrapper[4802]: I1206 04:07:26.216591 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qvk9p\" (UniqueName: \"kubernetes.io/projected/3c5d99f5-e46d-4af3-be5f-48b5f27d2beb-kube-api-access-qvk9p\") pod \"3c5d99f5-e46d-4af3-be5f-48b5f27d2beb\" (UID: \"3c5d99f5-e46d-4af3-be5f-48b5f27d2beb\") " Dec 06 04:07:26 crc kubenswrapper[4802]: I1206 04:07:26.216694 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c5d99f5-e46d-4af3-be5f-48b5f27d2beb-utilities\") pod \"3c5d99f5-e46d-4af3-be5f-48b5f27d2beb\" (UID: \"3c5d99f5-e46d-4af3-be5f-48b5f27d2beb\") " Dec 06 04:07:26 crc kubenswrapper[4802]: I1206 04:07:26.217488 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c5d99f5-e46d-4af3-be5f-48b5f27d2beb-utilities" (OuterVolumeSpecName: "utilities") pod "3c5d99f5-e46d-4af3-be5f-48b5f27d2beb" (UID: "3c5d99f5-e46d-4af3-be5f-48b5f27d2beb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:07:26 crc kubenswrapper[4802]: I1206 04:07:26.230947 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c5d99f5-e46d-4af3-be5f-48b5f27d2beb-kube-api-access-qvk9p" (OuterVolumeSpecName: "kube-api-access-qvk9p") pod "3c5d99f5-e46d-4af3-be5f-48b5f27d2beb" (UID: "3c5d99f5-e46d-4af3-be5f-48b5f27d2beb"). InnerVolumeSpecName "kube-api-access-qvk9p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:26 crc kubenswrapper[4802]: I1206 04:07:26.277739 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c5d99f5-e46d-4af3-be5f-48b5f27d2beb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3c5d99f5-e46d-4af3-be5f-48b5f27d2beb" (UID: "3c5d99f5-e46d-4af3-be5f-48b5f27d2beb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:07:26 crc kubenswrapper[4802]: I1206 04:07:26.318836 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qvk9p\" (UniqueName: \"kubernetes.io/projected/3c5d99f5-e46d-4af3-be5f-48b5f27d2beb-kube-api-access-qvk9p\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:26 crc kubenswrapper[4802]: I1206 04:07:26.318872 4802 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c5d99f5-e46d-4af3-be5f-48b5f27d2beb-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:26 crc kubenswrapper[4802]: I1206 04:07:26.318883 4802 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c5d99f5-e46d-4af3-be5f-48b5f27d2beb-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:27 crc kubenswrapper[4802]: I1206 04:07:27.132061 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mgf8b" event={"ID":"3c5d99f5-e46d-4af3-be5f-48b5f27d2beb","Type":"ContainerDied","Data":"266da08c17f05546f178fe82f46a760bd72211d0562898aa24716cd142202d61"} Dec 06 04:07:27 crc kubenswrapper[4802]: I1206 04:07:27.132131 4802 scope.go:117] "RemoveContainer" containerID="0207bd4517637824b47e22124079be19eba6ed41da0497779b10784e9a3fabc6" Dec 06 04:07:27 crc kubenswrapper[4802]: I1206 04:07:27.132169 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mgf8b" Dec 06 04:07:27 crc kubenswrapper[4802]: I1206 04:07:27.161504 4802 scope.go:117] "RemoveContainer" containerID="07beb1ae30b408369972b47089ef37c80389e5083d6f045e22931c08d9a85853" Dec 06 04:07:27 crc kubenswrapper[4802]: I1206 04:07:27.187047 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mgf8b"] Dec 06 04:07:27 crc kubenswrapper[4802]: I1206 04:07:27.202704 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-mgf8b"] Dec 06 04:07:27 crc kubenswrapper[4802]: I1206 04:07:27.231199 4802 scope.go:117] "RemoveContainer" containerID="a059274502d40dae0339bac34e0b1ccb53455c518b2f73ba6d1c91b0c3f3d5ce" Dec 06 04:07:27 crc kubenswrapper[4802]: I1206 04:07:27.476468 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c5d99f5-e46d-4af3-be5f-48b5f27d2beb" path="/var/lib/kubelet/pods/3c5d99f5-e46d-4af3-be5f-48b5f27d2beb/volumes" Dec 06 04:07:29 crc kubenswrapper[4802]: I1206 04:07:29.039806 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-49mm8"] Dec 06 04:07:29 crc kubenswrapper[4802]: I1206 04:07:29.052225 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-e5f1-account-create-update-zxdsz"] Dec 06 04:07:29 crc kubenswrapper[4802]: I1206 04:07:29.073174 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-e5f1-account-create-update-zxdsz"] Dec 06 04:07:29 crc kubenswrapper[4802]: I1206 04:07:29.084652 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-49mm8"] Dec 06 04:07:29 crc kubenswrapper[4802]: I1206 04:07:29.465571 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19fbf090-979f-4d84-b774-c9cf98ea1501" path="/var/lib/kubelet/pods/19fbf090-979f-4d84-b774-c9cf98ea1501/volumes" Dec 06 04:07:29 crc kubenswrapper[4802]: I1206 04:07:29.466448 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb7eaedf-a7ed-4c1f-8609-456afd84daa6" path="/var/lib/kubelet/pods/fb7eaedf-a7ed-4c1f-8609-456afd84daa6/volumes" Dec 06 04:07:30 crc kubenswrapper[4802]: I1206 04:07:30.041243 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-l2tpj"] Dec 06 04:07:30 crc kubenswrapper[4802]: I1206 04:07:30.051890 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-633c-account-create-update-ld6ht"] Dec 06 04:07:30 crc kubenswrapper[4802]: I1206 04:07:30.061554 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-l2tpj"] Dec 06 04:07:30 crc kubenswrapper[4802]: I1206 04:07:30.069866 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-633c-account-create-update-ld6ht"] Dec 06 04:07:30 crc kubenswrapper[4802]: I1206 04:07:30.080397 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-ckg82"] Dec 06 04:07:30 crc kubenswrapper[4802]: I1206 04:07:30.090459 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-785f-account-create-update-4x7n2"] Dec 06 04:07:30 crc kubenswrapper[4802]: I1206 04:07:30.098681 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-ckg82"] Dec 06 04:07:30 crc kubenswrapper[4802]: I1206 04:07:30.107465 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-785f-account-create-update-4x7n2"] Dec 06 04:07:31 crc kubenswrapper[4802]: I1206 04:07:31.450877 4802 scope.go:117] "RemoveContainer" containerID="3711adecb81191bc35ac469efa73faeb2346d30f4d5a4a148de8ac88071cdb00" Dec 06 04:07:31 crc kubenswrapper[4802]: E1206 04:07:31.451356 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:07:31 crc kubenswrapper[4802]: I1206 04:07:31.463527 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76accf7a-8aeb-48a1-88c5-3b64e8a0d25e" path="/var/lib/kubelet/pods/76accf7a-8aeb-48a1-88c5-3b64e8a0d25e/volumes" Dec 06 04:07:31 crc kubenswrapper[4802]: I1206 04:07:31.464368 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a4a22d0-ccb1-40ec-9872-05b41d788fd4" path="/var/lib/kubelet/pods/7a4a22d0-ccb1-40ec-9872-05b41d788fd4/volumes" Dec 06 04:07:31 crc kubenswrapper[4802]: I1206 04:07:31.465107 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2ab2ae8-12da-43f3-bcd8-1de34185d285" path="/var/lib/kubelet/pods/a2ab2ae8-12da-43f3-bcd8-1de34185d285/volumes" Dec 06 04:07:31 crc kubenswrapper[4802]: I1206 04:07:31.465843 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4efbdc9-de14-4774-9c69-66b4cb0f79b9" path="/var/lib/kubelet/pods/b4efbdc9-de14-4774-9c69-66b4cb0f79b9/volumes" Dec 06 04:07:42 crc kubenswrapper[4802]: I1206 04:07:42.450347 4802 scope.go:117] "RemoveContainer" containerID="3711adecb81191bc35ac469efa73faeb2346d30f4d5a4a148de8ac88071cdb00" Dec 06 04:07:42 crc kubenswrapper[4802]: E1206 04:07:42.452278 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:07:57 crc kubenswrapper[4802]: I1206 04:07:57.466750 4802 scope.go:117] "RemoveContainer" containerID="3711adecb81191bc35ac469efa73faeb2346d30f4d5a4a148de8ac88071cdb00" Dec 06 04:07:57 crc kubenswrapper[4802]: E1206 04:07:57.467613 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:08:09 crc kubenswrapper[4802]: I1206 04:08:09.450259 4802 scope.go:117] "RemoveContainer" containerID="3711adecb81191bc35ac469efa73faeb2346d30f4d5a4a148de8ac88071cdb00" Dec 06 04:08:09 crc kubenswrapper[4802]: E1206 04:08:09.451105 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:08:09 crc kubenswrapper[4802]: I1206 04:08:09.550680 4802 generic.go:334] "Generic (PLEG): container finished" podID="a07f3a26-501d-49cb-8dbf-648f1cac66da" containerID="984316e8889de93db585481aa2ce837159302ff37f09380b6a58357681518637" exitCode=0 Dec 06 04:08:09 crc kubenswrapper[4802]: I1206 04:08:09.550735 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-n5lkw" event={"ID":"a07f3a26-501d-49cb-8dbf-648f1cac66da","Type":"ContainerDied","Data":"984316e8889de93db585481aa2ce837159302ff37f09380b6a58357681518637"} Dec 06 04:08:11 crc kubenswrapper[4802]: I1206 04:08:11.032996 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-n5lkw" Dec 06 04:08:11 crc kubenswrapper[4802]: I1206 04:08:11.125973 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a07f3a26-501d-49cb-8dbf-648f1cac66da-ssh-key\") pod \"a07f3a26-501d-49cb-8dbf-648f1cac66da\" (UID: \"a07f3a26-501d-49cb-8dbf-648f1cac66da\") " Dec 06 04:08:11 crc kubenswrapper[4802]: I1206 04:08:11.126347 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnbxt\" (UniqueName: \"kubernetes.io/projected/a07f3a26-501d-49cb-8dbf-648f1cac66da-kube-api-access-rnbxt\") pod \"a07f3a26-501d-49cb-8dbf-648f1cac66da\" (UID: \"a07f3a26-501d-49cb-8dbf-648f1cac66da\") " Dec 06 04:08:11 crc kubenswrapper[4802]: I1206 04:08:11.126415 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a07f3a26-501d-49cb-8dbf-648f1cac66da-inventory\") pod \"a07f3a26-501d-49cb-8dbf-648f1cac66da\" (UID: \"a07f3a26-501d-49cb-8dbf-648f1cac66da\") " Dec 06 04:08:11 crc kubenswrapper[4802]: I1206 04:08:11.126628 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a07f3a26-501d-49cb-8dbf-648f1cac66da-bootstrap-combined-ca-bundle\") pod \"a07f3a26-501d-49cb-8dbf-648f1cac66da\" (UID: \"a07f3a26-501d-49cb-8dbf-648f1cac66da\") " Dec 06 04:08:11 crc kubenswrapper[4802]: I1206 04:08:11.135044 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a07f3a26-501d-49cb-8dbf-648f1cac66da-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "a07f3a26-501d-49cb-8dbf-648f1cac66da" (UID: "a07f3a26-501d-49cb-8dbf-648f1cac66da"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:08:11 crc kubenswrapper[4802]: I1206 04:08:11.135271 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a07f3a26-501d-49cb-8dbf-648f1cac66da-kube-api-access-rnbxt" (OuterVolumeSpecName: "kube-api-access-rnbxt") pod "a07f3a26-501d-49cb-8dbf-648f1cac66da" (UID: "a07f3a26-501d-49cb-8dbf-648f1cac66da"). InnerVolumeSpecName "kube-api-access-rnbxt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:08:11 crc kubenswrapper[4802]: I1206 04:08:11.169285 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a07f3a26-501d-49cb-8dbf-648f1cac66da-inventory" (OuterVolumeSpecName: "inventory") pod "a07f3a26-501d-49cb-8dbf-648f1cac66da" (UID: "a07f3a26-501d-49cb-8dbf-648f1cac66da"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:08:11 crc kubenswrapper[4802]: I1206 04:08:11.172792 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a07f3a26-501d-49cb-8dbf-648f1cac66da-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a07f3a26-501d-49cb-8dbf-648f1cac66da" (UID: "a07f3a26-501d-49cb-8dbf-648f1cac66da"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:08:11 crc kubenswrapper[4802]: I1206 04:08:11.228817 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnbxt\" (UniqueName: \"kubernetes.io/projected/a07f3a26-501d-49cb-8dbf-648f1cac66da-kube-api-access-rnbxt\") on node \"crc\" DevicePath \"\"" Dec 06 04:08:11 crc kubenswrapper[4802]: I1206 04:08:11.228865 4802 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a07f3a26-501d-49cb-8dbf-648f1cac66da-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 04:08:11 crc kubenswrapper[4802]: I1206 04:08:11.228880 4802 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a07f3a26-501d-49cb-8dbf-648f1cac66da-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:08:11 crc kubenswrapper[4802]: I1206 04:08:11.228891 4802 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a07f3a26-501d-49cb-8dbf-648f1cac66da-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 04:08:11 crc kubenswrapper[4802]: I1206 04:08:11.570975 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-n5lkw" event={"ID":"a07f3a26-501d-49cb-8dbf-648f1cac66da","Type":"ContainerDied","Data":"c696b288132f6ecc51dd2dc8ff43b41280a9bbbc41f1ec8132ab7cefcbcd47fd"} Dec 06 04:08:11 crc kubenswrapper[4802]: I1206 04:08:11.571018 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c696b288132f6ecc51dd2dc8ff43b41280a9bbbc41f1ec8132ab7cefcbcd47fd" Dec 06 04:08:11 crc kubenswrapper[4802]: I1206 04:08:11.571053 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-n5lkw" Dec 06 04:08:11 crc kubenswrapper[4802]: I1206 04:08:11.659522 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgxfp"] Dec 06 04:08:11 crc kubenswrapper[4802]: E1206 04:08:11.660001 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c5d99f5-e46d-4af3-be5f-48b5f27d2beb" containerName="extract-content" Dec 06 04:08:11 crc kubenswrapper[4802]: I1206 04:08:11.660024 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c5d99f5-e46d-4af3-be5f-48b5f27d2beb" containerName="extract-content" Dec 06 04:08:11 crc kubenswrapper[4802]: E1206 04:08:11.660045 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a07f3a26-501d-49cb-8dbf-648f1cac66da" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 06 04:08:11 crc kubenswrapper[4802]: I1206 04:08:11.660054 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="a07f3a26-501d-49cb-8dbf-648f1cac66da" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 06 04:08:11 crc kubenswrapper[4802]: E1206 04:08:11.660076 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c5d99f5-e46d-4af3-be5f-48b5f27d2beb" containerName="extract-utilities" Dec 06 04:08:11 crc kubenswrapper[4802]: I1206 04:08:11.660084 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c5d99f5-e46d-4af3-be5f-48b5f27d2beb" containerName="extract-utilities" Dec 06 04:08:11 crc kubenswrapper[4802]: E1206 04:08:11.660098 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c5d99f5-e46d-4af3-be5f-48b5f27d2beb" containerName="registry-server" Dec 06 04:08:11 crc kubenswrapper[4802]: I1206 04:08:11.660106 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c5d99f5-e46d-4af3-be5f-48b5f27d2beb" containerName="registry-server" Dec 06 04:08:11 crc kubenswrapper[4802]: I1206 04:08:11.660329 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c5d99f5-e46d-4af3-be5f-48b5f27d2beb" containerName="registry-server" Dec 06 04:08:11 crc kubenswrapper[4802]: I1206 04:08:11.660350 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="a07f3a26-501d-49cb-8dbf-648f1cac66da" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 06 04:08:11 crc kubenswrapper[4802]: I1206 04:08:11.661115 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgxfp" Dec 06 04:08:11 crc kubenswrapper[4802]: I1206 04:08:11.663565 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 04:08:11 crc kubenswrapper[4802]: I1206 04:08:11.663853 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 04:08:11 crc kubenswrapper[4802]: I1206 04:08:11.663930 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 04:08:11 crc kubenswrapper[4802]: I1206 04:08:11.664247 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-rgvjc" Dec 06 04:08:11 crc kubenswrapper[4802]: I1206 04:08:11.674059 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgxfp"] Dec 06 04:08:11 crc kubenswrapper[4802]: I1206 04:08:11.737833 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvcsc\" (UniqueName: \"kubernetes.io/projected/db0df1b2-ef63-42ae-96ff-cafa00acc8e1-kube-api-access-lvcsc\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-sgxfp\" (UID: \"db0df1b2-ef63-42ae-96ff-cafa00acc8e1\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgxfp" Dec 06 04:08:11 crc kubenswrapper[4802]: I1206 04:08:11.738058 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/db0df1b2-ef63-42ae-96ff-cafa00acc8e1-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-sgxfp\" (UID: \"db0df1b2-ef63-42ae-96ff-cafa00acc8e1\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgxfp" Dec 06 04:08:11 crc kubenswrapper[4802]: I1206 04:08:11.738132 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/db0df1b2-ef63-42ae-96ff-cafa00acc8e1-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-sgxfp\" (UID: \"db0df1b2-ef63-42ae-96ff-cafa00acc8e1\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgxfp" Dec 06 04:08:11 crc kubenswrapper[4802]: I1206 04:08:11.839644 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvcsc\" (UniqueName: \"kubernetes.io/projected/db0df1b2-ef63-42ae-96ff-cafa00acc8e1-kube-api-access-lvcsc\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-sgxfp\" (UID: \"db0df1b2-ef63-42ae-96ff-cafa00acc8e1\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgxfp" Dec 06 04:08:11 crc kubenswrapper[4802]: I1206 04:08:11.839750 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/db0df1b2-ef63-42ae-96ff-cafa00acc8e1-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-sgxfp\" (UID: \"db0df1b2-ef63-42ae-96ff-cafa00acc8e1\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgxfp" Dec 06 04:08:11 crc kubenswrapper[4802]: I1206 04:08:11.839822 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/db0df1b2-ef63-42ae-96ff-cafa00acc8e1-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-sgxfp\" (UID: \"db0df1b2-ef63-42ae-96ff-cafa00acc8e1\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgxfp" Dec 06 04:08:11 crc kubenswrapper[4802]: I1206 04:08:11.846385 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/db0df1b2-ef63-42ae-96ff-cafa00acc8e1-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-sgxfp\" (UID: \"db0df1b2-ef63-42ae-96ff-cafa00acc8e1\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgxfp" Dec 06 04:08:11 crc kubenswrapper[4802]: I1206 04:08:11.846514 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/db0df1b2-ef63-42ae-96ff-cafa00acc8e1-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-sgxfp\" (UID: \"db0df1b2-ef63-42ae-96ff-cafa00acc8e1\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgxfp" Dec 06 04:08:11 crc kubenswrapper[4802]: I1206 04:08:11.860645 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvcsc\" (UniqueName: \"kubernetes.io/projected/db0df1b2-ef63-42ae-96ff-cafa00acc8e1-kube-api-access-lvcsc\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-sgxfp\" (UID: \"db0df1b2-ef63-42ae-96ff-cafa00acc8e1\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgxfp" Dec 06 04:08:11 crc kubenswrapper[4802]: I1206 04:08:11.980289 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgxfp" Dec 06 04:08:12 crc kubenswrapper[4802]: I1206 04:08:12.565937 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgxfp"] Dec 06 04:08:12 crc kubenswrapper[4802]: W1206 04:08:12.568745 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddb0df1b2_ef63_42ae_96ff_cafa00acc8e1.slice/crio-aefa38e3b584b0bf6dcbe0ad2944c9a9672e80445c6af98c5d4e6b6cd0f66f53 WatchSource:0}: Error finding container aefa38e3b584b0bf6dcbe0ad2944c9a9672e80445c6af98c5d4e6b6cd0f66f53: Status 404 returned error can't find the container with id aefa38e3b584b0bf6dcbe0ad2944c9a9672e80445c6af98c5d4e6b6cd0f66f53 Dec 06 04:08:12 crc kubenswrapper[4802]: I1206 04:08:12.584220 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgxfp" event={"ID":"db0df1b2-ef63-42ae-96ff-cafa00acc8e1","Type":"ContainerStarted","Data":"aefa38e3b584b0bf6dcbe0ad2944c9a9672e80445c6af98c5d4e6b6cd0f66f53"} Dec 06 04:08:13 crc kubenswrapper[4802]: I1206 04:08:13.595288 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgxfp" event={"ID":"db0df1b2-ef63-42ae-96ff-cafa00acc8e1","Type":"ContainerStarted","Data":"a140b688777948d25cd90c0f584854dad6ee08e220dd1bda00ea073adf871131"} Dec 06 04:08:13 crc kubenswrapper[4802]: I1206 04:08:13.622085 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgxfp" podStartSLOduration=2.206939427 podStartE2EDuration="2.622065184s" podCreationTimestamp="2025-12-06 04:08:11 +0000 UTC" firstStartedPulling="2025-12-06 04:08:12.570862139 +0000 UTC m=+1685.442771291" lastFinishedPulling="2025-12-06 04:08:12.985987896 +0000 UTC m=+1685.857897048" observedRunningTime="2025-12-06 04:08:13.618744504 +0000 UTC m=+1686.490653656" watchObservedRunningTime="2025-12-06 04:08:13.622065184 +0000 UTC m=+1686.493974336" Dec 06 04:08:20 crc kubenswrapper[4802]: I1206 04:08:20.067864 4802 scope.go:117] "RemoveContainer" containerID="1fd6bec64ef8692b5a1e61e69e2d0eed99d6c95b54133c5bdf95b50381d88e38" Dec 06 04:08:20 crc kubenswrapper[4802]: I1206 04:08:20.097930 4802 scope.go:117] "RemoveContainer" containerID="f67ec7c76f73d0c62d661119f023a38670154a30378ef57061af626c7fd1edef" Dec 06 04:08:20 crc kubenswrapper[4802]: I1206 04:08:20.160249 4802 scope.go:117] "RemoveContainer" containerID="41aae8e61a7c611287056fd7264028019934d014219736f29d2f668182695dce" Dec 06 04:08:20 crc kubenswrapper[4802]: I1206 04:08:20.215409 4802 scope.go:117] "RemoveContainer" containerID="b6b1e5ff1fdbce8fcfff793b384cfd348effcdfc95f8a10a61794c9b801f0ddf" Dec 06 04:08:20 crc kubenswrapper[4802]: I1206 04:08:20.260514 4802 scope.go:117] "RemoveContainer" containerID="2770947b36cbd725b7ffe56bbc6f88a033ccb57bda0f1825d49e4634b4aac59c" Dec 06 04:08:20 crc kubenswrapper[4802]: I1206 04:08:20.289839 4802 scope.go:117] "RemoveContainer" containerID="a12f5de68366599bca0adb3315e75c4a95cce952726a6d08fce6030738b0db53" Dec 06 04:08:20 crc kubenswrapper[4802]: I1206 04:08:20.328529 4802 scope.go:117] "RemoveContainer" containerID="b59967445aad7babb9789c7d2853dc467dec4269e482d759935beb41a5ee1358" Dec 06 04:08:20 crc kubenswrapper[4802]: I1206 04:08:20.350940 4802 scope.go:117] "RemoveContainer" containerID="067fef1bf0135ad1b23802007a5d041456aea47d26cc06bd9c2d118a45dee413" Dec 06 04:08:24 crc kubenswrapper[4802]: I1206 04:08:24.450403 4802 scope.go:117] "RemoveContainer" containerID="3711adecb81191bc35ac469efa73faeb2346d30f4d5a4a148de8ac88071cdb00" Dec 06 04:08:24 crc kubenswrapper[4802]: E1206 04:08:24.451286 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:08:28 crc kubenswrapper[4802]: I1206 04:08:28.052783 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-1352-account-create-update-jhtbv"] Dec 06 04:08:28 crc kubenswrapper[4802]: I1206 04:08:28.069437 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-e48e-account-create-update-zw8gt"] Dec 06 04:08:28 crc kubenswrapper[4802]: I1206 04:08:28.081518 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-create-gdjlc"] Dec 06 04:08:28 crc kubenswrapper[4802]: I1206 04:08:28.090428 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-e48e-account-create-update-zw8gt"] Dec 06 04:08:28 crc kubenswrapper[4802]: I1206 04:08:28.099415 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-1352-account-create-update-jhtbv"] Dec 06 04:08:28 crc kubenswrapper[4802]: I1206 04:08:28.111664 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-create-gdjlc"] Dec 06 04:08:29 crc kubenswrapper[4802]: I1206 04:08:29.051887 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-sxrmq"] Dec 06 04:08:29 crc kubenswrapper[4802]: I1206 04:08:29.071926 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-da9d-account-create-update-6t8rq"] Dec 06 04:08:29 crc kubenswrapper[4802]: I1206 04:08:29.083030 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-ec26-account-create-update-fxz49"] Dec 06 04:08:29 crc kubenswrapper[4802]: I1206 04:08:29.092255 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-sxrmq"] Dec 06 04:08:29 crc kubenswrapper[4802]: I1206 04:08:29.100218 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-nv6kh"] Dec 06 04:08:29 crc kubenswrapper[4802]: I1206 04:08:29.107307 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-da9d-account-create-update-6t8rq"] Dec 06 04:08:29 crc kubenswrapper[4802]: I1206 04:08:29.115711 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-ec26-account-create-update-fxz49"] Dec 06 04:08:29 crc kubenswrapper[4802]: I1206 04:08:29.123653 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-nv6kh"] Dec 06 04:08:29 crc kubenswrapper[4802]: I1206 04:08:29.134138 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-v5t59"] Dec 06 04:08:29 crc kubenswrapper[4802]: I1206 04:08:29.144040 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-v5t59"] Dec 06 04:08:29 crc kubenswrapper[4802]: I1206 04:08:29.461102 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c4bfa7b-7d20-4c15-b4a7-181a05dcaab4" path="/var/lib/kubelet/pods/0c4bfa7b-7d20-4c15-b4a7-181a05dcaab4/volumes" Dec 06 04:08:29 crc kubenswrapper[4802]: I1206 04:08:29.461897 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="180473dd-14d1-409d-8ca8-059c9883c762" path="/var/lib/kubelet/pods/180473dd-14d1-409d-8ca8-059c9883c762/volumes" Dec 06 04:08:29 crc kubenswrapper[4802]: I1206 04:08:29.462604 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7653cfc2-6e34-40f2-843a-9a644165b0fb" path="/var/lib/kubelet/pods/7653cfc2-6e34-40f2-843a-9a644165b0fb/volumes" Dec 06 04:08:29 crc kubenswrapper[4802]: I1206 04:08:29.463360 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a076eb84-b8f5-4f75-922b-e44682ad8e67" path="/var/lib/kubelet/pods/a076eb84-b8f5-4f75-922b-e44682ad8e67/volumes" Dec 06 04:08:29 crc kubenswrapper[4802]: I1206 04:08:29.464720 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7d36678-2b9e-4380-8828-a32e159ce604" path="/var/lib/kubelet/pods/c7d36678-2b9e-4380-8828-a32e159ce604/volumes" Dec 06 04:08:29 crc kubenswrapper[4802]: I1206 04:08:29.465371 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e065079d-0d3d-40b5-8103-556cb4b7f338" path="/var/lib/kubelet/pods/e065079d-0d3d-40b5-8103-556cb4b7f338/volumes" Dec 06 04:08:29 crc kubenswrapper[4802]: I1206 04:08:29.466037 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e975aca2-19e6-4a92-a4b0-f0db131bdfec" path="/var/lib/kubelet/pods/e975aca2-19e6-4a92-a4b0-f0db131bdfec/volumes" Dec 06 04:08:29 crc kubenswrapper[4802]: I1206 04:08:29.467323 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ebeea130-3942-475e-ba17-51d624cde585" path="/var/lib/kubelet/pods/ebeea130-3942-475e-ba17-51d624cde585/volumes" Dec 06 04:08:37 crc kubenswrapper[4802]: I1206 04:08:37.035907 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-mrxg9"] Dec 06 04:08:37 crc kubenswrapper[4802]: I1206 04:08:37.047692 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-mrxg9"] Dec 06 04:08:37 crc kubenswrapper[4802]: I1206 04:08:37.459208 4802 scope.go:117] "RemoveContainer" containerID="3711adecb81191bc35ac469efa73faeb2346d30f4d5a4a148de8ac88071cdb00" Dec 06 04:08:37 crc kubenswrapper[4802]: E1206 04:08:37.459954 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:08:37 crc kubenswrapper[4802]: I1206 04:08:37.465704 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6f5b891-75b4-4b8c-95b1-7b92ee267793" path="/var/lib/kubelet/pods/c6f5b891-75b4-4b8c-95b1-7b92ee267793/volumes" Dec 06 04:08:38 crc kubenswrapper[4802]: I1206 04:08:38.035776 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-mlhfp"] Dec 06 04:08:38 crc kubenswrapper[4802]: I1206 04:08:38.047798 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-mlhfp"] Dec 06 04:08:39 crc kubenswrapper[4802]: I1206 04:08:39.465846 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d947661-37cd-49c2-bd89-e927e4034ba4" path="/var/lib/kubelet/pods/5d947661-37cd-49c2-bd89-e927e4034ba4/volumes" Dec 06 04:08:52 crc kubenswrapper[4802]: I1206 04:08:52.450927 4802 scope.go:117] "RemoveContainer" containerID="3711adecb81191bc35ac469efa73faeb2346d30f4d5a4a148de8ac88071cdb00" Dec 06 04:08:52 crc kubenswrapper[4802]: E1206 04:08:52.451855 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:09:03 crc kubenswrapper[4802]: I1206 04:09:03.450382 4802 scope.go:117] "RemoveContainer" containerID="3711adecb81191bc35ac469efa73faeb2346d30f4d5a4a148de8ac88071cdb00" Dec 06 04:09:03 crc kubenswrapper[4802]: E1206 04:09:03.451116 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:09:16 crc kubenswrapper[4802]: I1206 04:09:16.450200 4802 scope.go:117] "RemoveContainer" containerID="3711adecb81191bc35ac469efa73faeb2346d30f4d5a4a148de8ac88071cdb00" Dec 06 04:09:16 crc kubenswrapper[4802]: E1206 04:09:16.451076 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:09:18 crc kubenswrapper[4802]: I1206 04:09:18.052640 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-twnc4"] Dec 06 04:09:18 crc kubenswrapper[4802]: I1206 04:09:18.064025 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-twnc4"] Dec 06 04:09:19 crc kubenswrapper[4802]: I1206 04:09:19.038677 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-dxvz6"] Dec 06 04:09:19 crc kubenswrapper[4802]: I1206 04:09:19.048232 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-dxvz6"] Dec 06 04:09:19 crc kubenswrapper[4802]: I1206 04:09:19.459499 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4abd9ed3-07e6-46a6-8539-85fc5c624145" path="/var/lib/kubelet/pods/4abd9ed3-07e6-46a6-8539-85fc5c624145/volumes" Dec 06 04:09:19 crc kubenswrapper[4802]: I1206 04:09:19.460571 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f33c46c6-6721-4c7a-a602-48351824838a" path="/var/lib/kubelet/pods/f33c46c6-6721-4c7a-a602-48351824838a/volumes" Dec 06 04:09:20 crc kubenswrapper[4802]: I1206 04:09:20.071922 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-qtg4k"] Dec 06 04:09:20 crc kubenswrapper[4802]: I1206 04:09:20.080633 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-qtg4k"] Dec 06 04:09:20 crc kubenswrapper[4802]: I1206 04:09:20.550538 4802 scope.go:117] "RemoveContainer" containerID="9d2031d53e371d2056994227b30d5a30e7440a814f6f693c65b33b77e8a0aefe" Dec 06 04:09:20 crc kubenswrapper[4802]: I1206 04:09:20.592648 4802 scope.go:117] "RemoveContainer" containerID="3d3dd772d0f77aac781b02d536ba30f54b9ce0a6d82391bd065c0446c975682b" Dec 06 04:09:20 crc kubenswrapper[4802]: I1206 04:09:20.641441 4802 scope.go:117] "RemoveContainer" containerID="36903355ba2d3d638b25f476eaf7539b274899782c428bcb2387df084f4ddc5d" Dec 06 04:09:20 crc kubenswrapper[4802]: I1206 04:09:20.672594 4802 scope.go:117] "RemoveContainer" containerID="3f4b38426be7601d64bc873172769785c9046300daa2599fab6b14601ad2253e" Dec 06 04:09:20 crc kubenswrapper[4802]: I1206 04:09:20.720144 4802 scope.go:117] "RemoveContainer" containerID="2e82f2fc1c3bfcea72061447fb05fb9dd187735d00c8db3b4ed7beb9e7607c72" Dec 06 04:09:20 crc kubenswrapper[4802]: I1206 04:09:20.761291 4802 scope.go:117] "RemoveContainer" containerID="1b96b595560864842e116cdbda2b107003083adfdc7dc7068ee10aebc4ffd2e6" Dec 06 04:09:20 crc kubenswrapper[4802]: I1206 04:09:20.798402 4802 scope.go:117] "RemoveContainer" containerID="b5003c23488a9b4291d460938302ad94691a2a22a3ad10889c45ae5722d34b63" Dec 06 04:09:20 crc kubenswrapper[4802]: I1206 04:09:20.833461 4802 scope.go:117] "RemoveContainer" containerID="1e3c95d017fa3e0e5f5a8706195ef03d408a5d227166a54b29857a70dde385c2" Dec 06 04:09:20 crc kubenswrapper[4802]: I1206 04:09:20.863531 4802 scope.go:117] "RemoveContainer" containerID="8126e8f607f92361572ba1d15723069acd373ae541bfa0c6c20ffd7a082374ec" Dec 06 04:09:20 crc kubenswrapper[4802]: I1206 04:09:20.882308 4802 scope.go:117] "RemoveContainer" containerID="8af62147f7298df9918791bebb79842e0b68d2d15c42f552492667fa7bc01da7" Dec 06 04:09:20 crc kubenswrapper[4802]: I1206 04:09:20.904823 4802 scope.go:117] "RemoveContainer" containerID="46e6ea4ac88dfb898ea9b663b1a6d76d786318b1afea471fa0d10200b8a30b87" Dec 06 04:09:20 crc kubenswrapper[4802]: I1206 04:09:20.935795 4802 scope.go:117] "RemoveContainer" containerID="f920d0003b0a0bffb8461417928e8983607d12b907c7583af66b6fadb491020a" Dec 06 04:09:21 crc kubenswrapper[4802]: I1206 04:09:21.469297 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c636931-d435-47ec-b770-80dc1ee60756" path="/var/lib/kubelet/pods/5c636931-d435-47ec-b770-80dc1ee60756/volumes" Dec 06 04:09:29 crc kubenswrapper[4802]: I1206 04:09:29.451259 4802 scope.go:117] "RemoveContainer" containerID="3711adecb81191bc35ac469efa73faeb2346d30f4d5a4a148de8ac88071cdb00" Dec 06 04:09:29 crc kubenswrapper[4802]: E1206 04:09:29.452008 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:09:31 crc kubenswrapper[4802]: I1206 04:09:31.064718 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-qs8hv"] Dec 06 04:09:31 crc kubenswrapper[4802]: I1206 04:09:31.076352 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-qs8hv"] Dec 06 04:09:31 crc kubenswrapper[4802]: I1206 04:09:31.471683 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c25c327-e3cd-4f28-855a-249e4bb0c5f6" path="/var/lib/kubelet/pods/8c25c327-e3cd-4f28-855a-249e4bb0c5f6/volumes" Dec 06 04:09:43 crc kubenswrapper[4802]: I1206 04:09:43.038960 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-5hj5p"] Dec 06 04:09:43 crc kubenswrapper[4802]: I1206 04:09:43.049531 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-2hzlt"] Dec 06 04:09:43 crc kubenswrapper[4802]: I1206 04:09:43.058890 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-5hj5p"] Dec 06 04:09:43 crc kubenswrapper[4802]: I1206 04:09:43.068201 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-2hzlt"] Dec 06 04:09:43 crc kubenswrapper[4802]: I1206 04:09:43.451000 4802 scope.go:117] "RemoveContainer" containerID="3711adecb81191bc35ac469efa73faeb2346d30f4d5a4a148de8ac88071cdb00" Dec 06 04:09:43 crc kubenswrapper[4802]: E1206 04:09:43.451410 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:09:43 crc kubenswrapper[4802]: I1206 04:09:43.463716 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a639667-be04-429f-a2ce-b20a011425f8" path="/var/lib/kubelet/pods/2a639667-be04-429f-a2ce-b20a011425f8/volumes" Dec 06 04:09:43 crc kubenswrapper[4802]: I1206 04:09:43.464376 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1644907-b4b1-4e23-a273-19d535bd7f88" path="/var/lib/kubelet/pods/f1644907-b4b1-4e23-a273-19d535bd7f88/volumes" Dec 06 04:09:54 crc kubenswrapper[4802]: I1206 04:09:54.450519 4802 scope.go:117] "RemoveContainer" containerID="3711adecb81191bc35ac469efa73faeb2346d30f4d5a4a148de8ac88071cdb00" Dec 06 04:09:54 crc kubenswrapper[4802]: E1206 04:09:54.451394 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:09:56 crc kubenswrapper[4802]: I1206 04:09:56.618948 4802 generic.go:334] "Generic (PLEG): container finished" podID="db0df1b2-ef63-42ae-96ff-cafa00acc8e1" containerID="a140b688777948d25cd90c0f584854dad6ee08e220dd1bda00ea073adf871131" exitCode=0 Dec 06 04:09:56 crc kubenswrapper[4802]: I1206 04:09:56.619156 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgxfp" event={"ID":"db0df1b2-ef63-42ae-96ff-cafa00acc8e1","Type":"ContainerDied","Data":"a140b688777948d25cd90c0f584854dad6ee08e220dd1bda00ea073adf871131"} Dec 06 04:09:58 crc kubenswrapper[4802]: I1206 04:09:58.038771 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgxfp" Dec 06 04:09:58 crc kubenswrapper[4802]: I1206 04:09:58.111080 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/db0df1b2-ef63-42ae-96ff-cafa00acc8e1-ssh-key\") pod \"db0df1b2-ef63-42ae-96ff-cafa00acc8e1\" (UID: \"db0df1b2-ef63-42ae-96ff-cafa00acc8e1\") " Dec 06 04:09:58 crc kubenswrapper[4802]: I1206 04:09:58.111147 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lvcsc\" (UniqueName: \"kubernetes.io/projected/db0df1b2-ef63-42ae-96ff-cafa00acc8e1-kube-api-access-lvcsc\") pod \"db0df1b2-ef63-42ae-96ff-cafa00acc8e1\" (UID: \"db0df1b2-ef63-42ae-96ff-cafa00acc8e1\") " Dec 06 04:09:58 crc kubenswrapper[4802]: I1206 04:09:58.111227 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/db0df1b2-ef63-42ae-96ff-cafa00acc8e1-inventory\") pod \"db0df1b2-ef63-42ae-96ff-cafa00acc8e1\" (UID: \"db0df1b2-ef63-42ae-96ff-cafa00acc8e1\") " Dec 06 04:09:58 crc kubenswrapper[4802]: I1206 04:09:58.116942 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db0df1b2-ef63-42ae-96ff-cafa00acc8e1-kube-api-access-lvcsc" (OuterVolumeSpecName: "kube-api-access-lvcsc") pod "db0df1b2-ef63-42ae-96ff-cafa00acc8e1" (UID: "db0df1b2-ef63-42ae-96ff-cafa00acc8e1"). InnerVolumeSpecName "kube-api-access-lvcsc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:09:58 crc kubenswrapper[4802]: I1206 04:09:58.138652 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db0df1b2-ef63-42ae-96ff-cafa00acc8e1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "db0df1b2-ef63-42ae-96ff-cafa00acc8e1" (UID: "db0df1b2-ef63-42ae-96ff-cafa00acc8e1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:09:58 crc kubenswrapper[4802]: I1206 04:09:58.140942 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db0df1b2-ef63-42ae-96ff-cafa00acc8e1-inventory" (OuterVolumeSpecName: "inventory") pod "db0df1b2-ef63-42ae-96ff-cafa00acc8e1" (UID: "db0df1b2-ef63-42ae-96ff-cafa00acc8e1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:09:58 crc kubenswrapper[4802]: I1206 04:09:58.214271 4802 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/db0df1b2-ef63-42ae-96ff-cafa00acc8e1-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 04:09:58 crc kubenswrapper[4802]: I1206 04:09:58.214294 4802 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/db0df1b2-ef63-42ae-96ff-cafa00acc8e1-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 04:09:58 crc kubenswrapper[4802]: I1206 04:09:58.214303 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lvcsc\" (UniqueName: \"kubernetes.io/projected/db0df1b2-ef63-42ae-96ff-cafa00acc8e1-kube-api-access-lvcsc\") on node \"crc\" DevicePath \"\"" Dec 06 04:09:58 crc kubenswrapper[4802]: I1206 04:09:58.637570 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgxfp" event={"ID":"db0df1b2-ef63-42ae-96ff-cafa00acc8e1","Type":"ContainerDied","Data":"aefa38e3b584b0bf6dcbe0ad2944c9a9672e80445c6af98c5d4e6b6cd0f66f53"} Dec 06 04:09:58 crc kubenswrapper[4802]: I1206 04:09:58.637607 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aefa38e3b584b0bf6dcbe0ad2944c9a9672e80445c6af98c5d4e6b6cd0f66f53" Dec 06 04:09:58 crc kubenswrapper[4802]: I1206 04:09:58.637634 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgxfp" Dec 06 04:09:58 crc kubenswrapper[4802]: I1206 04:09:58.729331 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-22wfn"] Dec 06 04:09:58 crc kubenswrapper[4802]: E1206 04:09:58.730186 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db0df1b2-ef63-42ae-96ff-cafa00acc8e1" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 06 04:09:58 crc kubenswrapper[4802]: I1206 04:09:58.730212 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="db0df1b2-ef63-42ae-96ff-cafa00acc8e1" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 06 04:09:58 crc kubenswrapper[4802]: I1206 04:09:58.730443 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="db0df1b2-ef63-42ae-96ff-cafa00acc8e1" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 06 04:09:58 crc kubenswrapper[4802]: I1206 04:09:58.731309 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-22wfn" Dec 06 04:09:58 crc kubenswrapper[4802]: I1206 04:09:58.733661 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 04:09:58 crc kubenswrapper[4802]: I1206 04:09:58.734332 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 04:09:58 crc kubenswrapper[4802]: I1206 04:09:58.734589 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 04:09:58 crc kubenswrapper[4802]: I1206 04:09:58.735831 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-rgvjc" Dec 06 04:09:58 crc kubenswrapper[4802]: I1206 04:09:58.739994 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-22wfn"] Dec 06 04:09:58 crc kubenswrapper[4802]: I1206 04:09:58.925305 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7ljg\" (UniqueName: \"kubernetes.io/projected/7d07f677-e982-45d5-b4c0-92e0db096fd1-kube-api-access-h7ljg\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-22wfn\" (UID: \"7d07f677-e982-45d5-b4c0-92e0db096fd1\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-22wfn" Dec 06 04:09:58 crc kubenswrapper[4802]: I1206 04:09:58.925393 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7d07f677-e982-45d5-b4c0-92e0db096fd1-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-22wfn\" (UID: \"7d07f677-e982-45d5-b4c0-92e0db096fd1\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-22wfn" Dec 06 04:09:58 crc kubenswrapper[4802]: I1206 04:09:58.925486 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7d07f677-e982-45d5-b4c0-92e0db096fd1-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-22wfn\" (UID: \"7d07f677-e982-45d5-b4c0-92e0db096fd1\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-22wfn" Dec 06 04:09:59 crc kubenswrapper[4802]: I1206 04:09:59.027445 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7d07f677-e982-45d5-b4c0-92e0db096fd1-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-22wfn\" (UID: \"7d07f677-e982-45d5-b4c0-92e0db096fd1\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-22wfn" Dec 06 04:09:59 crc kubenswrapper[4802]: I1206 04:09:59.027581 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7ljg\" (UniqueName: \"kubernetes.io/projected/7d07f677-e982-45d5-b4c0-92e0db096fd1-kube-api-access-h7ljg\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-22wfn\" (UID: \"7d07f677-e982-45d5-b4c0-92e0db096fd1\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-22wfn" Dec 06 04:09:59 crc kubenswrapper[4802]: I1206 04:09:59.027628 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7d07f677-e982-45d5-b4c0-92e0db096fd1-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-22wfn\" (UID: \"7d07f677-e982-45d5-b4c0-92e0db096fd1\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-22wfn" Dec 06 04:09:59 crc kubenswrapper[4802]: I1206 04:09:59.032375 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7d07f677-e982-45d5-b4c0-92e0db096fd1-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-22wfn\" (UID: \"7d07f677-e982-45d5-b4c0-92e0db096fd1\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-22wfn" Dec 06 04:09:59 crc kubenswrapper[4802]: I1206 04:09:59.032723 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7d07f677-e982-45d5-b4c0-92e0db096fd1-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-22wfn\" (UID: \"7d07f677-e982-45d5-b4c0-92e0db096fd1\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-22wfn" Dec 06 04:09:59 crc kubenswrapper[4802]: I1206 04:09:59.048827 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7ljg\" (UniqueName: \"kubernetes.io/projected/7d07f677-e982-45d5-b4c0-92e0db096fd1-kube-api-access-h7ljg\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-22wfn\" (UID: \"7d07f677-e982-45d5-b4c0-92e0db096fd1\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-22wfn" Dec 06 04:09:59 crc kubenswrapper[4802]: I1206 04:09:59.349809 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-22wfn" Dec 06 04:09:59 crc kubenswrapper[4802]: I1206 04:09:59.971204 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-22wfn"] Dec 06 04:10:00 crc kubenswrapper[4802]: I1206 04:10:00.656386 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-22wfn" event={"ID":"7d07f677-e982-45d5-b4c0-92e0db096fd1","Type":"ContainerStarted","Data":"7660f249acf3e04882361b3d13c2fd26fc96d9c368b3494e4f398029752782fc"} Dec 06 04:10:00 crc kubenswrapper[4802]: I1206 04:10:00.656689 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-22wfn" event={"ID":"7d07f677-e982-45d5-b4c0-92e0db096fd1","Type":"ContainerStarted","Data":"ba2cb5778fb3f4c8332544a56da994efdf8b236ed884fa67d1a5d015d7870d94"} Dec 06 04:10:00 crc kubenswrapper[4802]: I1206 04:10:00.673105 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-22wfn" podStartSLOduration=2.249095688 podStartE2EDuration="2.673084164s" podCreationTimestamp="2025-12-06 04:09:58 +0000 UTC" firstStartedPulling="2025-12-06 04:09:59.976173138 +0000 UTC m=+1792.848082300" lastFinishedPulling="2025-12-06 04:10:00.400161624 +0000 UTC m=+1793.272070776" observedRunningTime="2025-12-06 04:10:00.673066203 +0000 UTC m=+1793.544975355" watchObservedRunningTime="2025-12-06 04:10:00.673084164 +0000 UTC m=+1793.544993316" Dec 06 04:10:05 crc kubenswrapper[4802]: I1206 04:10:05.450789 4802 scope.go:117] "RemoveContainer" containerID="3711adecb81191bc35ac469efa73faeb2346d30f4d5a4a148de8ac88071cdb00" Dec 06 04:10:05 crc kubenswrapper[4802]: E1206 04:10:05.451532 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:10:19 crc kubenswrapper[4802]: I1206 04:10:19.451414 4802 scope.go:117] "RemoveContainer" containerID="3711adecb81191bc35ac469efa73faeb2346d30f4d5a4a148de8ac88071cdb00" Dec 06 04:10:19 crc kubenswrapper[4802]: E1206 04:10:19.452349 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:10:21 crc kubenswrapper[4802]: I1206 04:10:21.153148 4802 scope.go:117] "RemoveContainer" containerID="860845ba16dd5a1de8338b6605815539607b7f5f02448fa1f96a5b3f02ee9a9e" Dec 06 04:10:21 crc kubenswrapper[4802]: I1206 04:10:21.206541 4802 scope.go:117] "RemoveContainer" containerID="66f602546227aa86f52ceb621c4f145c78c1b12628e754ab524c6e32f88a0151" Dec 06 04:10:21 crc kubenswrapper[4802]: I1206 04:10:21.264801 4802 scope.go:117] "RemoveContainer" containerID="22256b01fbd251cd8845a65d030b3e708b70e6bb74755d572798823340fae615" Dec 06 04:10:21 crc kubenswrapper[4802]: I1206 04:10:21.297405 4802 scope.go:117] "RemoveContainer" containerID="285d42f44e31edd7beb22eac4c447d12107075be937e5c0e69d394dc90c0ac17" Dec 06 04:10:30 crc kubenswrapper[4802]: I1206 04:10:30.071004 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-gv9nf"] Dec 06 04:10:30 crc kubenswrapper[4802]: I1206 04:10:30.089556 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-gv9nf"] Dec 06 04:10:30 crc kubenswrapper[4802]: I1206 04:10:30.450923 4802 scope.go:117] "RemoveContainer" containerID="3711adecb81191bc35ac469efa73faeb2346d30f4d5a4a148de8ac88071cdb00" Dec 06 04:10:30 crc kubenswrapper[4802]: E1206 04:10:30.451194 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:10:31 crc kubenswrapper[4802]: I1206 04:10:31.033597 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-b8sxp"] Dec 06 04:10:31 crc kubenswrapper[4802]: I1206 04:10:31.044761 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-b8sxp"] Dec 06 04:10:31 crc kubenswrapper[4802]: I1206 04:10:31.056667 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-rhcrb"] Dec 06 04:10:31 crc kubenswrapper[4802]: I1206 04:10:31.066553 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-1043-account-create-update-xzd9x"] Dec 06 04:10:31 crc kubenswrapper[4802]: I1206 04:10:31.077317 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-1043-account-create-update-xzd9x"] Dec 06 04:10:31 crc kubenswrapper[4802]: I1206 04:10:31.088772 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-rhcrb"] Dec 06 04:10:31 crc kubenswrapper[4802]: I1206 04:10:31.461565 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04a970e8-495a-491a-9f83-1b25159450b1" path="/var/lib/kubelet/pods/04a970e8-495a-491a-9f83-1b25159450b1/volumes" Dec 06 04:10:31 crc kubenswrapper[4802]: I1206 04:10:31.462187 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f247bdd-ddf7-4872-a1c5-597fec152cde" path="/var/lib/kubelet/pods/2f247bdd-ddf7-4872-a1c5-597fec152cde/volumes" Dec 06 04:10:31 crc kubenswrapper[4802]: I1206 04:10:31.462730 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59d551a5-c1ef-46ee-8ae2-f98526b2ab09" path="/var/lib/kubelet/pods/59d551a5-c1ef-46ee-8ae2-f98526b2ab09/volumes" Dec 06 04:10:31 crc kubenswrapper[4802]: I1206 04:10:31.463309 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2e9e18c-0730-4b74-8eaf-941cbeb08246" path="/var/lib/kubelet/pods/b2e9e18c-0730-4b74-8eaf-941cbeb08246/volumes" Dec 06 04:10:32 crc kubenswrapper[4802]: I1206 04:10:32.050139 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cdd3-account-create-update-bjk6c"] Dec 06 04:10:32 crc kubenswrapper[4802]: I1206 04:10:32.076464 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-66f6-account-create-update-th68q"] Dec 06 04:10:32 crc kubenswrapper[4802]: I1206 04:10:32.089226 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cdd3-account-create-update-bjk6c"] Dec 06 04:10:32 crc kubenswrapper[4802]: I1206 04:10:32.099148 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-66f6-account-create-update-th68q"] Dec 06 04:10:33 crc kubenswrapper[4802]: I1206 04:10:33.472236 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="164b8cc5-aab5-4e71-af01-d39679d4b3ba" path="/var/lib/kubelet/pods/164b8cc5-aab5-4e71-af01-d39679d4b3ba/volumes" Dec 06 04:10:33 crc kubenswrapper[4802]: I1206 04:10:33.473392 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a5a43d4-c219-4090-acc3-252e127e48e7" path="/var/lib/kubelet/pods/8a5a43d4-c219-4090-acc3-252e127e48e7/volumes" Dec 06 04:10:44 crc kubenswrapper[4802]: I1206 04:10:44.450708 4802 scope.go:117] "RemoveContainer" containerID="3711adecb81191bc35ac469efa73faeb2346d30f4d5a4a148de8ac88071cdb00" Dec 06 04:10:45 crc kubenswrapper[4802]: I1206 04:10:45.127872 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" event={"ID":"6df38316-e0d3-4018-8d27-3620eba3a68d","Type":"ContainerStarted","Data":"baa707a57be38104aa835e7ecacc0175c718b272c82894ed9c2693dc0041453b"} Dec 06 04:11:02 crc kubenswrapper[4802]: I1206 04:11:02.053352 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-xnnxk"] Dec 06 04:11:02 crc kubenswrapper[4802]: I1206 04:11:02.069821 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-xnnxk"] Dec 06 04:11:03 crc kubenswrapper[4802]: I1206 04:11:03.463361 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f09f7e4f-6e98-4461-bf3a-1a6aa9528c09" path="/var/lib/kubelet/pods/f09f7e4f-6e98-4461-bf3a-1a6aa9528c09/volumes" Dec 06 04:11:15 crc kubenswrapper[4802]: I1206 04:11:15.426219 4802 generic.go:334] "Generic (PLEG): container finished" podID="7d07f677-e982-45d5-b4c0-92e0db096fd1" containerID="7660f249acf3e04882361b3d13c2fd26fc96d9c368b3494e4f398029752782fc" exitCode=0 Dec 06 04:11:15 crc kubenswrapper[4802]: I1206 04:11:15.426351 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-22wfn" event={"ID":"7d07f677-e982-45d5-b4c0-92e0db096fd1","Type":"ContainerDied","Data":"7660f249acf3e04882361b3d13c2fd26fc96d9c368b3494e4f398029752782fc"} Dec 06 04:11:16 crc kubenswrapper[4802]: I1206 04:11:16.818895 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-22wfn" Dec 06 04:11:16 crc kubenswrapper[4802]: I1206 04:11:16.957802 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7d07f677-e982-45d5-b4c0-92e0db096fd1-ssh-key\") pod \"7d07f677-e982-45d5-b4c0-92e0db096fd1\" (UID: \"7d07f677-e982-45d5-b4c0-92e0db096fd1\") " Dec 06 04:11:16 crc kubenswrapper[4802]: I1206 04:11:16.958034 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h7ljg\" (UniqueName: \"kubernetes.io/projected/7d07f677-e982-45d5-b4c0-92e0db096fd1-kube-api-access-h7ljg\") pod \"7d07f677-e982-45d5-b4c0-92e0db096fd1\" (UID: \"7d07f677-e982-45d5-b4c0-92e0db096fd1\") " Dec 06 04:11:16 crc kubenswrapper[4802]: I1206 04:11:16.958147 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7d07f677-e982-45d5-b4c0-92e0db096fd1-inventory\") pod \"7d07f677-e982-45d5-b4c0-92e0db096fd1\" (UID: \"7d07f677-e982-45d5-b4c0-92e0db096fd1\") " Dec 06 04:11:16 crc kubenswrapper[4802]: I1206 04:11:16.967693 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d07f677-e982-45d5-b4c0-92e0db096fd1-kube-api-access-h7ljg" (OuterVolumeSpecName: "kube-api-access-h7ljg") pod "7d07f677-e982-45d5-b4c0-92e0db096fd1" (UID: "7d07f677-e982-45d5-b4c0-92e0db096fd1"). InnerVolumeSpecName "kube-api-access-h7ljg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:11:16 crc kubenswrapper[4802]: I1206 04:11:16.985211 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d07f677-e982-45d5-b4c0-92e0db096fd1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7d07f677-e982-45d5-b4c0-92e0db096fd1" (UID: "7d07f677-e982-45d5-b4c0-92e0db096fd1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:11:16 crc kubenswrapper[4802]: I1206 04:11:16.996628 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d07f677-e982-45d5-b4c0-92e0db096fd1-inventory" (OuterVolumeSpecName: "inventory") pod "7d07f677-e982-45d5-b4c0-92e0db096fd1" (UID: "7d07f677-e982-45d5-b4c0-92e0db096fd1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:11:17 crc kubenswrapper[4802]: I1206 04:11:17.060112 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h7ljg\" (UniqueName: \"kubernetes.io/projected/7d07f677-e982-45d5-b4c0-92e0db096fd1-kube-api-access-h7ljg\") on node \"crc\" DevicePath \"\"" Dec 06 04:11:17 crc kubenswrapper[4802]: I1206 04:11:17.060147 4802 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7d07f677-e982-45d5-b4c0-92e0db096fd1-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 04:11:17 crc kubenswrapper[4802]: I1206 04:11:17.060157 4802 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7d07f677-e982-45d5-b4c0-92e0db096fd1-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 04:11:17 crc kubenswrapper[4802]: I1206 04:11:17.445241 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-22wfn" event={"ID":"7d07f677-e982-45d5-b4c0-92e0db096fd1","Type":"ContainerDied","Data":"ba2cb5778fb3f4c8332544a56da994efdf8b236ed884fa67d1a5d015d7870d94"} Dec 06 04:11:17 crc kubenswrapper[4802]: I1206 04:11:17.445295 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ba2cb5778fb3f4c8332544a56da994efdf8b236ed884fa67d1a5d015d7870d94" Dec 06 04:11:17 crc kubenswrapper[4802]: I1206 04:11:17.445330 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-22wfn" Dec 06 04:11:17 crc kubenswrapper[4802]: I1206 04:11:17.538463 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7nbhk"] Dec 06 04:11:17 crc kubenswrapper[4802]: E1206 04:11:17.539091 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d07f677-e982-45d5-b4c0-92e0db096fd1" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 06 04:11:17 crc kubenswrapper[4802]: I1206 04:11:17.539125 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d07f677-e982-45d5-b4c0-92e0db096fd1" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 06 04:11:17 crc kubenswrapper[4802]: I1206 04:11:17.539496 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d07f677-e982-45d5-b4c0-92e0db096fd1" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 06 04:11:17 crc kubenswrapper[4802]: I1206 04:11:17.540632 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7nbhk" Dec 06 04:11:17 crc kubenswrapper[4802]: I1206 04:11:17.549920 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 04:11:17 crc kubenswrapper[4802]: I1206 04:11:17.554598 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 04:11:17 crc kubenswrapper[4802]: I1206 04:11:17.554608 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 04:11:17 crc kubenswrapper[4802]: I1206 04:11:17.555546 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-rgvjc" Dec 06 04:11:17 crc kubenswrapper[4802]: I1206 04:11:17.566963 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7nbhk"] Dec 06 04:11:17 crc kubenswrapper[4802]: I1206 04:11:17.680504 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rgmvj\" (UniqueName: \"kubernetes.io/projected/5fd4f8df-59a8-47cb-b9e7-4162a5811ed7-kube-api-access-rgmvj\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-7nbhk\" (UID: \"5fd4f8df-59a8-47cb-b9e7-4162a5811ed7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7nbhk" Dec 06 04:11:17 crc kubenswrapper[4802]: I1206 04:11:17.680707 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fd4f8df-59a8-47cb-b9e7-4162a5811ed7-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-7nbhk\" (UID: \"5fd4f8df-59a8-47cb-b9e7-4162a5811ed7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7nbhk" Dec 06 04:11:17 crc kubenswrapper[4802]: I1206 04:11:17.680779 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5fd4f8df-59a8-47cb-b9e7-4162a5811ed7-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-7nbhk\" (UID: \"5fd4f8df-59a8-47cb-b9e7-4162a5811ed7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7nbhk" Dec 06 04:11:17 crc kubenswrapper[4802]: I1206 04:11:17.782202 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rgmvj\" (UniqueName: \"kubernetes.io/projected/5fd4f8df-59a8-47cb-b9e7-4162a5811ed7-kube-api-access-rgmvj\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-7nbhk\" (UID: \"5fd4f8df-59a8-47cb-b9e7-4162a5811ed7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7nbhk" Dec 06 04:11:17 crc kubenswrapper[4802]: I1206 04:11:17.782581 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fd4f8df-59a8-47cb-b9e7-4162a5811ed7-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-7nbhk\" (UID: \"5fd4f8df-59a8-47cb-b9e7-4162a5811ed7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7nbhk" Dec 06 04:11:17 crc kubenswrapper[4802]: I1206 04:11:17.782614 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5fd4f8df-59a8-47cb-b9e7-4162a5811ed7-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-7nbhk\" (UID: \"5fd4f8df-59a8-47cb-b9e7-4162a5811ed7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7nbhk" Dec 06 04:11:17 crc kubenswrapper[4802]: I1206 04:11:17.789587 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fd4f8df-59a8-47cb-b9e7-4162a5811ed7-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-7nbhk\" (UID: \"5fd4f8df-59a8-47cb-b9e7-4162a5811ed7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7nbhk" Dec 06 04:11:17 crc kubenswrapper[4802]: I1206 04:11:17.791439 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5fd4f8df-59a8-47cb-b9e7-4162a5811ed7-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-7nbhk\" (UID: \"5fd4f8df-59a8-47cb-b9e7-4162a5811ed7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7nbhk" Dec 06 04:11:17 crc kubenswrapper[4802]: I1206 04:11:17.813945 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rgmvj\" (UniqueName: \"kubernetes.io/projected/5fd4f8df-59a8-47cb-b9e7-4162a5811ed7-kube-api-access-rgmvj\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-7nbhk\" (UID: \"5fd4f8df-59a8-47cb-b9e7-4162a5811ed7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7nbhk" Dec 06 04:11:17 crc kubenswrapper[4802]: I1206 04:11:17.857022 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7nbhk" Dec 06 04:11:18 crc kubenswrapper[4802]: I1206 04:11:18.398620 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7nbhk"] Dec 06 04:11:18 crc kubenswrapper[4802]: I1206 04:11:18.402912 4802 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 04:11:18 crc kubenswrapper[4802]: I1206 04:11:18.454197 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7nbhk" event={"ID":"5fd4f8df-59a8-47cb-b9e7-4162a5811ed7","Type":"ContainerStarted","Data":"c38f1708ba708b94a3a582e6a3f6eadaaf667ba27523a64543c286dbff3f64d9"} Dec 06 04:11:19 crc kubenswrapper[4802]: I1206 04:11:19.467210 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7nbhk" event={"ID":"5fd4f8df-59a8-47cb-b9e7-4162a5811ed7","Type":"ContainerStarted","Data":"baefc674b503bbcf3f468117ea45348af06b667cc09bd9d6c390be2134828cea"} Dec 06 04:11:21 crc kubenswrapper[4802]: I1206 04:11:21.412342 4802 scope.go:117] "RemoveContainer" containerID="5ab9087cf180f2a18ce6757d996a25ac4c3e22341ed3ae95f9624968cb5de9b2" Dec 06 04:11:21 crc kubenswrapper[4802]: I1206 04:11:21.438585 4802 scope.go:117] "RemoveContainer" containerID="71f041b5f79ae84d3ca26838b981a0e01fda8163adf89de78ed9c1bdb15ccff3" Dec 06 04:11:21 crc kubenswrapper[4802]: I1206 04:11:21.487706 4802 scope.go:117] "RemoveContainer" containerID="8f9b17b6778156f4f51764b0b22842a3799ddd93aca19a6b91890d38395cbde7" Dec 06 04:11:21 crc kubenswrapper[4802]: I1206 04:11:21.545541 4802 scope.go:117] "RemoveContainer" containerID="be5c2c6140d48e66a81cd2b5f33d8dd4541e5161488459f5109327ce75d6bccf" Dec 06 04:11:21 crc kubenswrapper[4802]: I1206 04:11:21.605962 4802 scope.go:117] "RemoveContainer" containerID="27a342c130d871ab7897d1d85dbab34c5507f64a62e6aa01189f4c55aa35e1ad" Dec 06 04:11:21 crc kubenswrapper[4802]: I1206 04:11:21.667887 4802 scope.go:117] "RemoveContainer" containerID="e646cfb755995e7f7666c31c64b92161aece638ad58007ab7195f0fdb2a45316" Dec 06 04:11:21 crc kubenswrapper[4802]: I1206 04:11:21.742336 4802 scope.go:117] "RemoveContainer" containerID="a6c5c28aa66df89c8e48ac36cf218e1d67940dcd025ae23684066eb35df96f0d" Dec 06 04:11:24 crc kubenswrapper[4802]: I1206 04:11:24.041186 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7nbhk" podStartSLOduration=6.45110434 podStartE2EDuration="7.041160719s" podCreationTimestamp="2025-12-06 04:11:17 +0000 UTC" firstStartedPulling="2025-12-06 04:11:18.402628099 +0000 UTC m=+1871.274537251" lastFinishedPulling="2025-12-06 04:11:18.992684478 +0000 UTC m=+1871.864593630" observedRunningTime="2025-12-06 04:11:19.497116236 +0000 UTC m=+1872.369025388" watchObservedRunningTime="2025-12-06 04:11:24.041160719 +0000 UTC m=+1876.913069881" Dec 06 04:11:24 crc kubenswrapper[4802]: I1206 04:11:24.046473 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-kx676"] Dec 06 04:11:24 crc kubenswrapper[4802]: I1206 04:11:24.061831 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-kx676"] Dec 06 04:11:24 crc kubenswrapper[4802]: I1206 04:11:24.510641 4802 generic.go:334] "Generic (PLEG): container finished" podID="5fd4f8df-59a8-47cb-b9e7-4162a5811ed7" containerID="baefc674b503bbcf3f468117ea45348af06b667cc09bd9d6c390be2134828cea" exitCode=0 Dec 06 04:11:24 crc kubenswrapper[4802]: I1206 04:11:24.510686 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7nbhk" event={"ID":"5fd4f8df-59a8-47cb-b9e7-4162a5811ed7","Type":"ContainerDied","Data":"baefc674b503bbcf3f468117ea45348af06b667cc09bd9d6c390be2134828cea"} Dec 06 04:11:25 crc kubenswrapper[4802]: I1206 04:11:25.035994 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-5sl8p"] Dec 06 04:11:25 crc kubenswrapper[4802]: I1206 04:11:25.047724 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-5sl8p"] Dec 06 04:11:25 crc kubenswrapper[4802]: I1206 04:11:25.476564 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="735dec51-8659-44bf-bd0b-afd1a1d78d86" path="/var/lib/kubelet/pods/735dec51-8659-44bf-bd0b-afd1a1d78d86/volumes" Dec 06 04:11:25 crc kubenswrapper[4802]: I1206 04:11:25.477304 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb7d4cc2-ea97-4675-bc65-8803fa1e3708" path="/var/lib/kubelet/pods/bb7d4cc2-ea97-4675-bc65-8803fa1e3708/volumes" Dec 06 04:11:25 crc kubenswrapper[4802]: I1206 04:11:25.926833 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7nbhk" Dec 06 04:11:25 crc kubenswrapper[4802]: I1206 04:11:25.954103 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5fd4f8df-59a8-47cb-b9e7-4162a5811ed7-ssh-key\") pod \"5fd4f8df-59a8-47cb-b9e7-4162a5811ed7\" (UID: \"5fd4f8df-59a8-47cb-b9e7-4162a5811ed7\") " Dec 06 04:11:25 crc kubenswrapper[4802]: I1206 04:11:25.954237 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fd4f8df-59a8-47cb-b9e7-4162a5811ed7-inventory\") pod \"5fd4f8df-59a8-47cb-b9e7-4162a5811ed7\" (UID: \"5fd4f8df-59a8-47cb-b9e7-4162a5811ed7\") " Dec 06 04:11:25 crc kubenswrapper[4802]: I1206 04:11:25.954271 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rgmvj\" (UniqueName: \"kubernetes.io/projected/5fd4f8df-59a8-47cb-b9e7-4162a5811ed7-kube-api-access-rgmvj\") pod \"5fd4f8df-59a8-47cb-b9e7-4162a5811ed7\" (UID: \"5fd4f8df-59a8-47cb-b9e7-4162a5811ed7\") " Dec 06 04:11:25 crc kubenswrapper[4802]: I1206 04:11:25.962006 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fd4f8df-59a8-47cb-b9e7-4162a5811ed7-kube-api-access-rgmvj" (OuterVolumeSpecName: "kube-api-access-rgmvj") pod "5fd4f8df-59a8-47cb-b9e7-4162a5811ed7" (UID: "5fd4f8df-59a8-47cb-b9e7-4162a5811ed7"). InnerVolumeSpecName "kube-api-access-rgmvj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:11:25 crc kubenswrapper[4802]: I1206 04:11:25.984491 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fd4f8df-59a8-47cb-b9e7-4162a5811ed7-inventory" (OuterVolumeSpecName: "inventory") pod "5fd4f8df-59a8-47cb-b9e7-4162a5811ed7" (UID: "5fd4f8df-59a8-47cb-b9e7-4162a5811ed7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:11:25 crc kubenswrapper[4802]: I1206 04:11:25.989116 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fd4f8df-59a8-47cb-b9e7-4162a5811ed7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5fd4f8df-59a8-47cb-b9e7-4162a5811ed7" (UID: "5fd4f8df-59a8-47cb-b9e7-4162a5811ed7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:11:26 crc kubenswrapper[4802]: I1206 04:11:26.056145 4802 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fd4f8df-59a8-47cb-b9e7-4162a5811ed7-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 04:11:26 crc kubenswrapper[4802]: I1206 04:11:26.056218 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rgmvj\" (UniqueName: \"kubernetes.io/projected/5fd4f8df-59a8-47cb-b9e7-4162a5811ed7-kube-api-access-rgmvj\") on node \"crc\" DevicePath \"\"" Dec 06 04:11:26 crc kubenswrapper[4802]: I1206 04:11:26.056229 4802 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5fd4f8df-59a8-47cb-b9e7-4162a5811ed7-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 04:11:26 crc kubenswrapper[4802]: I1206 04:11:26.527978 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7nbhk" event={"ID":"5fd4f8df-59a8-47cb-b9e7-4162a5811ed7","Type":"ContainerDied","Data":"c38f1708ba708b94a3a582e6a3f6eadaaf667ba27523a64543c286dbff3f64d9"} Dec 06 04:11:26 crc kubenswrapper[4802]: I1206 04:11:26.528021 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c38f1708ba708b94a3a582e6a3f6eadaaf667ba27523a64543c286dbff3f64d9" Dec 06 04:11:26 crc kubenswrapper[4802]: I1206 04:11:26.528028 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7nbhk" Dec 06 04:11:26 crc kubenswrapper[4802]: I1206 04:11:26.610367 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-g9qlw"] Dec 06 04:11:26 crc kubenswrapper[4802]: E1206 04:11:26.611111 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fd4f8df-59a8-47cb-b9e7-4162a5811ed7" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 06 04:11:26 crc kubenswrapper[4802]: I1206 04:11:26.611231 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fd4f8df-59a8-47cb-b9e7-4162a5811ed7" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 06 04:11:26 crc kubenswrapper[4802]: I1206 04:11:26.611468 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fd4f8df-59a8-47cb-b9e7-4162a5811ed7" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 06 04:11:26 crc kubenswrapper[4802]: I1206 04:11:26.612180 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g9qlw" Dec 06 04:11:26 crc kubenswrapper[4802]: I1206 04:11:26.616437 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-rgvjc" Dec 06 04:11:26 crc kubenswrapper[4802]: I1206 04:11:26.616731 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 04:11:26 crc kubenswrapper[4802]: I1206 04:11:26.616742 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 04:11:26 crc kubenswrapper[4802]: I1206 04:11:26.616824 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 04:11:26 crc kubenswrapper[4802]: I1206 04:11:26.618898 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-g9qlw"] Dec 06 04:11:26 crc kubenswrapper[4802]: I1206 04:11:26.770416 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b6c3bbd2-7993-420e-91d7-ffc72b1d2439-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-g9qlw\" (UID: \"b6c3bbd2-7993-420e-91d7-ffc72b1d2439\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g9qlw" Dec 06 04:11:26 crc kubenswrapper[4802]: I1206 04:11:26.770524 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b6c3bbd2-7993-420e-91d7-ffc72b1d2439-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-g9qlw\" (UID: \"b6c3bbd2-7993-420e-91d7-ffc72b1d2439\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g9qlw" Dec 06 04:11:26 crc kubenswrapper[4802]: I1206 04:11:26.770659 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9chh5\" (UniqueName: \"kubernetes.io/projected/b6c3bbd2-7993-420e-91d7-ffc72b1d2439-kube-api-access-9chh5\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-g9qlw\" (UID: \"b6c3bbd2-7993-420e-91d7-ffc72b1d2439\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g9qlw" Dec 06 04:11:26 crc kubenswrapper[4802]: I1206 04:11:26.872912 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9chh5\" (UniqueName: \"kubernetes.io/projected/b6c3bbd2-7993-420e-91d7-ffc72b1d2439-kube-api-access-9chh5\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-g9qlw\" (UID: \"b6c3bbd2-7993-420e-91d7-ffc72b1d2439\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g9qlw" Dec 06 04:11:26 crc kubenswrapper[4802]: I1206 04:11:26.873010 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b6c3bbd2-7993-420e-91d7-ffc72b1d2439-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-g9qlw\" (UID: \"b6c3bbd2-7993-420e-91d7-ffc72b1d2439\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g9qlw" Dec 06 04:11:26 crc kubenswrapper[4802]: I1206 04:11:26.873068 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b6c3bbd2-7993-420e-91d7-ffc72b1d2439-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-g9qlw\" (UID: \"b6c3bbd2-7993-420e-91d7-ffc72b1d2439\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g9qlw" Dec 06 04:11:26 crc kubenswrapper[4802]: I1206 04:11:26.881484 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b6c3bbd2-7993-420e-91d7-ffc72b1d2439-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-g9qlw\" (UID: \"b6c3bbd2-7993-420e-91d7-ffc72b1d2439\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g9qlw" Dec 06 04:11:26 crc kubenswrapper[4802]: I1206 04:11:26.885432 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b6c3bbd2-7993-420e-91d7-ffc72b1d2439-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-g9qlw\" (UID: \"b6c3bbd2-7993-420e-91d7-ffc72b1d2439\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g9qlw" Dec 06 04:11:26 crc kubenswrapper[4802]: I1206 04:11:26.892834 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9chh5\" (UniqueName: \"kubernetes.io/projected/b6c3bbd2-7993-420e-91d7-ffc72b1d2439-kube-api-access-9chh5\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-g9qlw\" (UID: \"b6c3bbd2-7993-420e-91d7-ffc72b1d2439\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g9qlw" Dec 06 04:11:26 crc kubenswrapper[4802]: I1206 04:11:26.942869 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g9qlw" Dec 06 04:11:27 crc kubenswrapper[4802]: I1206 04:11:27.377702 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-g9qlw"] Dec 06 04:11:27 crc kubenswrapper[4802]: I1206 04:11:27.537218 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g9qlw" event={"ID":"b6c3bbd2-7993-420e-91d7-ffc72b1d2439","Type":"ContainerStarted","Data":"a983ff46e59aecfdea1f7ffbff5a81fad820c73a5a49161771dc2d95130d1a50"} Dec 06 04:11:28 crc kubenswrapper[4802]: I1206 04:11:28.548630 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g9qlw" event={"ID":"b6c3bbd2-7993-420e-91d7-ffc72b1d2439","Type":"ContainerStarted","Data":"733ecca1d5e47ad47d78acbe2958d09431d50b98f3e25073f29b12a5739143e1"} Dec 06 04:11:28 crc kubenswrapper[4802]: I1206 04:11:28.579101 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g9qlw" podStartSLOduration=2.06179269 podStartE2EDuration="2.579079305s" podCreationTimestamp="2025-12-06 04:11:26 +0000 UTC" firstStartedPulling="2025-12-06 04:11:27.393828157 +0000 UTC m=+1880.265737309" lastFinishedPulling="2025-12-06 04:11:27.911114772 +0000 UTC m=+1880.783023924" observedRunningTime="2025-12-06 04:11:28.572119356 +0000 UTC m=+1881.444028518" watchObservedRunningTime="2025-12-06 04:11:28.579079305 +0000 UTC m=+1881.450988447" Dec 06 04:12:07 crc kubenswrapper[4802]: E1206 04:12:07.078844 4802 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb6c3bbd2_7993_420e_91d7_ffc72b1d2439.slice/crio-733ecca1d5e47ad47d78acbe2958d09431d50b98f3e25073f29b12a5739143e1.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb6c3bbd2_7993_420e_91d7_ffc72b1d2439.slice/crio-conmon-733ecca1d5e47ad47d78acbe2958d09431d50b98f3e25073f29b12a5739143e1.scope\": RecentStats: unable to find data in memory cache]" Dec 06 04:12:07 crc kubenswrapper[4802]: I1206 04:12:07.246600 4802 generic.go:334] "Generic (PLEG): container finished" podID="b6c3bbd2-7993-420e-91d7-ffc72b1d2439" containerID="733ecca1d5e47ad47d78acbe2958d09431d50b98f3e25073f29b12a5739143e1" exitCode=0 Dec 06 04:12:07 crc kubenswrapper[4802]: I1206 04:12:07.246695 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g9qlw" event={"ID":"b6c3bbd2-7993-420e-91d7-ffc72b1d2439","Type":"ContainerDied","Data":"733ecca1d5e47ad47d78acbe2958d09431d50b98f3e25073f29b12a5739143e1"} Dec 06 04:12:08 crc kubenswrapper[4802]: I1206 04:12:08.706048 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g9qlw" Dec 06 04:12:08 crc kubenswrapper[4802]: I1206 04:12:08.739666 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b6c3bbd2-7993-420e-91d7-ffc72b1d2439-ssh-key\") pod \"b6c3bbd2-7993-420e-91d7-ffc72b1d2439\" (UID: \"b6c3bbd2-7993-420e-91d7-ffc72b1d2439\") " Dec 06 04:12:08 crc kubenswrapper[4802]: I1206 04:12:08.739794 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9chh5\" (UniqueName: \"kubernetes.io/projected/b6c3bbd2-7993-420e-91d7-ffc72b1d2439-kube-api-access-9chh5\") pod \"b6c3bbd2-7993-420e-91d7-ffc72b1d2439\" (UID: \"b6c3bbd2-7993-420e-91d7-ffc72b1d2439\") " Dec 06 04:12:08 crc kubenswrapper[4802]: I1206 04:12:08.739919 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b6c3bbd2-7993-420e-91d7-ffc72b1d2439-inventory\") pod \"b6c3bbd2-7993-420e-91d7-ffc72b1d2439\" (UID: \"b6c3bbd2-7993-420e-91d7-ffc72b1d2439\") " Dec 06 04:12:08 crc kubenswrapper[4802]: I1206 04:12:08.751072 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6c3bbd2-7993-420e-91d7-ffc72b1d2439-kube-api-access-9chh5" (OuterVolumeSpecName: "kube-api-access-9chh5") pod "b6c3bbd2-7993-420e-91d7-ffc72b1d2439" (UID: "b6c3bbd2-7993-420e-91d7-ffc72b1d2439"). InnerVolumeSpecName "kube-api-access-9chh5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:12:08 crc kubenswrapper[4802]: I1206 04:12:08.776575 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6c3bbd2-7993-420e-91d7-ffc72b1d2439-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b6c3bbd2-7993-420e-91d7-ffc72b1d2439" (UID: "b6c3bbd2-7993-420e-91d7-ffc72b1d2439"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:12:08 crc kubenswrapper[4802]: I1206 04:12:08.783966 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6c3bbd2-7993-420e-91d7-ffc72b1d2439-inventory" (OuterVolumeSpecName: "inventory") pod "b6c3bbd2-7993-420e-91d7-ffc72b1d2439" (UID: "b6c3bbd2-7993-420e-91d7-ffc72b1d2439"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:12:08 crc kubenswrapper[4802]: I1206 04:12:08.842044 4802 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b6c3bbd2-7993-420e-91d7-ffc72b1d2439-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 04:12:08 crc kubenswrapper[4802]: I1206 04:12:08.842586 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9chh5\" (UniqueName: \"kubernetes.io/projected/b6c3bbd2-7993-420e-91d7-ffc72b1d2439-kube-api-access-9chh5\") on node \"crc\" DevicePath \"\"" Dec 06 04:12:08 crc kubenswrapper[4802]: I1206 04:12:08.842678 4802 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b6c3bbd2-7993-420e-91d7-ffc72b1d2439-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 04:12:09 crc kubenswrapper[4802]: I1206 04:12:09.271392 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g9qlw" event={"ID":"b6c3bbd2-7993-420e-91d7-ffc72b1d2439","Type":"ContainerDied","Data":"a983ff46e59aecfdea1f7ffbff5a81fad820c73a5a49161771dc2d95130d1a50"} Dec 06 04:12:09 crc kubenswrapper[4802]: I1206 04:12:09.271462 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a983ff46e59aecfdea1f7ffbff5a81fad820c73a5a49161771dc2d95130d1a50" Dec 06 04:12:09 crc kubenswrapper[4802]: I1206 04:12:09.271488 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g9qlw" Dec 06 04:12:09 crc kubenswrapper[4802]: I1206 04:12:09.391121 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zd97g"] Dec 06 04:12:09 crc kubenswrapper[4802]: E1206 04:12:09.391798 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6c3bbd2-7993-420e-91d7-ffc72b1d2439" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 06 04:12:09 crc kubenswrapper[4802]: I1206 04:12:09.391888 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6c3bbd2-7993-420e-91d7-ffc72b1d2439" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 06 04:12:09 crc kubenswrapper[4802]: I1206 04:12:09.392239 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6c3bbd2-7993-420e-91d7-ffc72b1d2439" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 06 04:12:09 crc kubenswrapper[4802]: I1206 04:12:09.393265 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zd97g" Dec 06 04:12:09 crc kubenswrapper[4802]: I1206 04:12:09.395363 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 04:12:09 crc kubenswrapper[4802]: I1206 04:12:09.396072 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-rgvjc" Dec 06 04:12:09 crc kubenswrapper[4802]: I1206 04:12:09.396769 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 04:12:09 crc kubenswrapper[4802]: I1206 04:12:09.402536 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 04:12:09 crc kubenswrapper[4802]: I1206 04:12:09.410419 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zd97g"] Dec 06 04:12:09 crc kubenswrapper[4802]: I1206 04:12:09.456222 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8nvm\" (UniqueName: \"kubernetes.io/projected/3d1e4ffc-0f3b-4bc2-9f36-6193b7f33785-kube-api-access-j8nvm\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-zd97g\" (UID: \"3d1e4ffc-0f3b-4bc2-9f36-6193b7f33785\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zd97g" Dec 06 04:12:09 crc kubenswrapper[4802]: I1206 04:12:09.456285 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3d1e4ffc-0f3b-4bc2-9f36-6193b7f33785-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-zd97g\" (UID: \"3d1e4ffc-0f3b-4bc2-9f36-6193b7f33785\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zd97g" Dec 06 04:12:09 crc kubenswrapper[4802]: I1206 04:12:09.456399 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3d1e4ffc-0f3b-4bc2-9f36-6193b7f33785-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-zd97g\" (UID: \"3d1e4ffc-0f3b-4bc2-9f36-6193b7f33785\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zd97g" Dec 06 04:12:09 crc kubenswrapper[4802]: I1206 04:12:09.558302 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8nvm\" (UniqueName: \"kubernetes.io/projected/3d1e4ffc-0f3b-4bc2-9f36-6193b7f33785-kube-api-access-j8nvm\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-zd97g\" (UID: \"3d1e4ffc-0f3b-4bc2-9f36-6193b7f33785\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zd97g" Dec 06 04:12:09 crc kubenswrapper[4802]: I1206 04:12:09.558897 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3d1e4ffc-0f3b-4bc2-9f36-6193b7f33785-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-zd97g\" (UID: \"3d1e4ffc-0f3b-4bc2-9f36-6193b7f33785\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zd97g" Dec 06 04:12:09 crc kubenswrapper[4802]: I1206 04:12:09.558942 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3d1e4ffc-0f3b-4bc2-9f36-6193b7f33785-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-zd97g\" (UID: \"3d1e4ffc-0f3b-4bc2-9f36-6193b7f33785\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zd97g" Dec 06 04:12:09 crc kubenswrapper[4802]: I1206 04:12:09.564486 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3d1e4ffc-0f3b-4bc2-9f36-6193b7f33785-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-zd97g\" (UID: \"3d1e4ffc-0f3b-4bc2-9f36-6193b7f33785\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zd97g" Dec 06 04:12:09 crc kubenswrapper[4802]: I1206 04:12:09.564520 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3d1e4ffc-0f3b-4bc2-9f36-6193b7f33785-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-zd97g\" (UID: \"3d1e4ffc-0f3b-4bc2-9f36-6193b7f33785\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zd97g" Dec 06 04:12:09 crc kubenswrapper[4802]: I1206 04:12:09.582644 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8nvm\" (UniqueName: \"kubernetes.io/projected/3d1e4ffc-0f3b-4bc2-9f36-6193b7f33785-kube-api-access-j8nvm\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-zd97g\" (UID: \"3d1e4ffc-0f3b-4bc2-9f36-6193b7f33785\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zd97g" Dec 06 04:12:09 crc kubenswrapper[4802]: I1206 04:12:09.712704 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zd97g" Dec 06 04:12:10 crc kubenswrapper[4802]: I1206 04:12:10.153067 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zd97g"] Dec 06 04:12:10 crc kubenswrapper[4802]: I1206 04:12:10.279818 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zd97g" event={"ID":"3d1e4ffc-0f3b-4bc2-9f36-6193b7f33785","Type":"ContainerStarted","Data":"38133c5f535fb374fd93d64276fd5fec32b7f6c64589d49be223c4fe3c5e7edf"} Dec 06 04:12:11 crc kubenswrapper[4802]: I1206 04:12:11.075799 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-l4lf9"] Dec 06 04:12:11 crc kubenswrapper[4802]: I1206 04:12:11.085928 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-l4lf9"] Dec 06 04:12:11 crc kubenswrapper[4802]: I1206 04:12:11.460274 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db51698a-4eea-44d9-bbe1-9dab86cd7a02" path="/var/lib/kubelet/pods/db51698a-4eea-44d9-bbe1-9dab86cd7a02/volumes" Dec 06 04:12:13 crc kubenswrapper[4802]: I1206 04:12:13.316813 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zd97g" event={"ID":"3d1e4ffc-0f3b-4bc2-9f36-6193b7f33785","Type":"ContainerStarted","Data":"027801b009c7626b47cb5de2ce0efa47dd5b68cb2dc49c4a01c2281c57492973"} Dec 06 04:12:13 crc kubenswrapper[4802]: I1206 04:12:13.342294 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zd97g" podStartSLOduration=2.319118949 podStartE2EDuration="4.342276792s" podCreationTimestamp="2025-12-06 04:12:09 +0000 UTC" firstStartedPulling="2025-12-06 04:12:10.15528009 +0000 UTC m=+1923.027189242" lastFinishedPulling="2025-12-06 04:12:12.178437933 +0000 UTC m=+1925.050347085" observedRunningTime="2025-12-06 04:12:13.336117575 +0000 UTC m=+1926.208026737" watchObservedRunningTime="2025-12-06 04:12:13.342276792 +0000 UTC m=+1926.214185954" Dec 06 04:12:21 crc kubenswrapper[4802]: I1206 04:12:21.863741 4802 scope.go:117] "RemoveContainer" containerID="8e649e34b5ec7777e026185ed91acf6eb44aeaded3f44a8818c9d17502eb366d" Dec 06 04:12:22 crc kubenswrapper[4802]: I1206 04:12:22.085603 4802 scope.go:117] "RemoveContainer" containerID="3a7525fcbc73d3441910081d90724b15d3f5451db8ed85974e99baea473d1247" Dec 06 04:12:22 crc kubenswrapper[4802]: I1206 04:12:22.142587 4802 scope.go:117] "RemoveContainer" containerID="93295a23f9c3d49cfdc9948bfe05cc2796671712333ba33a3f9f42614887987b" Dec 06 04:13:01 crc kubenswrapper[4802]: I1206 04:13:01.770631 4802 generic.go:334] "Generic (PLEG): container finished" podID="3d1e4ffc-0f3b-4bc2-9f36-6193b7f33785" containerID="027801b009c7626b47cb5de2ce0efa47dd5b68cb2dc49c4a01c2281c57492973" exitCode=0 Dec 06 04:13:01 crc kubenswrapper[4802]: I1206 04:13:01.770862 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zd97g" event={"ID":"3d1e4ffc-0f3b-4bc2-9f36-6193b7f33785","Type":"ContainerDied","Data":"027801b009c7626b47cb5de2ce0efa47dd5b68cb2dc49c4a01c2281c57492973"} Dec 06 04:13:03 crc kubenswrapper[4802]: I1206 04:13:03.218293 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zd97g" Dec 06 04:13:03 crc kubenswrapper[4802]: I1206 04:13:03.271594 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3d1e4ffc-0f3b-4bc2-9f36-6193b7f33785-inventory\") pod \"3d1e4ffc-0f3b-4bc2-9f36-6193b7f33785\" (UID: \"3d1e4ffc-0f3b-4bc2-9f36-6193b7f33785\") " Dec 06 04:13:03 crc kubenswrapper[4802]: I1206 04:13:03.271707 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j8nvm\" (UniqueName: \"kubernetes.io/projected/3d1e4ffc-0f3b-4bc2-9f36-6193b7f33785-kube-api-access-j8nvm\") pod \"3d1e4ffc-0f3b-4bc2-9f36-6193b7f33785\" (UID: \"3d1e4ffc-0f3b-4bc2-9f36-6193b7f33785\") " Dec 06 04:13:03 crc kubenswrapper[4802]: I1206 04:13:03.271861 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3d1e4ffc-0f3b-4bc2-9f36-6193b7f33785-ssh-key\") pod \"3d1e4ffc-0f3b-4bc2-9f36-6193b7f33785\" (UID: \"3d1e4ffc-0f3b-4bc2-9f36-6193b7f33785\") " Dec 06 04:13:03 crc kubenswrapper[4802]: I1206 04:13:03.277741 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d1e4ffc-0f3b-4bc2-9f36-6193b7f33785-kube-api-access-j8nvm" (OuterVolumeSpecName: "kube-api-access-j8nvm") pod "3d1e4ffc-0f3b-4bc2-9f36-6193b7f33785" (UID: "3d1e4ffc-0f3b-4bc2-9f36-6193b7f33785"). InnerVolumeSpecName "kube-api-access-j8nvm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:13:03 crc kubenswrapper[4802]: I1206 04:13:03.298068 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d1e4ffc-0f3b-4bc2-9f36-6193b7f33785-inventory" (OuterVolumeSpecName: "inventory") pod "3d1e4ffc-0f3b-4bc2-9f36-6193b7f33785" (UID: "3d1e4ffc-0f3b-4bc2-9f36-6193b7f33785"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:13:03 crc kubenswrapper[4802]: I1206 04:13:03.305036 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d1e4ffc-0f3b-4bc2-9f36-6193b7f33785-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3d1e4ffc-0f3b-4bc2-9f36-6193b7f33785" (UID: "3d1e4ffc-0f3b-4bc2-9f36-6193b7f33785"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:13:03 crc kubenswrapper[4802]: I1206 04:13:03.374379 4802 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3d1e4ffc-0f3b-4bc2-9f36-6193b7f33785-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 04:13:03 crc kubenswrapper[4802]: I1206 04:13:03.374437 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j8nvm\" (UniqueName: \"kubernetes.io/projected/3d1e4ffc-0f3b-4bc2-9f36-6193b7f33785-kube-api-access-j8nvm\") on node \"crc\" DevicePath \"\"" Dec 06 04:13:03 crc kubenswrapper[4802]: I1206 04:13:03.374447 4802 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3d1e4ffc-0f3b-4bc2-9f36-6193b7f33785-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 04:13:03 crc kubenswrapper[4802]: I1206 04:13:03.791976 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zd97g" event={"ID":"3d1e4ffc-0f3b-4bc2-9f36-6193b7f33785","Type":"ContainerDied","Data":"38133c5f535fb374fd93d64276fd5fec32b7f6c64589d49be223c4fe3c5e7edf"} Dec 06 04:13:03 crc kubenswrapper[4802]: I1206 04:13:03.792045 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="38133c5f535fb374fd93d64276fd5fec32b7f6c64589d49be223c4fe3c5e7edf" Dec 06 04:13:03 crc kubenswrapper[4802]: I1206 04:13:03.792182 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zd97g" Dec 06 04:13:03 crc kubenswrapper[4802]: I1206 04:13:03.898809 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-hgx7d"] Dec 06 04:13:03 crc kubenswrapper[4802]: E1206 04:13:03.899299 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d1e4ffc-0f3b-4bc2-9f36-6193b7f33785" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 06 04:13:03 crc kubenswrapper[4802]: I1206 04:13:03.899323 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d1e4ffc-0f3b-4bc2-9f36-6193b7f33785" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 06 04:13:03 crc kubenswrapper[4802]: I1206 04:13:03.899520 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d1e4ffc-0f3b-4bc2-9f36-6193b7f33785" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 06 04:13:03 crc kubenswrapper[4802]: I1206 04:13:03.901540 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-hgx7d" Dec 06 04:13:03 crc kubenswrapper[4802]: I1206 04:13:03.903857 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-rgvjc" Dec 06 04:13:03 crc kubenswrapper[4802]: I1206 04:13:03.904884 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 04:13:03 crc kubenswrapper[4802]: I1206 04:13:03.905024 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 04:13:03 crc kubenswrapper[4802]: I1206 04:13:03.905661 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 04:13:03 crc kubenswrapper[4802]: I1206 04:13:03.914909 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-hgx7d"] Dec 06 04:13:03 crc kubenswrapper[4802]: I1206 04:13:03.985109 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/815e5d24-83cc-4bbe-b602-6cd1da82039b-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-hgx7d\" (UID: \"815e5d24-83cc-4bbe-b602-6cd1da82039b\") " pod="openstack/ssh-known-hosts-edpm-deployment-hgx7d" Dec 06 04:13:03 crc kubenswrapper[4802]: I1206 04:13:03.985287 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/815e5d24-83cc-4bbe-b602-6cd1da82039b-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-hgx7d\" (UID: \"815e5d24-83cc-4bbe-b602-6cd1da82039b\") " pod="openstack/ssh-known-hosts-edpm-deployment-hgx7d" Dec 06 04:13:03 crc kubenswrapper[4802]: I1206 04:13:03.985337 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p69vc\" (UniqueName: \"kubernetes.io/projected/815e5d24-83cc-4bbe-b602-6cd1da82039b-kube-api-access-p69vc\") pod \"ssh-known-hosts-edpm-deployment-hgx7d\" (UID: \"815e5d24-83cc-4bbe-b602-6cd1da82039b\") " pod="openstack/ssh-known-hosts-edpm-deployment-hgx7d" Dec 06 04:13:04 crc kubenswrapper[4802]: I1206 04:13:04.087309 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/815e5d24-83cc-4bbe-b602-6cd1da82039b-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-hgx7d\" (UID: \"815e5d24-83cc-4bbe-b602-6cd1da82039b\") " pod="openstack/ssh-known-hosts-edpm-deployment-hgx7d" Dec 06 04:13:04 crc kubenswrapper[4802]: I1206 04:13:04.087654 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p69vc\" (UniqueName: \"kubernetes.io/projected/815e5d24-83cc-4bbe-b602-6cd1da82039b-kube-api-access-p69vc\") pod \"ssh-known-hosts-edpm-deployment-hgx7d\" (UID: \"815e5d24-83cc-4bbe-b602-6cd1da82039b\") " pod="openstack/ssh-known-hosts-edpm-deployment-hgx7d" Dec 06 04:13:04 crc kubenswrapper[4802]: I1206 04:13:04.087770 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/815e5d24-83cc-4bbe-b602-6cd1da82039b-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-hgx7d\" (UID: \"815e5d24-83cc-4bbe-b602-6cd1da82039b\") " pod="openstack/ssh-known-hosts-edpm-deployment-hgx7d" Dec 06 04:13:04 crc kubenswrapper[4802]: I1206 04:13:04.091529 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/815e5d24-83cc-4bbe-b602-6cd1da82039b-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-hgx7d\" (UID: \"815e5d24-83cc-4bbe-b602-6cd1da82039b\") " pod="openstack/ssh-known-hosts-edpm-deployment-hgx7d" Dec 06 04:13:04 crc kubenswrapper[4802]: I1206 04:13:04.095311 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/815e5d24-83cc-4bbe-b602-6cd1da82039b-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-hgx7d\" (UID: \"815e5d24-83cc-4bbe-b602-6cd1da82039b\") " pod="openstack/ssh-known-hosts-edpm-deployment-hgx7d" Dec 06 04:13:04 crc kubenswrapper[4802]: I1206 04:13:04.104636 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p69vc\" (UniqueName: \"kubernetes.io/projected/815e5d24-83cc-4bbe-b602-6cd1da82039b-kube-api-access-p69vc\") pod \"ssh-known-hosts-edpm-deployment-hgx7d\" (UID: \"815e5d24-83cc-4bbe-b602-6cd1da82039b\") " pod="openstack/ssh-known-hosts-edpm-deployment-hgx7d" Dec 06 04:13:04 crc kubenswrapper[4802]: I1206 04:13:04.218743 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-hgx7d" Dec 06 04:13:04 crc kubenswrapper[4802]: I1206 04:13:04.816625 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-hgx7d"] Dec 06 04:13:05 crc kubenswrapper[4802]: I1206 04:13:05.809632 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-hgx7d" event={"ID":"815e5d24-83cc-4bbe-b602-6cd1da82039b","Type":"ContainerStarted","Data":"ebc7f0b44abf895bf784208241a83b878a9869653a50e8bf4c713ab0e510aa8c"} Dec 06 04:13:05 crc kubenswrapper[4802]: I1206 04:13:05.810003 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-hgx7d" event={"ID":"815e5d24-83cc-4bbe-b602-6cd1da82039b","Type":"ContainerStarted","Data":"6d73a82e4bff4a8630014d94fc54d08d9ebd0ad50e1b1b65478e5f87473c506d"} Dec 06 04:13:05 crc kubenswrapper[4802]: I1206 04:13:05.841970 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-hgx7d" podStartSLOduration=2.445394068 podStartE2EDuration="2.841729993s" podCreationTimestamp="2025-12-06 04:13:03 +0000 UTC" firstStartedPulling="2025-12-06 04:13:04.81614743 +0000 UTC m=+1977.688056622" lastFinishedPulling="2025-12-06 04:13:05.212483405 +0000 UTC m=+1978.084392547" observedRunningTime="2025-12-06 04:13:05.831398732 +0000 UTC m=+1978.703307894" watchObservedRunningTime="2025-12-06 04:13:05.841729993 +0000 UTC m=+1978.713639175" Dec 06 04:13:12 crc kubenswrapper[4802]: I1206 04:13:12.881405 4802 generic.go:334] "Generic (PLEG): container finished" podID="815e5d24-83cc-4bbe-b602-6cd1da82039b" containerID="ebc7f0b44abf895bf784208241a83b878a9869653a50e8bf4c713ab0e510aa8c" exitCode=0 Dec 06 04:13:12 crc kubenswrapper[4802]: I1206 04:13:12.881495 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-hgx7d" event={"ID":"815e5d24-83cc-4bbe-b602-6cd1da82039b","Type":"ContainerDied","Data":"ebc7f0b44abf895bf784208241a83b878a9869653a50e8bf4c713ab0e510aa8c"} Dec 06 04:13:13 crc kubenswrapper[4802]: I1206 04:13:13.283491 4802 patch_prober.go:28] interesting pod/machine-config-daemon-zpxxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:13:13 crc kubenswrapper[4802]: I1206 04:13:13.283586 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:13:14 crc kubenswrapper[4802]: I1206 04:13:14.329737 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-hgx7d" Dec 06 04:13:14 crc kubenswrapper[4802]: I1206 04:13:14.425235 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/815e5d24-83cc-4bbe-b602-6cd1da82039b-inventory-0\") pod \"815e5d24-83cc-4bbe-b602-6cd1da82039b\" (UID: \"815e5d24-83cc-4bbe-b602-6cd1da82039b\") " Dec 06 04:13:14 crc kubenswrapper[4802]: I1206 04:13:14.425654 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p69vc\" (UniqueName: \"kubernetes.io/projected/815e5d24-83cc-4bbe-b602-6cd1da82039b-kube-api-access-p69vc\") pod \"815e5d24-83cc-4bbe-b602-6cd1da82039b\" (UID: \"815e5d24-83cc-4bbe-b602-6cd1da82039b\") " Dec 06 04:13:14 crc kubenswrapper[4802]: I1206 04:13:14.425791 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/815e5d24-83cc-4bbe-b602-6cd1da82039b-ssh-key-openstack-edpm-ipam\") pod \"815e5d24-83cc-4bbe-b602-6cd1da82039b\" (UID: \"815e5d24-83cc-4bbe-b602-6cd1da82039b\") " Dec 06 04:13:14 crc kubenswrapper[4802]: I1206 04:13:14.441316 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/815e5d24-83cc-4bbe-b602-6cd1da82039b-kube-api-access-p69vc" (OuterVolumeSpecName: "kube-api-access-p69vc") pod "815e5d24-83cc-4bbe-b602-6cd1da82039b" (UID: "815e5d24-83cc-4bbe-b602-6cd1da82039b"). InnerVolumeSpecName "kube-api-access-p69vc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:13:14 crc kubenswrapper[4802]: I1206 04:13:14.462236 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/815e5d24-83cc-4bbe-b602-6cd1da82039b-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "815e5d24-83cc-4bbe-b602-6cd1da82039b" (UID: "815e5d24-83cc-4bbe-b602-6cd1da82039b"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:13:14 crc kubenswrapper[4802]: I1206 04:13:14.473006 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/815e5d24-83cc-4bbe-b602-6cd1da82039b-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "815e5d24-83cc-4bbe-b602-6cd1da82039b" (UID: "815e5d24-83cc-4bbe-b602-6cd1da82039b"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:13:14 crc kubenswrapper[4802]: I1206 04:13:14.528283 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p69vc\" (UniqueName: \"kubernetes.io/projected/815e5d24-83cc-4bbe-b602-6cd1da82039b-kube-api-access-p69vc\") on node \"crc\" DevicePath \"\"" Dec 06 04:13:14 crc kubenswrapper[4802]: I1206 04:13:14.528324 4802 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/815e5d24-83cc-4bbe-b602-6cd1da82039b-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 06 04:13:14 crc kubenswrapper[4802]: I1206 04:13:14.528340 4802 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/815e5d24-83cc-4bbe-b602-6cd1da82039b-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 06 04:13:14 crc kubenswrapper[4802]: I1206 04:13:14.903684 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-hgx7d" event={"ID":"815e5d24-83cc-4bbe-b602-6cd1da82039b","Type":"ContainerDied","Data":"6d73a82e4bff4a8630014d94fc54d08d9ebd0ad50e1b1b65478e5f87473c506d"} Dec 06 04:13:14 crc kubenswrapper[4802]: I1206 04:13:14.903776 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-hgx7d" Dec 06 04:13:14 crc kubenswrapper[4802]: I1206 04:13:14.903786 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6d73a82e4bff4a8630014d94fc54d08d9ebd0ad50e1b1b65478e5f87473c506d" Dec 06 04:13:14 crc kubenswrapper[4802]: I1206 04:13:14.986972 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-4clcr"] Dec 06 04:13:14 crc kubenswrapper[4802]: E1206 04:13:14.987512 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="815e5d24-83cc-4bbe-b602-6cd1da82039b" containerName="ssh-known-hosts-edpm-deployment" Dec 06 04:13:14 crc kubenswrapper[4802]: I1206 04:13:14.987532 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="815e5d24-83cc-4bbe-b602-6cd1da82039b" containerName="ssh-known-hosts-edpm-deployment" Dec 06 04:13:14 crc kubenswrapper[4802]: I1206 04:13:14.987822 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="815e5d24-83cc-4bbe-b602-6cd1da82039b" containerName="ssh-known-hosts-edpm-deployment" Dec 06 04:13:14 crc kubenswrapper[4802]: I1206 04:13:14.988643 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-4clcr" Dec 06 04:13:14 crc kubenswrapper[4802]: I1206 04:13:14.990625 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 04:13:14 crc kubenswrapper[4802]: I1206 04:13:14.990648 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 04:13:14 crc kubenswrapper[4802]: I1206 04:13:14.991410 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-rgvjc" Dec 06 04:13:14 crc kubenswrapper[4802]: I1206 04:13:14.993305 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 04:13:15 crc kubenswrapper[4802]: I1206 04:13:15.000298 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-4clcr"] Dec 06 04:13:15 crc kubenswrapper[4802]: I1206 04:13:15.039100 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bcbbe04a-d1ab-4a9b-a4ff-11806f4928c8-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-4clcr\" (UID: \"bcbbe04a-d1ab-4a9b-a4ff-11806f4928c8\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-4clcr" Dec 06 04:13:15 crc kubenswrapper[4802]: I1206 04:13:15.039178 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bcbbe04a-d1ab-4a9b-a4ff-11806f4928c8-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-4clcr\" (UID: \"bcbbe04a-d1ab-4a9b-a4ff-11806f4928c8\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-4clcr" Dec 06 04:13:15 crc kubenswrapper[4802]: I1206 04:13:15.039227 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6hc6\" (UniqueName: \"kubernetes.io/projected/bcbbe04a-d1ab-4a9b-a4ff-11806f4928c8-kube-api-access-c6hc6\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-4clcr\" (UID: \"bcbbe04a-d1ab-4a9b-a4ff-11806f4928c8\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-4clcr" Dec 06 04:13:15 crc kubenswrapper[4802]: I1206 04:13:15.141561 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bcbbe04a-d1ab-4a9b-a4ff-11806f4928c8-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-4clcr\" (UID: \"bcbbe04a-d1ab-4a9b-a4ff-11806f4928c8\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-4clcr" Dec 06 04:13:15 crc kubenswrapper[4802]: I1206 04:13:15.141653 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bcbbe04a-d1ab-4a9b-a4ff-11806f4928c8-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-4clcr\" (UID: \"bcbbe04a-d1ab-4a9b-a4ff-11806f4928c8\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-4clcr" Dec 06 04:13:15 crc kubenswrapper[4802]: I1206 04:13:15.141696 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6hc6\" (UniqueName: \"kubernetes.io/projected/bcbbe04a-d1ab-4a9b-a4ff-11806f4928c8-kube-api-access-c6hc6\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-4clcr\" (UID: \"bcbbe04a-d1ab-4a9b-a4ff-11806f4928c8\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-4clcr" Dec 06 04:13:15 crc kubenswrapper[4802]: I1206 04:13:15.146240 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bcbbe04a-d1ab-4a9b-a4ff-11806f4928c8-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-4clcr\" (UID: \"bcbbe04a-d1ab-4a9b-a4ff-11806f4928c8\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-4clcr" Dec 06 04:13:15 crc kubenswrapper[4802]: I1206 04:13:15.153370 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bcbbe04a-d1ab-4a9b-a4ff-11806f4928c8-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-4clcr\" (UID: \"bcbbe04a-d1ab-4a9b-a4ff-11806f4928c8\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-4clcr" Dec 06 04:13:15 crc kubenswrapper[4802]: I1206 04:13:15.161058 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6hc6\" (UniqueName: \"kubernetes.io/projected/bcbbe04a-d1ab-4a9b-a4ff-11806f4928c8-kube-api-access-c6hc6\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-4clcr\" (UID: \"bcbbe04a-d1ab-4a9b-a4ff-11806f4928c8\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-4clcr" Dec 06 04:13:15 crc kubenswrapper[4802]: I1206 04:13:15.310923 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-4clcr" Dec 06 04:13:15 crc kubenswrapper[4802]: I1206 04:13:15.825058 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-4clcr"] Dec 06 04:13:15 crc kubenswrapper[4802]: W1206 04:13:15.827702 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbcbbe04a_d1ab_4a9b_a4ff_11806f4928c8.slice/crio-783b9fa875b6e12129a75d3a38d22dcdc3a2978172f3343f118568f94d02a10f WatchSource:0}: Error finding container 783b9fa875b6e12129a75d3a38d22dcdc3a2978172f3343f118568f94d02a10f: Status 404 returned error can't find the container with id 783b9fa875b6e12129a75d3a38d22dcdc3a2978172f3343f118568f94d02a10f Dec 06 04:13:15 crc kubenswrapper[4802]: I1206 04:13:15.913363 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-4clcr" event={"ID":"bcbbe04a-d1ab-4a9b-a4ff-11806f4928c8","Type":"ContainerStarted","Data":"783b9fa875b6e12129a75d3a38d22dcdc3a2978172f3343f118568f94d02a10f"} Dec 06 04:13:16 crc kubenswrapper[4802]: I1206 04:13:16.929065 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-4clcr" event={"ID":"bcbbe04a-d1ab-4a9b-a4ff-11806f4928c8","Type":"ContainerStarted","Data":"2f444a12b3c9371fcff71a178c0e9fa16a40dd87afd13b2b5fb778bd58440108"} Dec 06 04:13:16 crc kubenswrapper[4802]: I1206 04:13:16.984438 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-4clcr" podStartSLOduration=2.581860883 podStartE2EDuration="2.984416747s" podCreationTimestamp="2025-12-06 04:13:14 +0000 UTC" firstStartedPulling="2025-12-06 04:13:15.830214361 +0000 UTC m=+1988.702123513" lastFinishedPulling="2025-12-06 04:13:16.232770225 +0000 UTC m=+1989.104679377" observedRunningTime="2025-12-06 04:13:16.947847551 +0000 UTC m=+1989.819756733" watchObservedRunningTime="2025-12-06 04:13:16.984416747 +0000 UTC m=+1989.856325899" Dec 06 04:13:24 crc kubenswrapper[4802]: I1206 04:13:24.998526 4802 generic.go:334] "Generic (PLEG): container finished" podID="bcbbe04a-d1ab-4a9b-a4ff-11806f4928c8" containerID="2f444a12b3c9371fcff71a178c0e9fa16a40dd87afd13b2b5fb778bd58440108" exitCode=0 Dec 06 04:13:24 crc kubenswrapper[4802]: I1206 04:13:24.998583 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-4clcr" event={"ID":"bcbbe04a-d1ab-4a9b-a4ff-11806f4928c8","Type":"ContainerDied","Data":"2f444a12b3c9371fcff71a178c0e9fa16a40dd87afd13b2b5fb778bd58440108"} Dec 06 04:13:26 crc kubenswrapper[4802]: I1206 04:13:26.401811 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-4clcr" Dec 06 04:13:26 crc kubenswrapper[4802]: I1206 04:13:26.499737 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c6hc6\" (UniqueName: \"kubernetes.io/projected/bcbbe04a-d1ab-4a9b-a4ff-11806f4928c8-kube-api-access-c6hc6\") pod \"bcbbe04a-d1ab-4a9b-a4ff-11806f4928c8\" (UID: \"bcbbe04a-d1ab-4a9b-a4ff-11806f4928c8\") " Dec 06 04:13:26 crc kubenswrapper[4802]: I1206 04:13:26.499906 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bcbbe04a-d1ab-4a9b-a4ff-11806f4928c8-ssh-key\") pod \"bcbbe04a-d1ab-4a9b-a4ff-11806f4928c8\" (UID: \"bcbbe04a-d1ab-4a9b-a4ff-11806f4928c8\") " Dec 06 04:13:26 crc kubenswrapper[4802]: I1206 04:13:26.499938 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bcbbe04a-d1ab-4a9b-a4ff-11806f4928c8-inventory\") pod \"bcbbe04a-d1ab-4a9b-a4ff-11806f4928c8\" (UID: \"bcbbe04a-d1ab-4a9b-a4ff-11806f4928c8\") " Dec 06 04:13:26 crc kubenswrapper[4802]: I1206 04:13:26.506034 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcbbe04a-d1ab-4a9b-a4ff-11806f4928c8-kube-api-access-c6hc6" (OuterVolumeSpecName: "kube-api-access-c6hc6") pod "bcbbe04a-d1ab-4a9b-a4ff-11806f4928c8" (UID: "bcbbe04a-d1ab-4a9b-a4ff-11806f4928c8"). InnerVolumeSpecName "kube-api-access-c6hc6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:13:26 crc kubenswrapper[4802]: I1206 04:13:26.547966 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcbbe04a-d1ab-4a9b-a4ff-11806f4928c8-inventory" (OuterVolumeSpecName: "inventory") pod "bcbbe04a-d1ab-4a9b-a4ff-11806f4928c8" (UID: "bcbbe04a-d1ab-4a9b-a4ff-11806f4928c8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:13:26 crc kubenswrapper[4802]: I1206 04:13:26.548960 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcbbe04a-d1ab-4a9b-a4ff-11806f4928c8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "bcbbe04a-d1ab-4a9b-a4ff-11806f4928c8" (UID: "bcbbe04a-d1ab-4a9b-a4ff-11806f4928c8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:13:26 crc kubenswrapper[4802]: I1206 04:13:26.604922 4802 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bcbbe04a-d1ab-4a9b-a4ff-11806f4928c8-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 04:13:26 crc kubenswrapper[4802]: I1206 04:13:26.604970 4802 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bcbbe04a-d1ab-4a9b-a4ff-11806f4928c8-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 04:13:26 crc kubenswrapper[4802]: I1206 04:13:26.604988 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c6hc6\" (UniqueName: \"kubernetes.io/projected/bcbbe04a-d1ab-4a9b-a4ff-11806f4928c8-kube-api-access-c6hc6\") on node \"crc\" DevicePath \"\"" Dec 06 04:13:27 crc kubenswrapper[4802]: I1206 04:13:27.025697 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-4clcr" event={"ID":"bcbbe04a-d1ab-4a9b-a4ff-11806f4928c8","Type":"ContainerDied","Data":"783b9fa875b6e12129a75d3a38d22dcdc3a2978172f3343f118568f94d02a10f"} Dec 06 04:13:27 crc kubenswrapper[4802]: I1206 04:13:27.025975 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="783b9fa875b6e12129a75d3a38d22dcdc3a2978172f3343f118568f94d02a10f" Dec 06 04:13:27 crc kubenswrapper[4802]: I1206 04:13:27.025771 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-4clcr" Dec 06 04:13:27 crc kubenswrapper[4802]: I1206 04:13:27.111391 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7hgqr"] Dec 06 04:13:27 crc kubenswrapper[4802]: E1206 04:13:27.112053 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcbbe04a-d1ab-4a9b-a4ff-11806f4928c8" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 06 04:13:27 crc kubenswrapper[4802]: I1206 04:13:27.112088 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcbbe04a-d1ab-4a9b-a4ff-11806f4928c8" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 06 04:13:27 crc kubenswrapper[4802]: I1206 04:13:27.112424 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcbbe04a-d1ab-4a9b-a4ff-11806f4928c8" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 06 04:13:27 crc kubenswrapper[4802]: I1206 04:13:27.125342 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7hgqr"] Dec 06 04:13:27 crc kubenswrapper[4802]: I1206 04:13:27.125444 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7hgqr" Dec 06 04:13:27 crc kubenswrapper[4802]: I1206 04:13:27.134196 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-rgvjc" Dec 06 04:13:27 crc kubenswrapper[4802]: I1206 04:13:27.134579 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 04:13:27 crc kubenswrapper[4802]: I1206 04:13:27.134741 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 04:13:27 crc kubenswrapper[4802]: I1206 04:13:27.134968 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 04:13:27 crc kubenswrapper[4802]: I1206 04:13:27.220226 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmzzd\" (UniqueName: \"kubernetes.io/projected/0de67972-73b4-4f71-aca7-d77ae62d73b3-kube-api-access-zmzzd\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7hgqr\" (UID: \"0de67972-73b4-4f71-aca7-d77ae62d73b3\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7hgqr" Dec 06 04:13:27 crc kubenswrapper[4802]: I1206 04:13:27.220531 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0de67972-73b4-4f71-aca7-d77ae62d73b3-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7hgqr\" (UID: \"0de67972-73b4-4f71-aca7-d77ae62d73b3\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7hgqr" Dec 06 04:13:27 crc kubenswrapper[4802]: I1206 04:13:27.220687 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0de67972-73b4-4f71-aca7-d77ae62d73b3-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7hgqr\" (UID: \"0de67972-73b4-4f71-aca7-d77ae62d73b3\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7hgqr" Dec 06 04:13:27 crc kubenswrapper[4802]: I1206 04:13:27.322362 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0de67972-73b4-4f71-aca7-d77ae62d73b3-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7hgqr\" (UID: \"0de67972-73b4-4f71-aca7-d77ae62d73b3\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7hgqr" Dec 06 04:13:27 crc kubenswrapper[4802]: I1206 04:13:27.322443 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0de67972-73b4-4f71-aca7-d77ae62d73b3-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7hgqr\" (UID: \"0de67972-73b4-4f71-aca7-d77ae62d73b3\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7hgqr" Dec 06 04:13:27 crc kubenswrapper[4802]: I1206 04:13:27.322561 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmzzd\" (UniqueName: \"kubernetes.io/projected/0de67972-73b4-4f71-aca7-d77ae62d73b3-kube-api-access-zmzzd\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7hgqr\" (UID: \"0de67972-73b4-4f71-aca7-d77ae62d73b3\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7hgqr" Dec 06 04:13:27 crc kubenswrapper[4802]: I1206 04:13:27.333659 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0de67972-73b4-4f71-aca7-d77ae62d73b3-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7hgqr\" (UID: \"0de67972-73b4-4f71-aca7-d77ae62d73b3\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7hgqr" Dec 06 04:13:27 crc kubenswrapper[4802]: I1206 04:13:27.333842 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0de67972-73b4-4f71-aca7-d77ae62d73b3-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7hgqr\" (UID: \"0de67972-73b4-4f71-aca7-d77ae62d73b3\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7hgqr" Dec 06 04:13:27 crc kubenswrapper[4802]: I1206 04:13:27.346003 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmzzd\" (UniqueName: \"kubernetes.io/projected/0de67972-73b4-4f71-aca7-d77ae62d73b3-kube-api-access-zmzzd\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7hgqr\" (UID: \"0de67972-73b4-4f71-aca7-d77ae62d73b3\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7hgqr" Dec 06 04:13:27 crc kubenswrapper[4802]: I1206 04:13:27.444562 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7hgqr" Dec 06 04:13:27 crc kubenswrapper[4802]: I1206 04:13:27.946906 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7hgqr"] Dec 06 04:13:28 crc kubenswrapper[4802]: I1206 04:13:28.036639 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7hgqr" event={"ID":"0de67972-73b4-4f71-aca7-d77ae62d73b3","Type":"ContainerStarted","Data":"42e3f071a4995f47ffca00fc1f93749f284ee31333aeac02ba8939bfbfdd6490"} Dec 06 04:13:29 crc kubenswrapper[4802]: I1206 04:13:29.049113 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7hgqr" event={"ID":"0de67972-73b4-4f71-aca7-d77ae62d73b3","Type":"ContainerStarted","Data":"bfa5baf6ca685083911fa2672455b7c76cab9733fcaafc83addea3d1d3399a22"} Dec 06 04:13:29 crc kubenswrapper[4802]: I1206 04:13:29.069889 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7hgqr" podStartSLOduration=1.629769369 podStartE2EDuration="2.069870277s" podCreationTimestamp="2025-12-06 04:13:27 +0000 UTC" firstStartedPulling="2025-12-06 04:13:27.950828638 +0000 UTC m=+2000.822737800" lastFinishedPulling="2025-12-06 04:13:28.390929536 +0000 UTC m=+2001.262838708" observedRunningTime="2025-12-06 04:13:29.066101745 +0000 UTC m=+2001.938010907" watchObservedRunningTime="2025-12-06 04:13:29.069870277 +0000 UTC m=+2001.941779429" Dec 06 04:13:38 crc kubenswrapper[4802]: I1206 04:13:38.131092 4802 generic.go:334] "Generic (PLEG): container finished" podID="0de67972-73b4-4f71-aca7-d77ae62d73b3" containerID="bfa5baf6ca685083911fa2672455b7c76cab9733fcaafc83addea3d1d3399a22" exitCode=0 Dec 06 04:13:38 crc kubenswrapper[4802]: I1206 04:13:38.131196 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7hgqr" event={"ID":"0de67972-73b4-4f71-aca7-d77ae62d73b3","Type":"ContainerDied","Data":"bfa5baf6ca685083911fa2672455b7c76cab9733fcaafc83addea3d1d3399a22"} Dec 06 04:13:39 crc kubenswrapper[4802]: I1206 04:13:39.559026 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7hgqr" Dec 06 04:13:39 crc kubenswrapper[4802]: I1206 04:13:39.677432 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0de67972-73b4-4f71-aca7-d77ae62d73b3-inventory\") pod \"0de67972-73b4-4f71-aca7-d77ae62d73b3\" (UID: \"0de67972-73b4-4f71-aca7-d77ae62d73b3\") " Dec 06 04:13:39 crc kubenswrapper[4802]: I1206 04:13:39.677596 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zmzzd\" (UniqueName: \"kubernetes.io/projected/0de67972-73b4-4f71-aca7-d77ae62d73b3-kube-api-access-zmzzd\") pod \"0de67972-73b4-4f71-aca7-d77ae62d73b3\" (UID: \"0de67972-73b4-4f71-aca7-d77ae62d73b3\") " Dec 06 04:13:39 crc kubenswrapper[4802]: I1206 04:13:39.678302 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0de67972-73b4-4f71-aca7-d77ae62d73b3-ssh-key\") pod \"0de67972-73b4-4f71-aca7-d77ae62d73b3\" (UID: \"0de67972-73b4-4f71-aca7-d77ae62d73b3\") " Dec 06 04:13:39 crc kubenswrapper[4802]: I1206 04:13:39.684307 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0de67972-73b4-4f71-aca7-d77ae62d73b3-kube-api-access-zmzzd" (OuterVolumeSpecName: "kube-api-access-zmzzd") pod "0de67972-73b4-4f71-aca7-d77ae62d73b3" (UID: "0de67972-73b4-4f71-aca7-d77ae62d73b3"). InnerVolumeSpecName "kube-api-access-zmzzd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:13:39 crc kubenswrapper[4802]: I1206 04:13:39.707810 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0de67972-73b4-4f71-aca7-d77ae62d73b3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0de67972-73b4-4f71-aca7-d77ae62d73b3" (UID: "0de67972-73b4-4f71-aca7-d77ae62d73b3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:13:39 crc kubenswrapper[4802]: I1206 04:13:39.730074 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0de67972-73b4-4f71-aca7-d77ae62d73b3-inventory" (OuterVolumeSpecName: "inventory") pod "0de67972-73b4-4f71-aca7-d77ae62d73b3" (UID: "0de67972-73b4-4f71-aca7-d77ae62d73b3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:13:39 crc kubenswrapper[4802]: I1206 04:13:39.780766 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zmzzd\" (UniqueName: \"kubernetes.io/projected/0de67972-73b4-4f71-aca7-d77ae62d73b3-kube-api-access-zmzzd\") on node \"crc\" DevicePath \"\"" Dec 06 04:13:39 crc kubenswrapper[4802]: I1206 04:13:39.780810 4802 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0de67972-73b4-4f71-aca7-d77ae62d73b3-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 04:13:39 crc kubenswrapper[4802]: I1206 04:13:39.780820 4802 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0de67972-73b4-4f71-aca7-d77ae62d73b3-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.154920 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7hgqr" event={"ID":"0de67972-73b4-4f71-aca7-d77ae62d73b3","Type":"ContainerDied","Data":"42e3f071a4995f47ffca00fc1f93749f284ee31333aeac02ba8939bfbfdd6490"} Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.155273 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="42e3f071a4995f47ffca00fc1f93749f284ee31333aeac02ba8939bfbfdd6490" Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.154958 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7hgqr" Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.240914 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g"] Dec 06 04:13:40 crc kubenswrapper[4802]: E1206 04:13:40.241281 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0de67972-73b4-4f71-aca7-d77ae62d73b3" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.241298 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="0de67972-73b4-4f71-aca7-d77ae62d73b3" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.241492 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="0de67972-73b4-4f71-aca7-d77ae62d73b3" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.242103 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g" Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.245341 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.245455 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.245797 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-rgvjc" Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.245900 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.245949 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.246327 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.246562 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.247113 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.259316 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g"] Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.397022 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79de9688-9efb-44b4-9518-259e1f58b5b4-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g\" (UID: \"79de9688-9efb-44b4-9518-259e1f58b5b4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g" Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.397108 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/79de9688-9efb-44b4-9518-259e1f58b5b4-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g\" (UID: \"79de9688-9efb-44b4-9518-259e1f58b5b4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g" Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.397156 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/79de9688-9efb-44b4-9518-259e1f58b5b4-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g\" (UID: \"79de9688-9efb-44b4-9518-259e1f58b5b4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g" Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.397197 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79de9688-9efb-44b4-9518-259e1f58b5b4-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g\" (UID: \"79de9688-9efb-44b4-9518-259e1f58b5b4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g" Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.397312 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/79de9688-9efb-44b4-9518-259e1f58b5b4-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g\" (UID: \"79de9688-9efb-44b4-9518-259e1f58b5b4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g" Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.397335 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/79de9688-9efb-44b4-9518-259e1f58b5b4-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g\" (UID: \"79de9688-9efb-44b4-9518-259e1f58b5b4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g" Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.397360 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/79de9688-9efb-44b4-9518-259e1f58b5b4-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g\" (UID: \"79de9688-9efb-44b4-9518-259e1f58b5b4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g" Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.397396 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79de9688-9efb-44b4-9518-259e1f58b5b4-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g\" (UID: \"79de9688-9efb-44b4-9518-259e1f58b5b4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g" Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.397426 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79de9688-9efb-44b4-9518-259e1f58b5b4-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g\" (UID: \"79de9688-9efb-44b4-9518-259e1f58b5b4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g" Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.397446 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/79de9688-9efb-44b4-9518-259e1f58b5b4-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g\" (UID: \"79de9688-9efb-44b4-9518-259e1f58b5b4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g" Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.397470 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79de9688-9efb-44b4-9518-259e1f58b5b4-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g\" (UID: \"79de9688-9efb-44b4-9518-259e1f58b5b4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g" Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.397489 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xn4nq\" (UniqueName: \"kubernetes.io/projected/79de9688-9efb-44b4-9518-259e1f58b5b4-kube-api-access-xn4nq\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g\" (UID: \"79de9688-9efb-44b4-9518-259e1f58b5b4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g" Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.397530 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79de9688-9efb-44b4-9518-259e1f58b5b4-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g\" (UID: \"79de9688-9efb-44b4-9518-259e1f58b5b4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g" Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.397736 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79de9688-9efb-44b4-9518-259e1f58b5b4-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g\" (UID: \"79de9688-9efb-44b4-9518-259e1f58b5b4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g" Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.499446 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79de9688-9efb-44b4-9518-259e1f58b5b4-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g\" (UID: \"79de9688-9efb-44b4-9518-259e1f58b5b4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g" Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.499731 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79de9688-9efb-44b4-9518-259e1f58b5b4-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g\" (UID: \"79de9688-9efb-44b4-9518-259e1f58b5b4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g" Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.500044 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/79de9688-9efb-44b4-9518-259e1f58b5b4-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g\" (UID: \"79de9688-9efb-44b4-9518-259e1f58b5b4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g" Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.500101 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/79de9688-9efb-44b4-9518-259e1f58b5b4-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g\" (UID: \"79de9688-9efb-44b4-9518-259e1f58b5b4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g" Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.500156 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79de9688-9efb-44b4-9518-259e1f58b5b4-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g\" (UID: \"79de9688-9efb-44b4-9518-259e1f58b5b4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g" Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.500199 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/79de9688-9efb-44b4-9518-259e1f58b5b4-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g\" (UID: \"79de9688-9efb-44b4-9518-259e1f58b5b4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g" Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.502005 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/79de9688-9efb-44b4-9518-259e1f58b5b4-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g\" (UID: \"79de9688-9efb-44b4-9518-259e1f58b5b4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g" Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.502081 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/79de9688-9efb-44b4-9518-259e1f58b5b4-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g\" (UID: \"79de9688-9efb-44b4-9518-259e1f58b5b4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g" Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.502460 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79de9688-9efb-44b4-9518-259e1f58b5b4-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g\" (UID: \"79de9688-9efb-44b4-9518-259e1f58b5b4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g" Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.502558 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79de9688-9efb-44b4-9518-259e1f58b5b4-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g\" (UID: \"79de9688-9efb-44b4-9518-259e1f58b5b4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g" Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.502608 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/79de9688-9efb-44b4-9518-259e1f58b5b4-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g\" (UID: \"79de9688-9efb-44b4-9518-259e1f58b5b4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g" Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.502709 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79de9688-9efb-44b4-9518-259e1f58b5b4-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g\" (UID: \"79de9688-9efb-44b4-9518-259e1f58b5b4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g" Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.502771 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xn4nq\" (UniqueName: \"kubernetes.io/projected/79de9688-9efb-44b4-9518-259e1f58b5b4-kube-api-access-xn4nq\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g\" (UID: \"79de9688-9efb-44b4-9518-259e1f58b5b4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g" Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.502840 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79de9688-9efb-44b4-9518-259e1f58b5b4-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g\" (UID: \"79de9688-9efb-44b4-9518-259e1f58b5b4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g" Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.504305 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/79de9688-9efb-44b4-9518-259e1f58b5b4-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g\" (UID: \"79de9688-9efb-44b4-9518-259e1f58b5b4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g" Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.504342 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79de9688-9efb-44b4-9518-259e1f58b5b4-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g\" (UID: \"79de9688-9efb-44b4-9518-259e1f58b5b4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g" Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.504722 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/79de9688-9efb-44b4-9518-259e1f58b5b4-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g\" (UID: \"79de9688-9efb-44b4-9518-259e1f58b5b4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g" Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.506439 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/79de9688-9efb-44b4-9518-259e1f58b5b4-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g\" (UID: \"79de9688-9efb-44b4-9518-259e1f58b5b4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g" Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.506788 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79de9688-9efb-44b4-9518-259e1f58b5b4-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g\" (UID: \"79de9688-9efb-44b4-9518-259e1f58b5b4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g" Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.507446 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/79de9688-9efb-44b4-9518-259e1f58b5b4-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g\" (UID: \"79de9688-9efb-44b4-9518-259e1f58b5b4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g" Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.507906 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79de9688-9efb-44b4-9518-259e1f58b5b4-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g\" (UID: \"79de9688-9efb-44b4-9518-259e1f58b5b4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g" Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.508690 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/79de9688-9efb-44b4-9518-259e1f58b5b4-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g\" (UID: \"79de9688-9efb-44b4-9518-259e1f58b5b4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g" Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.509787 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/79de9688-9efb-44b4-9518-259e1f58b5b4-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g\" (UID: \"79de9688-9efb-44b4-9518-259e1f58b5b4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g" Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.510310 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79de9688-9efb-44b4-9518-259e1f58b5b4-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g\" (UID: \"79de9688-9efb-44b4-9518-259e1f58b5b4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g" Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.510554 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79de9688-9efb-44b4-9518-259e1f58b5b4-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g\" (UID: \"79de9688-9efb-44b4-9518-259e1f58b5b4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g" Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.513407 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79de9688-9efb-44b4-9518-259e1f58b5b4-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g\" (UID: \"79de9688-9efb-44b4-9518-259e1f58b5b4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g" Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.519648 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79de9688-9efb-44b4-9518-259e1f58b5b4-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g\" (UID: \"79de9688-9efb-44b4-9518-259e1f58b5b4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g" Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.524986 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xn4nq\" (UniqueName: \"kubernetes.io/projected/79de9688-9efb-44b4-9518-259e1f58b5b4-kube-api-access-xn4nq\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g\" (UID: \"79de9688-9efb-44b4-9518-259e1f58b5b4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g" Dec 06 04:13:40 crc kubenswrapper[4802]: I1206 04:13:40.602212 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g" Dec 06 04:13:41 crc kubenswrapper[4802]: I1206 04:13:41.177136 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g"] Dec 06 04:13:42 crc kubenswrapper[4802]: I1206 04:13:42.180352 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g" event={"ID":"79de9688-9efb-44b4-9518-259e1f58b5b4","Type":"ContainerStarted","Data":"c105f6f9d4b7fe2265b55c0ceabdc7e6c7bc54e61d014eaf6d3c7412bdad8711"} Dec 06 04:13:42 crc kubenswrapper[4802]: I1206 04:13:42.180400 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g" event={"ID":"79de9688-9efb-44b4-9518-259e1f58b5b4","Type":"ContainerStarted","Data":"ffc5c2ff5714f29b13766fb87da3599afa7f467436c32a0da772108b2163093d"} Dec 06 04:13:42 crc kubenswrapper[4802]: I1206 04:13:42.206317 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g" podStartSLOduration=1.806192454 podStartE2EDuration="2.206300512s" podCreationTimestamp="2025-12-06 04:13:40 +0000 UTC" firstStartedPulling="2025-12-06 04:13:41.188146581 +0000 UTC m=+2014.060055733" lastFinishedPulling="2025-12-06 04:13:41.588254639 +0000 UTC m=+2014.460163791" observedRunningTime="2025-12-06 04:13:42.205061879 +0000 UTC m=+2015.076971051" watchObservedRunningTime="2025-12-06 04:13:42.206300512 +0000 UTC m=+2015.078209664" Dec 06 04:13:43 crc kubenswrapper[4802]: I1206 04:13:43.283658 4802 patch_prober.go:28] interesting pod/machine-config-daemon-zpxxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:13:43 crc kubenswrapper[4802]: I1206 04:13:43.284040 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:14:13 crc kubenswrapper[4802]: I1206 04:14:13.283674 4802 patch_prober.go:28] interesting pod/machine-config-daemon-zpxxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:14:13 crc kubenswrapper[4802]: I1206 04:14:13.284691 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:14:13 crc kubenswrapper[4802]: I1206 04:14:13.284781 4802 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" Dec 06 04:14:13 crc kubenswrapper[4802]: I1206 04:14:13.285994 4802 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"baa707a57be38104aa835e7ecacc0175c718b272c82894ed9c2693dc0041453b"} pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 04:14:13 crc kubenswrapper[4802]: I1206 04:14:13.286143 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" containerID="cri-o://baa707a57be38104aa835e7ecacc0175c718b272c82894ed9c2693dc0041453b" gracePeriod=600 Dec 06 04:14:14 crc kubenswrapper[4802]: I1206 04:14:14.355831 4802 generic.go:334] "Generic (PLEG): container finished" podID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerID="baa707a57be38104aa835e7ecacc0175c718b272c82894ed9c2693dc0041453b" exitCode=0 Dec 06 04:14:14 crc kubenswrapper[4802]: I1206 04:14:14.355930 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" event={"ID":"6df38316-e0d3-4018-8d27-3620eba3a68d","Type":"ContainerDied","Data":"baa707a57be38104aa835e7ecacc0175c718b272c82894ed9c2693dc0041453b"} Dec 06 04:14:14 crc kubenswrapper[4802]: I1206 04:14:14.356378 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" event={"ID":"6df38316-e0d3-4018-8d27-3620eba3a68d","Type":"ContainerStarted","Data":"f8a17abbaa167592a628d9fdb3771d61462952618f6382d447c431c49f7c5690"} Dec 06 04:14:14 crc kubenswrapper[4802]: I1206 04:14:14.356400 4802 scope.go:117] "RemoveContainer" containerID="3711adecb81191bc35ac469efa73faeb2346d30f4d5a4a148de8ac88071cdb00" Dec 06 04:14:21 crc kubenswrapper[4802]: I1206 04:14:21.439278 4802 generic.go:334] "Generic (PLEG): container finished" podID="79de9688-9efb-44b4-9518-259e1f58b5b4" containerID="c105f6f9d4b7fe2265b55c0ceabdc7e6c7bc54e61d014eaf6d3c7412bdad8711" exitCode=0 Dec 06 04:14:21 crc kubenswrapper[4802]: I1206 04:14:21.439864 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g" event={"ID":"79de9688-9efb-44b4-9518-259e1f58b5b4","Type":"ContainerDied","Data":"c105f6f9d4b7fe2265b55c0ceabdc7e6c7bc54e61d014eaf6d3c7412bdad8711"} Dec 06 04:14:22 crc kubenswrapper[4802]: I1206 04:14:22.835866 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g" Dec 06 04:14:22 crc kubenswrapper[4802]: I1206 04:14:22.853076 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/79de9688-9efb-44b4-9518-259e1f58b5b4-ssh-key\") pod \"79de9688-9efb-44b4-9518-259e1f58b5b4\" (UID: \"79de9688-9efb-44b4-9518-259e1f58b5b4\") " Dec 06 04:14:22 crc kubenswrapper[4802]: I1206 04:14:22.853192 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/79de9688-9efb-44b4-9518-259e1f58b5b4-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"79de9688-9efb-44b4-9518-259e1f58b5b4\" (UID: \"79de9688-9efb-44b4-9518-259e1f58b5b4\") " Dec 06 04:14:22 crc kubenswrapper[4802]: I1206 04:14:22.853333 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/79de9688-9efb-44b4-9518-259e1f58b5b4-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"79de9688-9efb-44b4-9518-259e1f58b5b4\" (UID: \"79de9688-9efb-44b4-9518-259e1f58b5b4\") " Dec 06 04:14:22 crc kubenswrapper[4802]: I1206 04:14:22.853475 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79de9688-9efb-44b4-9518-259e1f58b5b4-bootstrap-combined-ca-bundle\") pod \"79de9688-9efb-44b4-9518-259e1f58b5b4\" (UID: \"79de9688-9efb-44b4-9518-259e1f58b5b4\") " Dec 06 04:14:22 crc kubenswrapper[4802]: I1206 04:14:22.853500 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79de9688-9efb-44b4-9518-259e1f58b5b4-nova-combined-ca-bundle\") pod \"79de9688-9efb-44b4-9518-259e1f58b5b4\" (UID: \"79de9688-9efb-44b4-9518-259e1f58b5b4\") " Dec 06 04:14:22 crc kubenswrapper[4802]: I1206 04:14:22.853575 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79de9688-9efb-44b4-9518-259e1f58b5b4-ovn-combined-ca-bundle\") pod \"79de9688-9efb-44b4-9518-259e1f58b5b4\" (UID: \"79de9688-9efb-44b4-9518-259e1f58b5b4\") " Dec 06 04:14:22 crc kubenswrapper[4802]: I1206 04:14:22.853605 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xn4nq\" (UniqueName: \"kubernetes.io/projected/79de9688-9efb-44b4-9518-259e1f58b5b4-kube-api-access-xn4nq\") pod \"79de9688-9efb-44b4-9518-259e1f58b5b4\" (UID: \"79de9688-9efb-44b4-9518-259e1f58b5b4\") " Dec 06 04:14:22 crc kubenswrapper[4802]: I1206 04:14:22.853654 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79de9688-9efb-44b4-9518-259e1f58b5b4-libvirt-combined-ca-bundle\") pod \"79de9688-9efb-44b4-9518-259e1f58b5b4\" (UID: \"79de9688-9efb-44b4-9518-259e1f58b5b4\") " Dec 06 04:14:22 crc kubenswrapper[4802]: I1206 04:14:22.853702 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79de9688-9efb-44b4-9518-259e1f58b5b4-repo-setup-combined-ca-bundle\") pod \"79de9688-9efb-44b4-9518-259e1f58b5b4\" (UID: \"79de9688-9efb-44b4-9518-259e1f58b5b4\") " Dec 06 04:14:22 crc kubenswrapper[4802]: I1206 04:14:22.853781 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/79de9688-9efb-44b4-9518-259e1f58b5b4-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"79de9688-9efb-44b4-9518-259e1f58b5b4\" (UID: \"79de9688-9efb-44b4-9518-259e1f58b5b4\") " Dec 06 04:14:22 crc kubenswrapper[4802]: I1206 04:14:22.853903 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79de9688-9efb-44b4-9518-259e1f58b5b4-telemetry-combined-ca-bundle\") pod \"79de9688-9efb-44b4-9518-259e1f58b5b4\" (UID: \"79de9688-9efb-44b4-9518-259e1f58b5b4\") " Dec 06 04:14:22 crc kubenswrapper[4802]: I1206 04:14:22.853959 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79de9688-9efb-44b4-9518-259e1f58b5b4-neutron-metadata-combined-ca-bundle\") pod \"79de9688-9efb-44b4-9518-259e1f58b5b4\" (UID: \"79de9688-9efb-44b4-9518-259e1f58b5b4\") " Dec 06 04:14:22 crc kubenswrapper[4802]: I1206 04:14:22.853989 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/79de9688-9efb-44b4-9518-259e1f58b5b4-openstack-edpm-ipam-ovn-default-certs-0\") pod \"79de9688-9efb-44b4-9518-259e1f58b5b4\" (UID: \"79de9688-9efb-44b4-9518-259e1f58b5b4\") " Dec 06 04:14:22 crc kubenswrapper[4802]: I1206 04:14:22.854011 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/79de9688-9efb-44b4-9518-259e1f58b5b4-inventory\") pod \"79de9688-9efb-44b4-9518-259e1f58b5b4\" (UID: \"79de9688-9efb-44b4-9518-259e1f58b5b4\") " Dec 06 04:14:22 crc kubenswrapper[4802]: I1206 04:14:22.864359 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79de9688-9efb-44b4-9518-259e1f58b5b4-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "79de9688-9efb-44b4-9518-259e1f58b5b4" (UID: "79de9688-9efb-44b4-9518-259e1f58b5b4"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:14:22 crc kubenswrapper[4802]: I1206 04:14:22.866719 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79de9688-9efb-44b4-9518-259e1f58b5b4-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "79de9688-9efb-44b4-9518-259e1f58b5b4" (UID: "79de9688-9efb-44b4-9518-259e1f58b5b4"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:14:22 crc kubenswrapper[4802]: I1206 04:14:22.873888 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79de9688-9efb-44b4-9518-259e1f58b5b4-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "79de9688-9efb-44b4-9518-259e1f58b5b4" (UID: "79de9688-9efb-44b4-9518-259e1f58b5b4"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:14:22 crc kubenswrapper[4802]: I1206 04:14:22.874009 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79de9688-9efb-44b4-9518-259e1f58b5b4-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "79de9688-9efb-44b4-9518-259e1f58b5b4" (UID: "79de9688-9efb-44b4-9518-259e1f58b5b4"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:14:22 crc kubenswrapper[4802]: I1206 04:14:22.874036 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79de9688-9efb-44b4-9518-259e1f58b5b4-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "79de9688-9efb-44b4-9518-259e1f58b5b4" (UID: "79de9688-9efb-44b4-9518-259e1f58b5b4"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:14:22 crc kubenswrapper[4802]: I1206 04:14:22.874341 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79de9688-9efb-44b4-9518-259e1f58b5b4-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "79de9688-9efb-44b4-9518-259e1f58b5b4" (UID: "79de9688-9efb-44b4-9518-259e1f58b5b4"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:14:22 crc kubenswrapper[4802]: I1206 04:14:22.875772 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79de9688-9efb-44b4-9518-259e1f58b5b4-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "79de9688-9efb-44b4-9518-259e1f58b5b4" (UID: "79de9688-9efb-44b4-9518-259e1f58b5b4"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:14:22 crc kubenswrapper[4802]: I1206 04:14:22.876560 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79de9688-9efb-44b4-9518-259e1f58b5b4-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "79de9688-9efb-44b4-9518-259e1f58b5b4" (UID: "79de9688-9efb-44b4-9518-259e1f58b5b4"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:14:22 crc kubenswrapper[4802]: I1206 04:14:22.876866 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79de9688-9efb-44b4-9518-259e1f58b5b4-kube-api-access-xn4nq" (OuterVolumeSpecName: "kube-api-access-xn4nq") pod "79de9688-9efb-44b4-9518-259e1f58b5b4" (UID: "79de9688-9efb-44b4-9518-259e1f58b5b4"). InnerVolumeSpecName "kube-api-access-xn4nq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:14:22 crc kubenswrapper[4802]: I1206 04:14:22.878417 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79de9688-9efb-44b4-9518-259e1f58b5b4-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "79de9688-9efb-44b4-9518-259e1f58b5b4" (UID: "79de9688-9efb-44b4-9518-259e1f58b5b4"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:14:22 crc kubenswrapper[4802]: I1206 04:14:22.878472 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79de9688-9efb-44b4-9518-259e1f58b5b4-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "79de9688-9efb-44b4-9518-259e1f58b5b4" (UID: "79de9688-9efb-44b4-9518-259e1f58b5b4"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:14:22 crc kubenswrapper[4802]: I1206 04:14:22.878878 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79de9688-9efb-44b4-9518-259e1f58b5b4-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "79de9688-9efb-44b4-9518-259e1f58b5b4" (UID: "79de9688-9efb-44b4-9518-259e1f58b5b4"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:14:22 crc kubenswrapper[4802]: I1206 04:14:22.896801 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79de9688-9efb-44b4-9518-259e1f58b5b4-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "79de9688-9efb-44b4-9518-259e1f58b5b4" (UID: "79de9688-9efb-44b4-9518-259e1f58b5b4"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:14:22 crc kubenswrapper[4802]: I1206 04:14:22.898022 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79de9688-9efb-44b4-9518-259e1f58b5b4-inventory" (OuterVolumeSpecName: "inventory") pod "79de9688-9efb-44b4-9518-259e1f58b5b4" (UID: "79de9688-9efb-44b4-9518-259e1f58b5b4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:14:22 crc kubenswrapper[4802]: I1206 04:14:22.959469 4802 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/79de9688-9efb-44b4-9518-259e1f58b5b4-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 04:14:22 crc kubenswrapper[4802]: I1206 04:14:22.959510 4802 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/79de9688-9efb-44b4-9518-259e1f58b5b4-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 06 04:14:22 crc kubenswrapper[4802]: I1206 04:14:22.959526 4802 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/79de9688-9efb-44b4-9518-259e1f58b5b4-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 06 04:14:22 crc kubenswrapper[4802]: I1206 04:14:22.959543 4802 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79de9688-9efb-44b4-9518-259e1f58b5b4-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:14:22 crc kubenswrapper[4802]: I1206 04:14:22.959556 4802 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79de9688-9efb-44b4-9518-259e1f58b5b4-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:14:22 crc kubenswrapper[4802]: I1206 04:14:22.959568 4802 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79de9688-9efb-44b4-9518-259e1f58b5b4-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:14:22 crc kubenswrapper[4802]: I1206 04:14:22.959581 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xn4nq\" (UniqueName: \"kubernetes.io/projected/79de9688-9efb-44b4-9518-259e1f58b5b4-kube-api-access-xn4nq\") on node \"crc\" DevicePath \"\"" Dec 06 04:14:22 crc kubenswrapper[4802]: I1206 04:14:22.959593 4802 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79de9688-9efb-44b4-9518-259e1f58b5b4-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:14:22 crc kubenswrapper[4802]: I1206 04:14:22.959605 4802 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79de9688-9efb-44b4-9518-259e1f58b5b4-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:14:22 crc kubenswrapper[4802]: I1206 04:14:22.959620 4802 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/79de9688-9efb-44b4-9518-259e1f58b5b4-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 06 04:14:22 crc kubenswrapper[4802]: I1206 04:14:22.959634 4802 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79de9688-9efb-44b4-9518-259e1f58b5b4-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:14:22 crc kubenswrapper[4802]: I1206 04:14:22.959648 4802 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79de9688-9efb-44b4-9518-259e1f58b5b4-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:14:22 crc kubenswrapper[4802]: I1206 04:14:22.959660 4802 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/79de9688-9efb-44b4-9518-259e1f58b5b4-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 06 04:14:22 crc kubenswrapper[4802]: I1206 04:14:22.959674 4802 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/79de9688-9efb-44b4-9518-259e1f58b5b4-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 04:14:23 crc kubenswrapper[4802]: I1206 04:14:23.461560 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g" Dec 06 04:14:23 crc kubenswrapper[4802]: I1206 04:14:23.465619 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g" event={"ID":"79de9688-9efb-44b4-9518-259e1f58b5b4","Type":"ContainerDied","Data":"ffc5c2ff5714f29b13766fb87da3599afa7f467436c32a0da772108b2163093d"} Dec 06 04:14:23 crc kubenswrapper[4802]: I1206 04:14:23.465678 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ffc5c2ff5714f29b13766fb87da3599afa7f467436c32a0da772108b2163093d" Dec 06 04:14:23 crc kubenswrapper[4802]: I1206 04:14:23.616482 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-9hxgw"] Dec 06 04:14:23 crc kubenswrapper[4802]: E1206 04:14:23.617223 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79de9688-9efb-44b4-9518-259e1f58b5b4" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 06 04:14:23 crc kubenswrapper[4802]: I1206 04:14:23.617253 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="79de9688-9efb-44b4-9518-259e1f58b5b4" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 06 04:14:23 crc kubenswrapper[4802]: I1206 04:14:23.617499 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="79de9688-9efb-44b4-9518-259e1f58b5b4" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 06 04:14:23 crc kubenswrapper[4802]: I1206 04:14:23.618355 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9hxgw" Dec 06 04:14:23 crc kubenswrapper[4802]: I1206 04:14:23.620976 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 04:14:23 crc kubenswrapper[4802]: I1206 04:14:23.621253 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 04:14:23 crc kubenswrapper[4802]: I1206 04:14:23.621666 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-rgvjc" Dec 06 04:14:23 crc kubenswrapper[4802]: I1206 04:14:23.621891 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Dec 06 04:14:23 crc kubenswrapper[4802]: I1206 04:14:23.623305 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 04:14:23 crc kubenswrapper[4802]: I1206 04:14:23.632714 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-9hxgw"] Dec 06 04:14:23 crc kubenswrapper[4802]: I1206 04:14:23.679849 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8tklv\" (UniqueName: \"kubernetes.io/projected/e2001db0-90f1-40a7-ba6e-982eabebc117-kube-api-access-8tklv\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9hxgw\" (UID: \"e2001db0-90f1-40a7-ba6e-982eabebc117\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9hxgw" Dec 06 04:14:23 crc kubenswrapper[4802]: I1206 04:14:23.680007 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e2001db0-90f1-40a7-ba6e-982eabebc117-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9hxgw\" (UID: \"e2001db0-90f1-40a7-ba6e-982eabebc117\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9hxgw" Dec 06 04:14:23 crc kubenswrapper[4802]: I1206 04:14:23.680080 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/e2001db0-90f1-40a7-ba6e-982eabebc117-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9hxgw\" (UID: \"e2001db0-90f1-40a7-ba6e-982eabebc117\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9hxgw" Dec 06 04:14:23 crc kubenswrapper[4802]: I1206 04:14:23.680102 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e2001db0-90f1-40a7-ba6e-982eabebc117-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9hxgw\" (UID: \"e2001db0-90f1-40a7-ba6e-982eabebc117\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9hxgw" Dec 06 04:14:23 crc kubenswrapper[4802]: I1206 04:14:23.680135 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2001db0-90f1-40a7-ba6e-982eabebc117-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9hxgw\" (UID: \"e2001db0-90f1-40a7-ba6e-982eabebc117\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9hxgw" Dec 06 04:14:23 crc kubenswrapper[4802]: I1206 04:14:23.781495 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8tklv\" (UniqueName: \"kubernetes.io/projected/e2001db0-90f1-40a7-ba6e-982eabebc117-kube-api-access-8tklv\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9hxgw\" (UID: \"e2001db0-90f1-40a7-ba6e-982eabebc117\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9hxgw" Dec 06 04:14:23 crc kubenswrapper[4802]: I1206 04:14:23.781629 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e2001db0-90f1-40a7-ba6e-982eabebc117-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9hxgw\" (UID: \"e2001db0-90f1-40a7-ba6e-982eabebc117\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9hxgw" Dec 06 04:14:23 crc kubenswrapper[4802]: I1206 04:14:23.781673 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/e2001db0-90f1-40a7-ba6e-982eabebc117-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9hxgw\" (UID: \"e2001db0-90f1-40a7-ba6e-982eabebc117\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9hxgw" Dec 06 04:14:23 crc kubenswrapper[4802]: I1206 04:14:23.781698 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e2001db0-90f1-40a7-ba6e-982eabebc117-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9hxgw\" (UID: \"e2001db0-90f1-40a7-ba6e-982eabebc117\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9hxgw" Dec 06 04:14:23 crc kubenswrapper[4802]: I1206 04:14:23.781719 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2001db0-90f1-40a7-ba6e-982eabebc117-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9hxgw\" (UID: \"e2001db0-90f1-40a7-ba6e-982eabebc117\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9hxgw" Dec 06 04:14:23 crc kubenswrapper[4802]: I1206 04:14:23.783491 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/e2001db0-90f1-40a7-ba6e-982eabebc117-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9hxgw\" (UID: \"e2001db0-90f1-40a7-ba6e-982eabebc117\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9hxgw" Dec 06 04:14:23 crc kubenswrapper[4802]: I1206 04:14:23.789532 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e2001db0-90f1-40a7-ba6e-982eabebc117-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9hxgw\" (UID: \"e2001db0-90f1-40a7-ba6e-982eabebc117\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9hxgw" Dec 06 04:14:23 crc kubenswrapper[4802]: I1206 04:14:23.789940 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e2001db0-90f1-40a7-ba6e-982eabebc117-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9hxgw\" (UID: \"e2001db0-90f1-40a7-ba6e-982eabebc117\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9hxgw" Dec 06 04:14:23 crc kubenswrapper[4802]: I1206 04:14:23.790490 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2001db0-90f1-40a7-ba6e-982eabebc117-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9hxgw\" (UID: \"e2001db0-90f1-40a7-ba6e-982eabebc117\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9hxgw" Dec 06 04:14:23 crc kubenswrapper[4802]: I1206 04:14:23.804254 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8tklv\" (UniqueName: \"kubernetes.io/projected/e2001db0-90f1-40a7-ba6e-982eabebc117-kube-api-access-8tklv\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9hxgw\" (UID: \"e2001db0-90f1-40a7-ba6e-982eabebc117\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9hxgw" Dec 06 04:14:23 crc kubenswrapper[4802]: I1206 04:14:23.936527 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9hxgw" Dec 06 04:14:24 crc kubenswrapper[4802]: I1206 04:14:24.428253 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-9hxgw"] Dec 06 04:14:24 crc kubenswrapper[4802]: I1206 04:14:24.479251 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9hxgw" event={"ID":"e2001db0-90f1-40a7-ba6e-982eabebc117","Type":"ContainerStarted","Data":"2da96709c1fa418850efce465292ff8ccc3b2cc56418b6491ca3ccfee1016cc4"} Dec 06 04:14:25 crc kubenswrapper[4802]: I1206 04:14:25.489299 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9hxgw" event={"ID":"e2001db0-90f1-40a7-ba6e-982eabebc117","Type":"ContainerStarted","Data":"7ee734134363a2e85ca9795f1c3e9d7e7ff551bf967dfe6b1c9dfc0130c0cbba"} Dec 06 04:14:25 crc kubenswrapper[4802]: I1206 04:14:25.516570 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9hxgw" podStartSLOduration=2.088831824 podStartE2EDuration="2.516549973s" podCreationTimestamp="2025-12-06 04:14:23 +0000 UTC" firstStartedPulling="2025-12-06 04:14:24.42741015 +0000 UTC m=+2057.299319302" lastFinishedPulling="2025-12-06 04:14:24.855128299 +0000 UTC m=+2057.727037451" observedRunningTime="2025-12-06 04:14:25.512962446 +0000 UTC m=+2058.384871648" watchObservedRunningTime="2025-12-06 04:14:25.516549973 +0000 UTC m=+2058.388459135" Dec 06 04:15:00 crc kubenswrapper[4802]: I1206 04:15:00.146229 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416575-fwqzl"] Dec 06 04:15:00 crc kubenswrapper[4802]: I1206 04:15:00.148413 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416575-fwqzl" Dec 06 04:15:00 crc kubenswrapper[4802]: I1206 04:15:00.151040 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 04:15:00 crc kubenswrapper[4802]: I1206 04:15:00.151363 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 04:15:00 crc kubenswrapper[4802]: I1206 04:15:00.156365 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416575-fwqzl"] Dec 06 04:15:00 crc kubenswrapper[4802]: I1206 04:15:00.295463 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkxn8\" (UniqueName: \"kubernetes.io/projected/8189b4bc-09c3-4075-b959-bb67bf6ccd34-kube-api-access-nkxn8\") pod \"collect-profiles-29416575-fwqzl\" (UID: \"8189b4bc-09c3-4075-b959-bb67bf6ccd34\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416575-fwqzl" Dec 06 04:15:00 crc kubenswrapper[4802]: I1206 04:15:00.295708 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8189b4bc-09c3-4075-b959-bb67bf6ccd34-secret-volume\") pod \"collect-profiles-29416575-fwqzl\" (UID: \"8189b4bc-09c3-4075-b959-bb67bf6ccd34\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416575-fwqzl" Dec 06 04:15:00 crc kubenswrapper[4802]: I1206 04:15:00.295779 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8189b4bc-09c3-4075-b959-bb67bf6ccd34-config-volume\") pod \"collect-profiles-29416575-fwqzl\" (UID: \"8189b4bc-09c3-4075-b959-bb67bf6ccd34\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416575-fwqzl" Dec 06 04:15:00 crc kubenswrapper[4802]: I1206 04:15:00.397850 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8189b4bc-09c3-4075-b959-bb67bf6ccd34-secret-volume\") pod \"collect-profiles-29416575-fwqzl\" (UID: \"8189b4bc-09c3-4075-b959-bb67bf6ccd34\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416575-fwqzl" Dec 06 04:15:00 crc kubenswrapper[4802]: I1206 04:15:00.397923 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8189b4bc-09c3-4075-b959-bb67bf6ccd34-config-volume\") pod \"collect-profiles-29416575-fwqzl\" (UID: \"8189b4bc-09c3-4075-b959-bb67bf6ccd34\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416575-fwqzl" Dec 06 04:15:00 crc kubenswrapper[4802]: I1206 04:15:00.397993 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkxn8\" (UniqueName: \"kubernetes.io/projected/8189b4bc-09c3-4075-b959-bb67bf6ccd34-kube-api-access-nkxn8\") pod \"collect-profiles-29416575-fwqzl\" (UID: \"8189b4bc-09c3-4075-b959-bb67bf6ccd34\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416575-fwqzl" Dec 06 04:15:00 crc kubenswrapper[4802]: I1206 04:15:00.398961 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8189b4bc-09c3-4075-b959-bb67bf6ccd34-config-volume\") pod \"collect-profiles-29416575-fwqzl\" (UID: \"8189b4bc-09c3-4075-b959-bb67bf6ccd34\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416575-fwqzl" Dec 06 04:15:00 crc kubenswrapper[4802]: I1206 04:15:00.412645 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8189b4bc-09c3-4075-b959-bb67bf6ccd34-secret-volume\") pod \"collect-profiles-29416575-fwqzl\" (UID: \"8189b4bc-09c3-4075-b959-bb67bf6ccd34\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416575-fwqzl" Dec 06 04:15:00 crc kubenswrapper[4802]: I1206 04:15:00.421784 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkxn8\" (UniqueName: \"kubernetes.io/projected/8189b4bc-09c3-4075-b959-bb67bf6ccd34-kube-api-access-nkxn8\") pod \"collect-profiles-29416575-fwqzl\" (UID: \"8189b4bc-09c3-4075-b959-bb67bf6ccd34\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416575-fwqzl" Dec 06 04:15:00 crc kubenswrapper[4802]: I1206 04:15:00.479430 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416575-fwqzl" Dec 06 04:15:00 crc kubenswrapper[4802]: I1206 04:15:00.924501 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416575-fwqzl"] Dec 06 04:15:01 crc kubenswrapper[4802]: I1206 04:15:01.828483 4802 generic.go:334] "Generic (PLEG): container finished" podID="8189b4bc-09c3-4075-b959-bb67bf6ccd34" containerID="be56c3fe487fc382c34351345754aedaa82b6dbfce9e8c1b32005d67202a5a56" exitCode=0 Dec 06 04:15:01 crc kubenswrapper[4802]: I1206 04:15:01.828533 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416575-fwqzl" event={"ID":"8189b4bc-09c3-4075-b959-bb67bf6ccd34","Type":"ContainerDied","Data":"be56c3fe487fc382c34351345754aedaa82b6dbfce9e8c1b32005d67202a5a56"} Dec 06 04:15:01 crc kubenswrapper[4802]: I1206 04:15:01.828913 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416575-fwqzl" event={"ID":"8189b4bc-09c3-4075-b959-bb67bf6ccd34","Type":"ContainerStarted","Data":"45c5b20716fa07ee332b93c3c1e21511dfe0b8dbea68e4bea7d032580985353e"} Dec 06 04:15:03 crc kubenswrapper[4802]: I1206 04:15:03.183025 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416575-fwqzl" Dec 06 04:15:03 crc kubenswrapper[4802]: I1206 04:15:03.350749 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8189b4bc-09c3-4075-b959-bb67bf6ccd34-secret-volume\") pod \"8189b4bc-09c3-4075-b959-bb67bf6ccd34\" (UID: \"8189b4bc-09c3-4075-b959-bb67bf6ccd34\") " Dec 06 04:15:03 crc kubenswrapper[4802]: I1206 04:15:03.350906 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nkxn8\" (UniqueName: \"kubernetes.io/projected/8189b4bc-09c3-4075-b959-bb67bf6ccd34-kube-api-access-nkxn8\") pod \"8189b4bc-09c3-4075-b959-bb67bf6ccd34\" (UID: \"8189b4bc-09c3-4075-b959-bb67bf6ccd34\") " Dec 06 04:15:03 crc kubenswrapper[4802]: I1206 04:15:03.350935 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8189b4bc-09c3-4075-b959-bb67bf6ccd34-config-volume\") pod \"8189b4bc-09c3-4075-b959-bb67bf6ccd34\" (UID: \"8189b4bc-09c3-4075-b959-bb67bf6ccd34\") " Dec 06 04:15:03 crc kubenswrapper[4802]: I1206 04:15:03.351699 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8189b4bc-09c3-4075-b959-bb67bf6ccd34-config-volume" (OuterVolumeSpecName: "config-volume") pod "8189b4bc-09c3-4075-b959-bb67bf6ccd34" (UID: "8189b4bc-09c3-4075-b959-bb67bf6ccd34"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:15:03 crc kubenswrapper[4802]: I1206 04:15:03.357508 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8189b4bc-09c3-4075-b959-bb67bf6ccd34-kube-api-access-nkxn8" (OuterVolumeSpecName: "kube-api-access-nkxn8") pod "8189b4bc-09c3-4075-b959-bb67bf6ccd34" (UID: "8189b4bc-09c3-4075-b959-bb67bf6ccd34"). InnerVolumeSpecName "kube-api-access-nkxn8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:15:03 crc kubenswrapper[4802]: I1206 04:15:03.357829 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8189b4bc-09c3-4075-b959-bb67bf6ccd34-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "8189b4bc-09c3-4075-b959-bb67bf6ccd34" (UID: "8189b4bc-09c3-4075-b959-bb67bf6ccd34"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:15:03 crc kubenswrapper[4802]: I1206 04:15:03.453707 4802 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8189b4bc-09c3-4075-b959-bb67bf6ccd34-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 04:15:03 crc kubenswrapper[4802]: I1206 04:15:03.453801 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nkxn8\" (UniqueName: \"kubernetes.io/projected/8189b4bc-09c3-4075-b959-bb67bf6ccd34-kube-api-access-nkxn8\") on node \"crc\" DevicePath \"\"" Dec 06 04:15:03 crc kubenswrapper[4802]: I1206 04:15:03.453823 4802 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8189b4bc-09c3-4075-b959-bb67bf6ccd34-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 04:15:03 crc kubenswrapper[4802]: I1206 04:15:03.850329 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416575-fwqzl" event={"ID":"8189b4bc-09c3-4075-b959-bb67bf6ccd34","Type":"ContainerDied","Data":"45c5b20716fa07ee332b93c3c1e21511dfe0b8dbea68e4bea7d032580985353e"} Dec 06 04:15:03 crc kubenswrapper[4802]: I1206 04:15:03.850378 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="45c5b20716fa07ee332b93c3c1e21511dfe0b8dbea68e4bea7d032580985353e" Dec 06 04:15:03 crc kubenswrapper[4802]: I1206 04:15:03.850384 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416575-fwqzl" Dec 06 04:15:04 crc kubenswrapper[4802]: I1206 04:15:04.263372 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416530-88l5w"] Dec 06 04:15:04 crc kubenswrapper[4802]: I1206 04:15:04.279534 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416530-88l5w"] Dec 06 04:15:05 crc kubenswrapper[4802]: I1206 04:15:05.463333 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19ade530-d0de-44d1-be83-96136753a35e" path="/var/lib/kubelet/pods/19ade530-d0de-44d1-be83-96136753a35e/volumes" Dec 06 04:15:22 crc kubenswrapper[4802]: I1206 04:15:22.300425 4802 scope.go:117] "RemoveContainer" containerID="017399694bb75b5df9ba1d00a8d0ea5dd6ffc293c970cfffc5e1bdf4bc7e000a" Dec 06 04:15:30 crc kubenswrapper[4802]: I1206 04:15:30.090296 4802 generic.go:334] "Generic (PLEG): container finished" podID="e2001db0-90f1-40a7-ba6e-982eabebc117" containerID="7ee734134363a2e85ca9795f1c3e9d7e7ff551bf967dfe6b1c9dfc0130c0cbba" exitCode=0 Dec 06 04:15:30 crc kubenswrapper[4802]: I1206 04:15:30.090375 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9hxgw" event={"ID":"e2001db0-90f1-40a7-ba6e-982eabebc117","Type":"ContainerDied","Data":"7ee734134363a2e85ca9795f1c3e9d7e7ff551bf967dfe6b1c9dfc0130c0cbba"} Dec 06 04:15:31 crc kubenswrapper[4802]: I1206 04:15:31.525843 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9hxgw" Dec 06 04:15:31 crc kubenswrapper[4802]: I1206 04:15:31.624896 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tklv\" (UniqueName: \"kubernetes.io/projected/e2001db0-90f1-40a7-ba6e-982eabebc117-kube-api-access-8tklv\") pod \"e2001db0-90f1-40a7-ba6e-982eabebc117\" (UID: \"e2001db0-90f1-40a7-ba6e-982eabebc117\") " Dec 06 04:15:31 crc kubenswrapper[4802]: I1206 04:15:31.625094 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e2001db0-90f1-40a7-ba6e-982eabebc117-inventory\") pod \"e2001db0-90f1-40a7-ba6e-982eabebc117\" (UID: \"e2001db0-90f1-40a7-ba6e-982eabebc117\") " Dec 06 04:15:31 crc kubenswrapper[4802]: I1206 04:15:31.625261 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2001db0-90f1-40a7-ba6e-982eabebc117-ovn-combined-ca-bundle\") pod \"e2001db0-90f1-40a7-ba6e-982eabebc117\" (UID: \"e2001db0-90f1-40a7-ba6e-982eabebc117\") " Dec 06 04:15:31 crc kubenswrapper[4802]: I1206 04:15:31.625345 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/e2001db0-90f1-40a7-ba6e-982eabebc117-ovncontroller-config-0\") pod \"e2001db0-90f1-40a7-ba6e-982eabebc117\" (UID: \"e2001db0-90f1-40a7-ba6e-982eabebc117\") " Dec 06 04:15:31 crc kubenswrapper[4802]: I1206 04:15:31.625424 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e2001db0-90f1-40a7-ba6e-982eabebc117-ssh-key\") pod \"e2001db0-90f1-40a7-ba6e-982eabebc117\" (UID: \"e2001db0-90f1-40a7-ba6e-982eabebc117\") " Dec 06 04:15:31 crc kubenswrapper[4802]: I1206 04:15:31.630818 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2001db0-90f1-40a7-ba6e-982eabebc117-kube-api-access-8tklv" (OuterVolumeSpecName: "kube-api-access-8tklv") pod "e2001db0-90f1-40a7-ba6e-982eabebc117" (UID: "e2001db0-90f1-40a7-ba6e-982eabebc117"). InnerVolumeSpecName "kube-api-access-8tklv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:15:31 crc kubenswrapper[4802]: I1206 04:15:31.631106 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2001db0-90f1-40a7-ba6e-982eabebc117-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "e2001db0-90f1-40a7-ba6e-982eabebc117" (UID: "e2001db0-90f1-40a7-ba6e-982eabebc117"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:15:31 crc kubenswrapper[4802]: I1206 04:15:31.652955 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2001db0-90f1-40a7-ba6e-982eabebc117-inventory" (OuterVolumeSpecName: "inventory") pod "e2001db0-90f1-40a7-ba6e-982eabebc117" (UID: "e2001db0-90f1-40a7-ba6e-982eabebc117"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:15:31 crc kubenswrapper[4802]: I1206 04:15:31.658061 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2001db0-90f1-40a7-ba6e-982eabebc117-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "e2001db0-90f1-40a7-ba6e-982eabebc117" (UID: "e2001db0-90f1-40a7-ba6e-982eabebc117"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:15:31 crc kubenswrapper[4802]: I1206 04:15:31.660617 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2001db0-90f1-40a7-ba6e-982eabebc117-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e2001db0-90f1-40a7-ba6e-982eabebc117" (UID: "e2001db0-90f1-40a7-ba6e-982eabebc117"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:15:31 crc kubenswrapper[4802]: I1206 04:15:31.729084 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tklv\" (UniqueName: \"kubernetes.io/projected/e2001db0-90f1-40a7-ba6e-982eabebc117-kube-api-access-8tklv\") on node \"crc\" DevicePath \"\"" Dec 06 04:15:31 crc kubenswrapper[4802]: I1206 04:15:31.729118 4802 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e2001db0-90f1-40a7-ba6e-982eabebc117-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 04:15:31 crc kubenswrapper[4802]: I1206 04:15:31.729127 4802 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2001db0-90f1-40a7-ba6e-982eabebc117-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:15:31 crc kubenswrapper[4802]: I1206 04:15:31.729139 4802 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/e2001db0-90f1-40a7-ba6e-982eabebc117-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 06 04:15:31 crc kubenswrapper[4802]: I1206 04:15:31.729146 4802 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e2001db0-90f1-40a7-ba6e-982eabebc117-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 04:15:32 crc kubenswrapper[4802]: I1206 04:15:32.115287 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9hxgw" event={"ID":"e2001db0-90f1-40a7-ba6e-982eabebc117","Type":"ContainerDied","Data":"2da96709c1fa418850efce465292ff8ccc3b2cc56418b6491ca3ccfee1016cc4"} Dec 06 04:15:32 crc kubenswrapper[4802]: I1206 04:15:32.115346 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2da96709c1fa418850efce465292ff8ccc3b2cc56418b6491ca3ccfee1016cc4" Dec 06 04:15:32 crc kubenswrapper[4802]: I1206 04:15:32.115382 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9hxgw" Dec 06 04:15:32 crc kubenswrapper[4802]: I1206 04:15:32.220159 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-lpc95"] Dec 06 04:15:32 crc kubenswrapper[4802]: E1206 04:15:32.220550 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2001db0-90f1-40a7-ba6e-982eabebc117" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 06 04:15:32 crc kubenswrapper[4802]: I1206 04:15:32.220567 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2001db0-90f1-40a7-ba6e-982eabebc117" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 06 04:15:32 crc kubenswrapper[4802]: E1206 04:15:32.220615 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8189b4bc-09c3-4075-b959-bb67bf6ccd34" containerName="collect-profiles" Dec 06 04:15:32 crc kubenswrapper[4802]: I1206 04:15:32.220622 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="8189b4bc-09c3-4075-b959-bb67bf6ccd34" containerName="collect-profiles" Dec 06 04:15:32 crc kubenswrapper[4802]: I1206 04:15:32.221046 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2001db0-90f1-40a7-ba6e-982eabebc117" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 06 04:15:32 crc kubenswrapper[4802]: I1206 04:15:32.221083 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="8189b4bc-09c3-4075-b959-bb67bf6ccd34" containerName="collect-profiles" Dec 06 04:15:32 crc kubenswrapper[4802]: I1206 04:15:32.221711 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-lpc95" Dec 06 04:15:32 crc kubenswrapper[4802]: I1206 04:15:32.224012 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-rgvjc" Dec 06 04:15:32 crc kubenswrapper[4802]: I1206 04:15:32.224200 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Dec 06 04:15:32 crc kubenswrapper[4802]: I1206 04:15:32.226156 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Dec 06 04:15:32 crc kubenswrapper[4802]: I1206 04:15:32.226208 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 04:15:32 crc kubenswrapper[4802]: I1206 04:15:32.226261 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 04:15:32 crc kubenswrapper[4802]: I1206 04:15:32.226594 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 04:15:32 crc kubenswrapper[4802]: I1206 04:15:32.255296 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-lpc95"] Dec 06 04:15:32 crc kubenswrapper[4802]: I1206 04:15:32.342709 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64272e53-eccf-4cb7-9176-0bc95c5b7e10-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-lpc95\" (UID: \"64272e53-eccf-4cb7-9176-0bc95c5b7e10\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-lpc95" Dec 06 04:15:32 crc kubenswrapper[4802]: I1206 04:15:32.342805 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/64272e53-eccf-4cb7-9176-0bc95c5b7e10-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-lpc95\" (UID: \"64272e53-eccf-4cb7-9176-0bc95c5b7e10\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-lpc95" Dec 06 04:15:32 crc kubenswrapper[4802]: I1206 04:15:32.342850 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxtcx\" (UniqueName: \"kubernetes.io/projected/64272e53-eccf-4cb7-9176-0bc95c5b7e10-kube-api-access-kxtcx\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-lpc95\" (UID: \"64272e53-eccf-4cb7-9176-0bc95c5b7e10\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-lpc95" Dec 06 04:15:32 crc kubenswrapper[4802]: I1206 04:15:32.343002 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/64272e53-eccf-4cb7-9176-0bc95c5b7e10-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-lpc95\" (UID: \"64272e53-eccf-4cb7-9176-0bc95c5b7e10\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-lpc95" Dec 06 04:15:32 crc kubenswrapper[4802]: I1206 04:15:32.343041 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/64272e53-eccf-4cb7-9176-0bc95c5b7e10-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-lpc95\" (UID: \"64272e53-eccf-4cb7-9176-0bc95c5b7e10\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-lpc95" Dec 06 04:15:32 crc kubenswrapper[4802]: I1206 04:15:32.343408 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/64272e53-eccf-4cb7-9176-0bc95c5b7e10-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-lpc95\" (UID: \"64272e53-eccf-4cb7-9176-0bc95c5b7e10\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-lpc95" Dec 06 04:15:32 crc kubenswrapper[4802]: I1206 04:15:32.447120 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64272e53-eccf-4cb7-9176-0bc95c5b7e10-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-lpc95\" (UID: \"64272e53-eccf-4cb7-9176-0bc95c5b7e10\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-lpc95" Dec 06 04:15:32 crc kubenswrapper[4802]: I1206 04:15:32.447182 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/64272e53-eccf-4cb7-9176-0bc95c5b7e10-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-lpc95\" (UID: \"64272e53-eccf-4cb7-9176-0bc95c5b7e10\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-lpc95" Dec 06 04:15:32 crc kubenswrapper[4802]: I1206 04:15:32.447229 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxtcx\" (UniqueName: \"kubernetes.io/projected/64272e53-eccf-4cb7-9176-0bc95c5b7e10-kube-api-access-kxtcx\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-lpc95\" (UID: \"64272e53-eccf-4cb7-9176-0bc95c5b7e10\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-lpc95" Dec 06 04:15:32 crc kubenswrapper[4802]: I1206 04:15:32.447308 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/64272e53-eccf-4cb7-9176-0bc95c5b7e10-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-lpc95\" (UID: \"64272e53-eccf-4cb7-9176-0bc95c5b7e10\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-lpc95" Dec 06 04:15:32 crc kubenswrapper[4802]: I1206 04:15:32.447352 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/64272e53-eccf-4cb7-9176-0bc95c5b7e10-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-lpc95\" (UID: \"64272e53-eccf-4cb7-9176-0bc95c5b7e10\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-lpc95" Dec 06 04:15:32 crc kubenswrapper[4802]: I1206 04:15:32.447489 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/64272e53-eccf-4cb7-9176-0bc95c5b7e10-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-lpc95\" (UID: \"64272e53-eccf-4cb7-9176-0bc95c5b7e10\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-lpc95" Dec 06 04:15:32 crc kubenswrapper[4802]: I1206 04:15:32.452986 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/64272e53-eccf-4cb7-9176-0bc95c5b7e10-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-lpc95\" (UID: \"64272e53-eccf-4cb7-9176-0bc95c5b7e10\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-lpc95" Dec 06 04:15:32 crc kubenswrapper[4802]: I1206 04:15:32.453902 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/64272e53-eccf-4cb7-9176-0bc95c5b7e10-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-lpc95\" (UID: \"64272e53-eccf-4cb7-9176-0bc95c5b7e10\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-lpc95" Dec 06 04:15:32 crc kubenswrapper[4802]: I1206 04:15:32.454074 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/64272e53-eccf-4cb7-9176-0bc95c5b7e10-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-lpc95\" (UID: \"64272e53-eccf-4cb7-9176-0bc95c5b7e10\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-lpc95" Dec 06 04:15:32 crc kubenswrapper[4802]: I1206 04:15:32.454195 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64272e53-eccf-4cb7-9176-0bc95c5b7e10-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-lpc95\" (UID: \"64272e53-eccf-4cb7-9176-0bc95c5b7e10\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-lpc95" Dec 06 04:15:32 crc kubenswrapper[4802]: I1206 04:15:32.459985 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/64272e53-eccf-4cb7-9176-0bc95c5b7e10-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-lpc95\" (UID: \"64272e53-eccf-4cb7-9176-0bc95c5b7e10\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-lpc95" Dec 06 04:15:32 crc kubenswrapper[4802]: I1206 04:15:32.464022 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxtcx\" (UniqueName: \"kubernetes.io/projected/64272e53-eccf-4cb7-9176-0bc95c5b7e10-kube-api-access-kxtcx\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-lpc95\" (UID: \"64272e53-eccf-4cb7-9176-0bc95c5b7e10\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-lpc95" Dec 06 04:15:32 crc kubenswrapper[4802]: I1206 04:15:32.546762 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-lpc95" Dec 06 04:15:33 crc kubenswrapper[4802]: I1206 04:15:33.074001 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-lpc95"] Dec 06 04:15:33 crc kubenswrapper[4802]: I1206 04:15:33.127928 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-lpc95" event={"ID":"64272e53-eccf-4cb7-9176-0bc95c5b7e10","Type":"ContainerStarted","Data":"9cd4a6ddaec87837861f48fe0b80f3d5be5f7938241a9cf0338cbbf429b8374f"} Dec 06 04:15:34 crc kubenswrapper[4802]: I1206 04:15:34.137083 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-lpc95" event={"ID":"64272e53-eccf-4cb7-9176-0bc95c5b7e10","Type":"ContainerStarted","Data":"e0b86fa299a56135e33fb0ee5732f2432c7b6fd89adee72c6c88071d6e83c479"} Dec 06 04:15:34 crc kubenswrapper[4802]: I1206 04:15:34.160507 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-lpc95" podStartSLOduration=1.7097722499999999 podStartE2EDuration="2.160490226s" podCreationTimestamp="2025-12-06 04:15:32 +0000 UTC" firstStartedPulling="2025-12-06 04:15:33.082866905 +0000 UTC m=+2125.954776057" lastFinishedPulling="2025-12-06 04:15:33.533584841 +0000 UTC m=+2126.405494033" observedRunningTime="2025-12-06 04:15:34.156034684 +0000 UTC m=+2127.027943836" watchObservedRunningTime="2025-12-06 04:15:34.160490226 +0000 UTC m=+2127.032399378" Dec 06 04:16:13 crc kubenswrapper[4802]: I1206 04:16:13.283845 4802 patch_prober.go:28] interesting pod/machine-config-daemon-zpxxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:16:13 crc kubenswrapper[4802]: I1206 04:16:13.284414 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:16:24 crc kubenswrapper[4802]: I1206 04:16:24.666037 4802 generic.go:334] "Generic (PLEG): container finished" podID="64272e53-eccf-4cb7-9176-0bc95c5b7e10" containerID="e0b86fa299a56135e33fb0ee5732f2432c7b6fd89adee72c6c88071d6e83c479" exitCode=0 Dec 06 04:16:24 crc kubenswrapper[4802]: I1206 04:16:24.666427 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-lpc95" event={"ID":"64272e53-eccf-4cb7-9176-0bc95c5b7e10","Type":"ContainerDied","Data":"e0b86fa299a56135e33fb0ee5732f2432c7b6fd89adee72c6c88071d6e83c479"} Dec 06 04:16:26 crc kubenswrapper[4802]: I1206 04:16:26.138573 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-lpc95" Dec 06 04:16:26 crc kubenswrapper[4802]: I1206 04:16:26.242185 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/64272e53-eccf-4cb7-9176-0bc95c5b7e10-ssh-key\") pod \"64272e53-eccf-4cb7-9176-0bc95c5b7e10\" (UID: \"64272e53-eccf-4cb7-9176-0bc95c5b7e10\") " Dec 06 04:16:26 crc kubenswrapper[4802]: I1206 04:16:26.242230 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kxtcx\" (UniqueName: \"kubernetes.io/projected/64272e53-eccf-4cb7-9176-0bc95c5b7e10-kube-api-access-kxtcx\") pod \"64272e53-eccf-4cb7-9176-0bc95c5b7e10\" (UID: \"64272e53-eccf-4cb7-9176-0bc95c5b7e10\") " Dec 06 04:16:26 crc kubenswrapper[4802]: I1206 04:16:26.242261 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/64272e53-eccf-4cb7-9176-0bc95c5b7e10-inventory\") pod \"64272e53-eccf-4cb7-9176-0bc95c5b7e10\" (UID: \"64272e53-eccf-4cb7-9176-0bc95c5b7e10\") " Dec 06 04:16:26 crc kubenswrapper[4802]: I1206 04:16:26.242308 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/64272e53-eccf-4cb7-9176-0bc95c5b7e10-neutron-ovn-metadata-agent-neutron-config-0\") pod \"64272e53-eccf-4cb7-9176-0bc95c5b7e10\" (UID: \"64272e53-eccf-4cb7-9176-0bc95c5b7e10\") " Dec 06 04:16:26 crc kubenswrapper[4802]: I1206 04:16:26.242337 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64272e53-eccf-4cb7-9176-0bc95c5b7e10-neutron-metadata-combined-ca-bundle\") pod \"64272e53-eccf-4cb7-9176-0bc95c5b7e10\" (UID: \"64272e53-eccf-4cb7-9176-0bc95c5b7e10\") " Dec 06 04:16:26 crc kubenswrapper[4802]: I1206 04:16:26.242384 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/64272e53-eccf-4cb7-9176-0bc95c5b7e10-nova-metadata-neutron-config-0\") pod \"64272e53-eccf-4cb7-9176-0bc95c5b7e10\" (UID: \"64272e53-eccf-4cb7-9176-0bc95c5b7e10\") " Dec 06 04:16:26 crc kubenswrapper[4802]: I1206 04:16:26.249240 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64272e53-eccf-4cb7-9176-0bc95c5b7e10-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "64272e53-eccf-4cb7-9176-0bc95c5b7e10" (UID: "64272e53-eccf-4cb7-9176-0bc95c5b7e10"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:16:26 crc kubenswrapper[4802]: I1206 04:16:26.249707 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64272e53-eccf-4cb7-9176-0bc95c5b7e10-kube-api-access-kxtcx" (OuterVolumeSpecName: "kube-api-access-kxtcx") pod "64272e53-eccf-4cb7-9176-0bc95c5b7e10" (UID: "64272e53-eccf-4cb7-9176-0bc95c5b7e10"). InnerVolumeSpecName "kube-api-access-kxtcx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:16:26 crc kubenswrapper[4802]: I1206 04:16:26.271902 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64272e53-eccf-4cb7-9176-0bc95c5b7e10-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "64272e53-eccf-4cb7-9176-0bc95c5b7e10" (UID: "64272e53-eccf-4cb7-9176-0bc95c5b7e10"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:16:26 crc kubenswrapper[4802]: I1206 04:16:26.273141 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64272e53-eccf-4cb7-9176-0bc95c5b7e10-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "64272e53-eccf-4cb7-9176-0bc95c5b7e10" (UID: "64272e53-eccf-4cb7-9176-0bc95c5b7e10"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:16:26 crc kubenswrapper[4802]: I1206 04:16:26.278370 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64272e53-eccf-4cb7-9176-0bc95c5b7e10-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "64272e53-eccf-4cb7-9176-0bc95c5b7e10" (UID: "64272e53-eccf-4cb7-9176-0bc95c5b7e10"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:16:26 crc kubenswrapper[4802]: I1206 04:16:26.283411 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64272e53-eccf-4cb7-9176-0bc95c5b7e10-inventory" (OuterVolumeSpecName: "inventory") pod "64272e53-eccf-4cb7-9176-0bc95c5b7e10" (UID: "64272e53-eccf-4cb7-9176-0bc95c5b7e10"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:16:26 crc kubenswrapper[4802]: I1206 04:16:26.344306 4802 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/64272e53-eccf-4cb7-9176-0bc95c5b7e10-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 04:16:26 crc kubenswrapper[4802]: I1206 04:16:26.344345 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kxtcx\" (UniqueName: \"kubernetes.io/projected/64272e53-eccf-4cb7-9176-0bc95c5b7e10-kube-api-access-kxtcx\") on node \"crc\" DevicePath \"\"" Dec 06 04:16:26 crc kubenswrapper[4802]: I1206 04:16:26.344356 4802 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/64272e53-eccf-4cb7-9176-0bc95c5b7e10-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 04:16:26 crc kubenswrapper[4802]: I1206 04:16:26.344366 4802 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/64272e53-eccf-4cb7-9176-0bc95c5b7e10-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 06 04:16:26 crc kubenswrapper[4802]: I1206 04:16:26.344377 4802 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64272e53-eccf-4cb7-9176-0bc95c5b7e10-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:16:26 crc kubenswrapper[4802]: I1206 04:16:26.344388 4802 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/64272e53-eccf-4cb7-9176-0bc95c5b7e10-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 06 04:16:26 crc kubenswrapper[4802]: I1206 04:16:26.688267 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-lpc95" event={"ID":"64272e53-eccf-4cb7-9176-0bc95c5b7e10","Type":"ContainerDied","Data":"9cd4a6ddaec87837861f48fe0b80f3d5be5f7938241a9cf0338cbbf429b8374f"} Dec 06 04:16:26 crc kubenswrapper[4802]: I1206 04:16:26.688325 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9cd4a6ddaec87837861f48fe0b80f3d5be5f7938241a9cf0338cbbf429b8374f" Dec 06 04:16:26 crc kubenswrapper[4802]: I1206 04:16:26.688367 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-lpc95" Dec 06 04:16:26 crc kubenswrapper[4802]: I1206 04:16:26.817628 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tckw2"] Dec 06 04:16:26 crc kubenswrapper[4802]: E1206 04:16:26.819022 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64272e53-eccf-4cb7-9176-0bc95c5b7e10" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 06 04:16:26 crc kubenswrapper[4802]: I1206 04:16:26.819056 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="64272e53-eccf-4cb7-9176-0bc95c5b7e10" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 06 04:16:26 crc kubenswrapper[4802]: I1206 04:16:26.819413 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="64272e53-eccf-4cb7-9176-0bc95c5b7e10" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 06 04:16:26 crc kubenswrapper[4802]: I1206 04:16:26.820626 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tckw2" Dec 06 04:16:26 crc kubenswrapper[4802]: I1206 04:16:26.822927 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 04:16:26 crc kubenswrapper[4802]: I1206 04:16:26.823924 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 04:16:26 crc kubenswrapper[4802]: I1206 04:16:26.824002 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Dec 06 04:16:26 crc kubenswrapper[4802]: I1206 04:16:26.824053 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 04:16:26 crc kubenswrapper[4802]: I1206 04:16:26.826665 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-rgvjc" Dec 06 04:16:26 crc kubenswrapper[4802]: I1206 04:16:26.827499 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tckw2"] Dec 06 04:16:26 crc kubenswrapper[4802]: I1206 04:16:26.955877 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4j6b\" (UniqueName: \"kubernetes.io/projected/18179f2f-334f-48c2-a6be-6d1c1792e600-kube-api-access-f4j6b\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tckw2\" (UID: \"18179f2f-334f-48c2-a6be-6d1c1792e600\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tckw2" Dec 06 04:16:26 crc kubenswrapper[4802]: I1206 04:16:26.956315 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18179f2f-334f-48c2-a6be-6d1c1792e600-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tckw2\" (UID: \"18179f2f-334f-48c2-a6be-6d1c1792e600\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tckw2" Dec 06 04:16:26 crc kubenswrapper[4802]: I1206 04:16:26.956530 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/18179f2f-334f-48c2-a6be-6d1c1792e600-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tckw2\" (UID: \"18179f2f-334f-48c2-a6be-6d1c1792e600\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tckw2" Dec 06 04:16:26 crc kubenswrapper[4802]: I1206 04:16:26.956854 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18179f2f-334f-48c2-a6be-6d1c1792e600-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tckw2\" (UID: \"18179f2f-334f-48c2-a6be-6d1c1792e600\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tckw2" Dec 06 04:16:26 crc kubenswrapper[4802]: I1206 04:16:26.957238 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/18179f2f-334f-48c2-a6be-6d1c1792e600-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tckw2\" (UID: \"18179f2f-334f-48c2-a6be-6d1c1792e600\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tckw2" Dec 06 04:16:27 crc kubenswrapper[4802]: I1206 04:16:27.059455 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/18179f2f-334f-48c2-a6be-6d1c1792e600-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tckw2\" (UID: \"18179f2f-334f-48c2-a6be-6d1c1792e600\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tckw2" Dec 06 04:16:27 crc kubenswrapper[4802]: I1206 04:16:27.059548 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4j6b\" (UniqueName: \"kubernetes.io/projected/18179f2f-334f-48c2-a6be-6d1c1792e600-kube-api-access-f4j6b\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tckw2\" (UID: \"18179f2f-334f-48c2-a6be-6d1c1792e600\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tckw2" Dec 06 04:16:27 crc kubenswrapper[4802]: I1206 04:16:27.059580 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18179f2f-334f-48c2-a6be-6d1c1792e600-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tckw2\" (UID: \"18179f2f-334f-48c2-a6be-6d1c1792e600\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tckw2" Dec 06 04:16:27 crc kubenswrapper[4802]: I1206 04:16:27.059634 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/18179f2f-334f-48c2-a6be-6d1c1792e600-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tckw2\" (UID: \"18179f2f-334f-48c2-a6be-6d1c1792e600\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tckw2" Dec 06 04:16:27 crc kubenswrapper[4802]: I1206 04:16:27.059719 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18179f2f-334f-48c2-a6be-6d1c1792e600-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tckw2\" (UID: \"18179f2f-334f-48c2-a6be-6d1c1792e600\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tckw2" Dec 06 04:16:27 crc kubenswrapper[4802]: I1206 04:16:27.063995 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/18179f2f-334f-48c2-a6be-6d1c1792e600-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tckw2\" (UID: \"18179f2f-334f-48c2-a6be-6d1c1792e600\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tckw2" Dec 06 04:16:27 crc kubenswrapper[4802]: I1206 04:16:27.064898 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18179f2f-334f-48c2-a6be-6d1c1792e600-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tckw2\" (UID: \"18179f2f-334f-48c2-a6be-6d1c1792e600\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tckw2" Dec 06 04:16:27 crc kubenswrapper[4802]: I1206 04:16:27.066818 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/18179f2f-334f-48c2-a6be-6d1c1792e600-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tckw2\" (UID: \"18179f2f-334f-48c2-a6be-6d1c1792e600\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tckw2" Dec 06 04:16:27 crc kubenswrapper[4802]: I1206 04:16:27.066963 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18179f2f-334f-48c2-a6be-6d1c1792e600-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tckw2\" (UID: \"18179f2f-334f-48c2-a6be-6d1c1792e600\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tckw2" Dec 06 04:16:27 crc kubenswrapper[4802]: I1206 04:16:27.084899 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4j6b\" (UniqueName: \"kubernetes.io/projected/18179f2f-334f-48c2-a6be-6d1c1792e600-kube-api-access-f4j6b\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tckw2\" (UID: \"18179f2f-334f-48c2-a6be-6d1c1792e600\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tckw2" Dec 06 04:16:27 crc kubenswrapper[4802]: I1206 04:16:27.141721 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tckw2" Dec 06 04:16:27 crc kubenswrapper[4802]: I1206 04:16:27.727850 4802 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 04:16:27 crc kubenswrapper[4802]: I1206 04:16:27.729218 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tckw2"] Dec 06 04:16:28 crc kubenswrapper[4802]: I1206 04:16:28.711173 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tckw2" event={"ID":"18179f2f-334f-48c2-a6be-6d1c1792e600","Type":"ContainerStarted","Data":"c4b2ce0306615d6526357d3ee8392af4a43bf9e7c8579cbaf02c2c03100e3ab4"} Dec 06 04:16:28 crc kubenswrapper[4802]: I1206 04:16:28.711510 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tckw2" event={"ID":"18179f2f-334f-48c2-a6be-6d1c1792e600","Type":"ContainerStarted","Data":"e07fc280ce32e557c76f79d7eefeceb843334dfea6d5e0a0a40571cfb091499c"} Dec 06 04:16:28 crc kubenswrapper[4802]: I1206 04:16:28.736630 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tckw2" podStartSLOduration=2.31155165 podStartE2EDuration="2.736613259s" podCreationTimestamp="2025-12-06 04:16:26 +0000 UTC" firstStartedPulling="2025-12-06 04:16:27.727622671 +0000 UTC m=+2180.599531823" lastFinishedPulling="2025-12-06 04:16:28.15268428 +0000 UTC m=+2181.024593432" observedRunningTime="2025-12-06 04:16:28.729134394 +0000 UTC m=+2181.601043546" watchObservedRunningTime="2025-12-06 04:16:28.736613259 +0000 UTC m=+2181.608522411" Dec 06 04:16:31 crc kubenswrapper[4802]: I1206 04:16:31.612690 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hb8q5"] Dec 06 04:16:31 crc kubenswrapper[4802]: I1206 04:16:31.615321 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hb8q5" Dec 06 04:16:31 crc kubenswrapper[4802]: I1206 04:16:31.648606 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hb8q5"] Dec 06 04:16:31 crc kubenswrapper[4802]: I1206 04:16:31.753153 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4146a6b-8ed7-4ec3-aa94-14e034cf6f47-catalog-content\") pod \"certified-operators-hb8q5\" (UID: \"d4146a6b-8ed7-4ec3-aa94-14e034cf6f47\") " pod="openshift-marketplace/certified-operators-hb8q5" Dec 06 04:16:31 crc kubenswrapper[4802]: I1206 04:16:31.753214 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7srf\" (UniqueName: \"kubernetes.io/projected/d4146a6b-8ed7-4ec3-aa94-14e034cf6f47-kube-api-access-n7srf\") pod \"certified-operators-hb8q5\" (UID: \"d4146a6b-8ed7-4ec3-aa94-14e034cf6f47\") " pod="openshift-marketplace/certified-operators-hb8q5" Dec 06 04:16:31 crc kubenswrapper[4802]: I1206 04:16:31.753371 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4146a6b-8ed7-4ec3-aa94-14e034cf6f47-utilities\") pod \"certified-operators-hb8q5\" (UID: \"d4146a6b-8ed7-4ec3-aa94-14e034cf6f47\") " pod="openshift-marketplace/certified-operators-hb8q5" Dec 06 04:16:31 crc kubenswrapper[4802]: I1206 04:16:31.855535 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4146a6b-8ed7-4ec3-aa94-14e034cf6f47-utilities\") pod \"certified-operators-hb8q5\" (UID: \"d4146a6b-8ed7-4ec3-aa94-14e034cf6f47\") " pod="openshift-marketplace/certified-operators-hb8q5" Dec 06 04:16:31 crc kubenswrapper[4802]: I1206 04:16:31.856126 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4146a6b-8ed7-4ec3-aa94-14e034cf6f47-utilities\") pod \"certified-operators-hb8q5\" (UID: \"d4146a6b-8ed7-4ec3-aa94-14e034cf6f47\") " pod="openshift-marketplace/certified-operators-hb8q5" Dec 06 04:16:31 crc kubenswrapper[4802]: I1206 04:16:31.856533 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4146a6b-8ed7-4ec3-aa94-14e034cf6f47-catalog-content\") pod \"certified-operators-hb8q5\" (UID: \"d4146a6b-8ed7-4ec3-aa94-14e034cf6f47\") " pod="openshift-marketplace/certified-operators-hb8q5" Dec 06 04:16:31 crc kubenswrapper[4802]: I1206 04:16:31.856563 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4146a6b-8ed7-4ec3-aa94-14e034cf6f47-catalog-content\") pod \"certified-operators-hb8q5\" (UID: \"d4146a6b-8ed7-4ec3-aa94-14e034cf6f47\") " pod="openshift-marketplace/certified-operators-hb8q5" Dec 06 04:16:31 crc kubenswrapper[4802]: I1206 04:16:31.856584 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7srf\" (UniqueName: \"kubernetes.io/projected/d4146a6b-8ed7-4ec3-aa94-14e034cf6f47-kube-api-access-n7srf\") pod \"certified-operators-hb8q5\" (UID: \"d4146a6b-8ed7-4ec3-aa94-14e034cf6f47\") " pod="openshift-marketplace/certified-operators-hb8q5" Dec 06 04:16:31 crc kubenswrapper[4802]: I1206 04:16:31.879536 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7srf\" (UniqueName: \"kubernetes.io/projected/d4146a6b-8ed7-4ec3-aa94-14e034cf6f47-kube-api-access-n7srf\") pod \"certified-operators-hb8q5\" (UID: \"d4146a6b-8ed7-4ec3-aa94-14e034cf6f47\") " pod="openshift-marketplace/certified-operators-hb8q5" Dec 06 04:16:31 crc kubenswrapper[4802]: I1206 04:16:31.946351 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hb8q5" Dec 06 04:16:32 crc kubenswrapper[4802]: I1206 04:16:32.517740 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hb8q5"] Dec 06 04:16:32 crc kubenswrapper[4802]: W1206 04:16:32.519099 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd4146a6b_8ed7_4ec3_aa94_14e034cf6f47.slice/crio-a29e5eef21a7c6a6d5d59f1520aa6c9be68cd5bd07dbfb5fd47e8627753fdda1 WatchSource:0}: Error finding container a29e5eef21a7c6a6d5d59f1520aa6c9be68cd5bd07dbfb5fd47e8627753fdda1: Status 404 returned error can't find the container with id a29e5eef21a7c6a6d5d59f1520aa6c9be68cd5bd07dbfb5fd47e8627753fdda1 Dec 06 04:16:32 crc kubenswrapper[4802]: I1206 04:16:32.764711 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hb8q5" event={"ID":"d4146a6b-8ed7-4ec3-aa94-14e034cf6f47","Type":"ContainerStarted","Data":"a29e5eef21a7c6a6d5d59f1520aa6c9be68cd5bd07dbfb5fd47e8627753fdda1"} Dec 06 04:16:33 crc kubenswrapper[4802]: I1206 04:16:33.776107 4802 generic.go:334] "Generic (PLEG): container finished" podID="d4146a6b-8ed7-4ec3-aa94-14e034cf6f47" containerID="c4f3beca5b25690f77ce0d8d8b9a9bec8812afaddc2d3339ed09832ee744c319" exitCode=0 Dec 06 04:16:33 crc kubenswrapper[4802]: I1206 04:16:33.776248 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hb8q5" event={"ID":"d4146a6b-8ed7-4ec3-aa94-14e034cf6f47","Type":"ContainerDied","Data":"c4f3beca5b25690f77ce0d8d8b9a9bec8812afaddc2d3339ed09832ee744c319"} Dec 06 04:16:34 crc kubenswrapper[4802]: I1206 04:16:34.787288 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hb8q5" event={"ID":"d4146a6b-8ed7-4ec3-aa94-14e034cf6f47","Type":"ContainerStarted","Data":"a62ca9f55103b91a299754ff12e2ad7c42bc23e348c7d20b14884c56d7067433"} Dec 06 04:16:35 crc kubenswrapper[4802]: I1206 04:16:35.797685 4802 generic.go:334] "Generic (PLEG): container finished" podID="d4146a6b-8ed7-4ec3-aa94-14e034cf6f47" containerID="a62ca9f55103b91a299754ff12e2ad7c42bc23e348c7d20b14884c56d7067433" exitCode=0 Dec 06 04:16:35 crc kubenswrapper[4802]: I1206 04:16:35.797745 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hb8q5" event={"ID":"d4146a6b-8ed7-4ec3-aa94-14e034cf6f47","Type":"ContainerDied","Data":"a62ca9f55103b91a299754ff12e2ad7c42bc23e348c7d20b14884c56d7067433"} Dec 06 04:16:36 crc kubenswrapper[4802]: I1206 04:16:36.808812 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hb8q5" event={"ID":"d4146a6b-8ed7-4ec3-aa94-14e034cf6f47","Type":"ContainerStarted","Data":"a9a17a682dfb9482cf57515a5b02164c76b04e0b6790d994f18106ff234985c1"} Dec 06 04:16:36 crc kubenswrapper[4802]: I1206 04:16:36.830662 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hb8q5" podStartSLOduration=3.38320506 podStartE2EDuration="5.830640284s" podCreationTimestamp="2025-12-06 04:16:31 +0000 UTC" firstStartedPulling="2025-12-06 04:16:33.778598196 +0000 UTC m=+2186.650507348" lastFinishedPulling="2025-12-06 04:16:36.22603342 +0000 UTC m=+2189.097942572" observedRunningTime="2025-12-06 04:16:36.830088649 +0000 UTC m=+2189.701997811" watchObservedRunningTime="2025-12-06 04:16:36.830640284 +0000 UTC m=+2189.702549456" Dec 06 04:16:41 crc kubenswrapper[4802]: I1206 04:16:41.946939 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hb8q5" Dec 06 04:16:41 crc kubenswrapper[4802]: I1206 04:16:41.947361 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hb8q5" Dec 06 04:16:41 crc kubenswrapper[4802]: I1206 04:16:41.992060 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hb8q5" Dec 06 04:16:42 crc kubenswrapper[4802]: I1206 04:16:42.902498 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hb8q5" Dec 06 04:16:42 crc kubenswrapper[4802]: I1206 04:16:42.945207 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hb8q5"] Dec 06 04:16:43 crc kubenswrapper[4802]: I1206 04:16:43.283521 4802 patch_prober.go:28] interesting pod/machine-config-daemon-zpxxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:16:43 crc kubenswrapper[4802]: I1206 04:16:43.283901 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:16:44 crc kubenswrapper[4802]: I1206 04:16:44.640910 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5kq9w"] Dec 06 04:16:44 crc kubenswrapper[4802]: I1206 04:16:44.644425 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5kq9w" Dec 06 04:16:44 crc kubenswrapper[4802]: I1206 04:16:44.657371 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5kq9w"] Dec 06 04:16:44 crc kubenswrapper[4802]: I1206 04:16:44.701142 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87ffb33b-59aa-4aac-9f02-0fa126759dbc-utilities\") pod \"redhat-marketplace-5kq9w\" (UID: \"87ffb33b-59aa-4aac-9f02-0fa126759dbc\") " pod="openshift-marketplace/redhat-marketplace-5kq9w" Dec 06 04:16:44 crc kubenswrapper[4802]: I1206 04:16:44.701405 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jtzwc\" (UniqueName: \"kubernetes.io/projected/87ffb33b-59aa-4aac-9f02-0fa126759dbc-kube-api-access-jtzwc\") pod \"redhat-marketplace-5kq9w\" (UID: \"87ffb33b-59aa-4aac-9f02-0fa126759dbc\") " pod="openshift-marketplace/redhat-marketplace-5kq9w" Dec 06 04:16:44 crc kubenswrapper[4802]: I1206 04:16:44.701819 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87ffb33b-59aa-4aac-9f02-0fa126759dbc-catalog-content\") pod \"redhat-marketplace-5kq9w\" (UID: \"87ffb33b-59aa-4aac-9f02-0fa126759dbc\") " pod="openshift-marketplace/redhat-marketplace-5kq9w" Dec 06 04:16:44 crc kubenswrapper[4802]: I1206 04:16:44.803183 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87ffb33b-59aa-4aac-9f02-0fa126759dbc-catalog-content\") pod \"redhat-marketplace-5kq9w\" (UID: \"87ffb33b-59aa-4aac-9f02-0fa126759dbc\") " pod="openshift-marketplace/redhat-marketplace-5kq9w" Dec 06 04:16:44 crc kubenswrapper[4802]: I1206 04:16:44.803254 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87ffb33b-59aa-4aac-9f02-0fa126759dbc-utilities\") pod \"redhat-marketplace-5kq9w\" (UID: \"87ffb33b-59aa-4aac-9f02-0fa126759dbc\") " pod="openshift-marketplace/redhat-marketplace-5kq9w" Dec 06 04:16:44 crc kubenswrapper[4802]: I1206 04:16:44.803325 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jtzwc\" (UniqueName: \"kubernetes.io/projected/87ffb33b-59aa-4aac-9f02-0fa126759dbc-kube-api-access-jtzwc\") pod \"redhat-marketplace-5kq9w\" (UID: \"87ffb33b-59aa-4aac-9f02-0fa126759dbc\") " pod="openshift-marketplace/redhat-marketplace-5kq9w" Dec 06 04:16:44 crc kubenswrapper[4802]: I1206 04:16:44.803680 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87ffb33b-59aa-4aac-9f02-0fa126759dbc-catalog-content\") pod \"redhat-marketplace-5kq9w\" (UID: \"87ffb33b-59aa-4aac-9f02-0fa126759dbc\") " pod="openshift-marketplace/redhat-marketplace-5kq9w" Dec 06 04:16:44 crc kubenswrapper[4802]: I1206 04:16:44.803772 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87ffb33b-59aa-4aac-9f02-0fa126759dbc-utilities\") pod \"redhat-marketplace-5kq9w\" (UID: \"87ffb33b-59aa-4aac-9f02-0fa126759dbc\") " pod="openshift-marketplace/redhat-marketplace-5kq9w" Dec 06 04:16:44 crc kubenswrapper[4802]: I1206 04:16:44.836811 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jtzwc\" (UniqueName: \"kubernetes.io/projected/87ffb33b-59aa-4aac-9f02-0fa126759dbc-kube-api-access-jtzwc\") pod \"redhat-marketplace-5kq9w\" (UID: \"87ffb33b-59aa-4aac-9f02-0fa126759dbc\") " pod="openshift-marketplace/redhat-marketplace-5kq9w" Dec 06 04:16:44 crc kubenswrapper[4802]: I1206 04:16:44.877769 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-hb8q5" podUID="d4146a6b-8ed7-4ec3-aa94-14e034cf6f47" containerName="registry-server" containerID="cri-o://a9a17a682dfb9482cf57515a5b02164c76b04e0b6790d994f18106ff234985c1" gracePeriod=2 Dec 06 04:16:44 crc kubenswrapper[4802]: I1206 04:16:44.979829 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5kq9w" Dec 06 04:16:45 crc kubenswrapper[4802]: I1206 04:16:45.434152 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hb8q5" Dec 06 04:16:45 crc kubenswrapper[4802]: I1206 04:16:45.513359 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4146a6b-8ed7-4ec3-aa94-14e034cf6f47-catalog-content\") pod \"d4146a6b-8ed7-4ec3-aa94-14e034cf6f47\" (UID: \"d4146a6b-8ed7-4ec3-aa94-14e034cf6f47\") " Dec 06 04:16:45 crc kubenswrapper[4802]: I1206 04:16:45.513470 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4146a6b-8ed7-4ec3-aa94-14e034cf6f47-utilities\") pod \"d4146a6b-8ed7-4ec3-aa94-14e034cf6f47\" (UID: \"d4146a6b-8ed7-4ec3-aa94-14e034cf6f47\") " Dec 06 04:16:45 crc kubenswrapper[4802]: I1206 04:16:45.513676 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n7srf\" (UniqueName: \"kubernetes.io/projected/d4146a6b-8ed7-4ec3-aa94-14e034cf6f47-kube-api-access-n7srf\") pod \"d4146a6b-8ed7-4ec3-aa94-14e034cf6f47\" (UID: \"d4146a6b-8ed7-4ec3-aa94-14e034cf6f47\") " Dec 06 04:16:45 crc kubenswrapper[4802]: I1206 04:16:45.514291 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4146a6b-8ed7-4ec3-aa94-14e034cf6f47-utilities" (OuterVolumeSpecName: "utilities") pod "d4146a6b-8ed7-4ec3-aa94-14e034cf6f47" (UID: "d4146a6b-8ed7-4ec3-aa94-14e034cf6f47"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:16:45 crc kubenswrapper[4802]: I1206 04:16:45.518361 4802 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4146a6b-8ed7-4ec3-aa94-14e034cf6f47-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 04:16:45 crc kubenswrapper[4802]: I1206 04:16:45.518381 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4146a6b-8ed7-4ec3-aa94-14e034cf6f47-kube-api-access-n7srf" (OuterVolumeSpecName: "kube-api-access-n7srf") pod "d4146a6b-8ed7-4ec3-aa94-14e034cf6f47" (UID: "d4146a6b-8ed7-4ec3-aa94-14e034cf6f47"). InnerVolumeSpecName "kube-api-access-n7srf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:16:45 crc kubenswrapper[4802]: I1206 04:16:45.519379 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5kq9w"] Dec 06 04:16:45 crc kubenswrapper[4802]: I1206 04:16:45.576468 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4146a6b-8ed7-4ec3-aa94-14e034cf6f47-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d4146a6b-8ed7-4ec3-aa94-14e034cf6f47" (UID: "d4146a6b-8ed7-4ec3-aa94-14e034cf6f47"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:16:45 crc kubenswrapper[4802]: I1206 04:16:45.620442 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n7srf\" (UniqueName: \"kubernetes.io/projected/d4146a6b-8ed7-4ec3-aa94-14e034cf6f47-kube-api-access-n7srf\") on node \"crc\" DevicePath \"\"" Dec 06 04:16:45 crc kubenswrapper[4802]: I1206 04:16:45.620473 4802 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4146a6b-8ed7-4ec3-aa94-14e034cf6f47-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 04:16:45 crc kubenswrapper[4802]: I1206 04:16:45.889888 4802 generic.go:334] "Generic (PLEG): container finished" podID="87ffb33b-59aa-4aac-9f02-0fa126759dbc" containerID="d69d195ca5b0f276bde28b0cc7eb21eef2faa557845d07c57f877e579dff8079" exitCode=0 Dec 06 04:16:45 crc kubenswrapper[4802]: I1206 04:16:45.889994 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5kq9w" event={"ID":"87ffb33b-59aa-4aac-9f02-0fa126759dbc","Type":"ContainerDied","Data":"d69d195ca5b0f276bde28b0cc7eb21eef2faa557845d07c57f877e579dff8079"} Dec 06 04:16:45 crc kubenswrapper[4802]: I1206 04:16:45.890041 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5kq9w" event={"ID":"87ffb33b-59aa-4aac-9f02-0fa126759dbc","Type":"ContainerStarted","Data":"95d87ae6acf7cd73c47596473c0f4f6d678841e822dab06a02e3d94757989e08"} Dec 06 04:16:45 crc kubenswrapper[4802]: I1206 04:16:45.896831 4802 generic.go:334] "Generic (PLEG): container finished" podID="d4146a6b-8ed7-4ec3-aa94-14e034cf6f47" containerID="a9a17a682dfb9482cf57515a5b02164c76b04e0b6790d994f18106ff234985c1" exitCode=0 Dec 06 04:16:45 crc kubenswrapper[4802]: I1206 04:16:45.896885 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hb8q5" event={"ID":"d4146a6b-8ed7-4ec3-aa94-14e034cf6f47","Type":"ContainerDied","Data":"a9a17a682dfb9482cf57515a5b02164c76b04e0b6790d994f18106ff234985c1"} Dec 06 04:16:45 crc kubenswrapper[4802]: I1206 04:16:45.896913 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hb8q5" Dec 06 04:16:45 crc kubenswrapper[4802]: I1206 04:16:45.896925 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hb8q5" event={"ID":"d4146a6b-8ed7-4ec3-aa94-14e034cf6f47","Type":"ContainerDied","Data":"a29e5eef21a7c6a6d5d59f1520aa6c9be68cd5bd07dbfb5fd47e8627753fdda1"} Dec 06 04:16:45 crc kubenswrapper[4802]: I1206 04:16:45.896955 4802 scope.go:117] "RemoveContainer" containerID="a9a17a682dfb9482cf57515a5b02164c76b04e0b6790d994f18106ff234985c1" Dec 06 04:16:45 crc kubenswrapper[4802]: I1206 04:16:45.966343 4802 scope.go:117] "RemoveContainer" containerID="a62ca9f55103b91a299754ff12e2ad7c42bc23e348c7d20b14884c56d7067433" Dec 06 04:16:45 crc kubenswrapper[4802]: I1206 04:16:45.978747 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hb8q5"] Dec 06 04:16:45 crc kubenswrapper[4802]: I1206 04:16:45.986490 4802 scope.go:117] "RemoveContainer" containerID="c4f3beca5b25690f77ce0d8d8b9a9bec8812afaddc2d3339ed09832ee744c319" Dec 06 04:16:45 crc kubenswrapper[4802]: I1206 04:16:45.989596 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-hb8q5"] Dec 06 04:16:46 crc kubenswrapper[4802]: I1206 04:16:46.028488 4802 scope.go:117] "RemoveContainer" containerID="a9a17a682dfb9482cf57515a5b02164c76b04e0b6790d994f18106ff234985c1" Dec 06 04:16:46 crc kubenswrapper[4802]: E1206 04:16:46.029013 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9a17a682dfb9482cf57515a5b02164c76b04e0b6790d994f18106ff234985c1\": container with ID starting with a9a17a682dfb9482cf57515a5b02164c76b04e0b6790d994f18106ff234985c1 not found: ID does not exist" containerID="a9a17a682dfb9482cf57515a5b02164c76b04e0b6790d994f18106ff234985c1" Dec 06 04:16:46 crc kubenswrapper[4802]: I1206 04:16:46.029049 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9a17a682dfb9482cf57515a5b02164c76b04e0b6790d994f18106ff234985c1"} err="failed to get container status \"a9a17a682dfb9482cf57515a5b02164c76b04e0b6790d994f18106ff234985c1\": rpc error: code = NotFound desc = could not find container \"a9a17a682dfb9482cf57515a5b02164c76b04e0b6790d994f18106ff234985c1\": container with ID starting with a9a17a682dfb9482cf57515a5b02164c76b04e0b6790d994f18106ff234985c1 not found: ID does not exist" Dec 06 04:16:46 crc kubenswrapper[4802]: I1206 04:16:46.029074 4802 scope.go:117] "RemoveContainer" containerID="a62ca9f55103b91a299754ff12e2ad7c42bc23e348c7d20b14884c56d7067433" Dec 06 04:16:46 crc kubenswrapper[4802]: E1206 04:16:46.032044 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a62ca9f55103b91a299754ff12e2ad7c42bc23e348c7d20b14884c56d7067433\": container with ID starting with a62ca9f55103b91a299754ff12e2ad7c42bc23e348c7d20b14884c56d7067433 not found: ID does not exist" containerID="a62ca9f55103b91a299754ff12e2ad7c42bc23e348c7d20b14884c56d7067433" Dec 06 04:16:46 crc kubenswrapper[4802]: I1206 04:16:46.032074 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a62ca9f55103b91a299754ff12e2ad7c42bc23e348c7d20b14884c56d7067433"} err="failed to get container status \"a62ca9f55103b91a299754ff12e2ad7c42bc23e348c7d20b14884c56d7067433\": rpc error: code = NotFound desc = could not find container \"a62ca9f55103b91a299754ff12e2ad7c42bc23e348c7d20b14884c56d7067433\": container with ID starting with a62ca9f55103b91a299754ff12e2ad7c42bc23e348c7d20b14884c56d7067433 not found: ID does not exist" Dec 06 04:16:46 crc kubenswrapper[4802]: I1206 04:16:46.032090 4802 scope.go:117] "RemoveContainer" containerID="c4f3beca5b25690f77ce0d8d8b9a9bec8812afaddc2d3339ed09832ee744c319" Dec 06 04:16:46 crc kubenswrapper[4802]: E1206 04:16:46.032449 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4f3beca5b25690f77ce0d8d8b9a9bec8812afaddc2d3339ed09832ee744c319\": container with ID starting with c4f3beca5b25690f77ce0d8d8b9a9bec8812afaddc2d3339ed09832ee744c319 not found: ID does not exist" containerID="c4f3beca5b25690f77ce0d8d8b9a9bec8812afaddc2d3339ed09832ee744c319" Dec 06 04:16:46 crc kubenswrapper[4802]: I1206 04:16:46.032486 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4f3beca5b25690f77ce0d8d8b9a9bec8812afaddc2d3339ed09832ee744c319"} err="failed to get container status \"c4f3beca5b25690f77ce0d8d8b9a9bec8812afaddc2d3339ed09832ee744c319\": rpc error: code = NotFound desc = could not find container \"c4f3beca5b25690f77ce0d8d8b9a9bec8812afaddc2d3339ed09832ee744c319\": container with ID starting with c4f3beca5b25690f77ce0d8d8b9a9bec8812afaddc2d3339ed09832ee744c319 not found: ID does not exist" Dec 06 04:16:47 crc kubenswrapper[4802]: I1206 04:16:47.464156 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4146a6b-8ed7-4ec3-aa94-14e034cf6f47" path="/var/lib/kubelet/pods/d4146a6b-8ed7-4ec3-aa94-14e034cf6f47/volumes" Dec 06 04:16:47 crc kubenswrapper[4802]: I1206 04:16:47.916210 4802 generic.go:334] "Generic (PLEG): container finished" podID="87ffb33b-59aa-4aac-9f02-0fa126759dbc" containerID="cb8052c043f88229fb92096f8d90e7dfcfa988e4921a12ab0a65a2a0467cd197" exitCode=0 Dec 06 04:16:47 crc kubenswrapper[4802]: I1206 04:16:47.916253 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5kq9w" event={"ID":"87ffb33b-59aa-4aac-9f02-0fa126759dbc","Type":"ContainerDied","Data":"cb8052c043f88229fb92096f8d90e7dfcfa988e4921a12ab0a65a2a0467cd197"} Dec 06 04:16:48 crc kubenswrapper[4802]: I1206 04:16:48.925563 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5kq9w" event={"ID":"87ffb33b-59aa-4aac-9f02-0fa126759dbc","Type":"ContainerStarted","Data":"6d52b2544e265338526c4ed36401c11a4c43b31826d3244c9c721a51ad836409"} Dec 06 04:16:48 crc kubenswrapper[4802]: I1206 04:16:48.951972 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5kq9w" podStartSLOduration=2.539951506 podStartE2EDuration="4.951949863s" podCreationTimestamp="2025-12-06 04:16:44 +0000 UTC" firstStartedPulling="2025-12-06 04:16:45.891946458 +0000 UTC m=+2198.763855650" lastFinishedPulling="2025-12-06 04:16:48.303944855 +0000 UTC m=+2201.175854007" observedRunningTime="2025-12-06 04:16:48.945028305 +0000 UTC m=+2201.816937477" watchObservedRunningTime="2025-12-06 04:16:48.951949863 +0000 UTC m=+2201.823859015" Dec 06 04:16:51 crc kubenswrapper[4802]: I1206 04:16:51.836467 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-tddmb"] Dec 06 04:16:51 crc kubenswrapper[4802]: E1206 04:16:51.837021 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4146a6b-8ed7-4ec3-aa94-14e034cf6f47" containerName="registry-server" Dec 06 04:16:51 crc kubenswrapper[4802]: I1206 04:16:51.837038 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4146a6b-8ed7-4ec3-aa94-14e034cf6f47" containerName="registry-server" Dec 06 04:16:51 crc kubenswrapper[4802]: E1206 04:16:51.837072 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4146a6b-8ed7-4ec3-aa94-14e034cf6f47" containerName="extract-utilities" Dec 06 04:16:51 crc kubenswrapper[4802]: I1206 04:16:51.837081 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4146a6b-8ed7-4ec3-aa94-14e034cf6f47" containerName="extract-utilities" Dec 06 04:16:51 crc kubenswrapper[4802]: E1206 04:16:51.837105 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4146a6b-8ed7-4ec3-aa94-14e034cf6f47" containerName="extract-content" Dec 06 04:16:51 crc kubenswrapper[4802]: I1206 04:16:51.837114 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4146a6b-8ed7-4ec3-aa94-14e034cf6f47" containerName="extract-content" Dec 06 04:16:51 crc kubenswrapper[4802]: I1206 04:16:51.837375 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4146a6b-8ed7-4ec3-aa94-14e034cf6f47" containerName="registry-server" Dec 06 04:16:51 crc kubenswrapper[4802]: I1206 04:16:51.839256 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tddmb" Dec 06 04:16:51 crc kubenswrapper[4802]: I1206 04:16:51.850451 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tddmb"] Dec 06 04:16:51 crc kubenswrapper[4802]: I1206 04:16:51.961347 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4tjwc\" (UniqueName: \"kubernetes.io/projected/71fbffa3-ba3d-4318-9403-cbc002acd0f8-kube-api-access-4tjwc\") pod \"redhat-operators-tddmb\" (UID: \"71fbffa3-ba3d-4318-9403-cbc002acd0f8\") " pod="openshift-marketplace/redhat-operators-tddmb" Dec 06 04:16:51 crc kubenswrapper[4802]: I1206 04:16:51.961640 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71fbffa3-ba3d-4318-9403-cbc002acd0f8-utilities\") pod \"redhat-operators-tddmb\" (UID: \"71fbffa3-ba3d-4318-9403-cbc002acd0f8\") " pod="openshift-marketplace/redhat-operators-tddmb" Dec 06 04:16:51 crc kubenswrapper[4802]: I1206 04:16:51.961994 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71fbffa3-ba3d-4318-9403-cbc002acd0f8-catalog-content\") pod \"redhat-operators-tddmb\" (UID: \"71fbffa3-ba3d-4318-9403-cbc002acd0f8\") " pod="openshift-marketplace/redhat-operators-tddmb" Dec 06 04:16:52 crc kubenswrapper[4802]: I1206 04:16:52.064054 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71fbffa3-ba3d-4318-9403-cbc002acd0f8-utilities\") pod \"redhat-operators-tddmb\" (UID: \"71fbffa3-ba3d-4318-9403-cbc002acd0f8\") " pod="openshift-marketplace/redhat-operators-tddmb" Dec 06 04:16:52 crc kubenswrapper[4802]: I1206 04:16:52.064322 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71fbffa3-ba3d-4318-9403-cbc002acd0f8-catalog-content\") pod \"redhat-operators-tddmb\" (UID: \"71fbffa3-ba3d-4318-9403-cbc002acd0f8\") " pod="openshift-marketplace/redhat-operators-tddmb" Dec 06 04:16:52 crc kubenswrapper[4802]: I1206 04:16:52.064414 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4tjwc\" (UniqueName: \"kubernetes.io/projected/71fbffa3-ba3d-4318-9403-cbc002acd0f8-kube-api-access-4tjwc\") pod \"redhat-operators-tddmb\" (UID: \"71fbffa3-ba3d-4318-9403-cbc002acd0f8\") " pod="openshift-marketplace/redhat-operators-tddmb" Dec 06 04:16:52 crc kubenswrapper[4802]: I1206 04:16:52.064703 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71fbffa3-ba3d-4318-9403-cbc002acd0f8-utilities\") pod \"redhat-operators-tddmb\" (UID: \"71fbffa3-ba3d-4318-9403-cbc002acd0f8\") " pod="openshift-marketplace/redhat-operators-tddmb" Dec 06 04:16:52 crc kubenswrapper[4802]: I1206 04:16:52.064786 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71fbffa3-ba3d-4318-9403-cbc002acd0f8-catalog-content\") pod \"redhat-operators-tddmb\" (UID: \"71fbffa3-ba3d-4318-9403-cbc002acd0f8\") " pod="openshift-marketplace/redhat-operators-tddmb" Dec 06 04:16:52 crc kubenswrapper[4802]: I1206 04:16:52.091341 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4tjwc\" (UniqueName: \"kubernetes.io/projected/71fbffa3-ba3d-4318-9403-cbc002acd0f8-kube-api-access-4tjwc\") pod \"redhat-operators-tddmb\" (UID: \"71fbffa3-ba3d-4318-9403-cbc002acd0f8\") " pod="openshift-marketplace/redhat-operators-tddmb" Dec 06 04:16:52 crc kubenswrapper[4802]: I1206 04:16:52.157591 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tddmb" Dec 06 04:16:52 crc kubenswrapper[4802]: I1206 04:16:52.714410 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tddmb"] Dec 06 04:16:52 crc kubenswrapper[4802]: I1206 04:16:52.964966 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tddmb" event={"ID":"71fbffa3-ba3d-4318-9403-cbc002acd0f8","Type":"ContainerStarted","Data":"406770fd18d8bc742f4810c2fdaa39189a4a9a26b058201abc25e6d05da55191"} Dec 06 04:16:53 crc kubenswrapper[4802]: I1206 04:16:53.973553 4802 generic.go:334] "Generic (PLEG): container finished" podID="71fbffa3-ba3d-4318-9403-cbc002acd0f8" containerID="5ef85a835b633bca739e3caad4f81de8d04e02fd52b702a3633a767f9f61a6cf" exitCode=0 Dec 06 04:16:53 crc kubenswrapper[4802]: I1206 04:16:53.973604 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tddmb" event={"ID":"71fbffa3-ba3d-4318-9403-cbc002acd0f8","Type":"ContainerDied","Data":"5ef85a835b633bca739e3caad4f81de8d04e02fd52b702a3633a767f9f61a6cf"} Dec 06 04:16:54 crc kubenswrapper[4802]: I1206 04:16:54.981018 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5kq9w" Dec 06 04:16:54 crc kubenswrapper[4802]: I1206 04:16:54.981589 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5kq9w" Dec 06 04:16:54 crc kubenswrapper[4802]: I1206 04:16:54.984011 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tddmb" event={"ID":"71fbffa3-ba3d-4318-9403-cbc002acd0f8","Type":"ContainerStarted","Data":"ce2036f823d663aa27a4e8bc5b3acf351cace0d1856e260bf9bea6d541c11b2b"} Dec 06 04:16:55 crc kubenswrapper[4802]: I1206 04:16:55.049516 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5kq9w" Dec 06 04:16:56 crc kubenswrapper[4802]: I1206 04:16:56.080833 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5kq9w" Dec 06 04:16:57 crc kubenswrapper[4802]: I1206 04:16:57.005527 4802 generic.go:334] "Generic (PLEG): container finished" podID="71fbffa3-ba3d-4318-9403-cbc002acd0f8" containerID="ce2036f823d663aa27a4e8bc5b3acf351cace0d1856e260bf9bea6d541c11b2b" exitCode=0 Dec 06 04:16:57 crc kubenswrapper[4802]: I1206 04:16:57.005628 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tddmb" event={"ID":"71fbffa3-ba3d-4318-9403-cbc002acd0f8","Type":"ContainerDied","Data":"ce2036f823d663aa27a4e8bc5b3acf351cace0d1856e260bf9bea6d541c11b2b"} Dec 06 04:16:57 crc kubenswrapper[4802]: I1206 04:16:57.035251 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5kq9w"] Dec 06 04:16:58 crc kubenswrapper[4802]: I1206 04:16:58.017336 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tddmb" event={"ID":"71fbffa3-ba3d-4318-9403-cbc002acd0f8","Type":"ContainerStarted","Data":"fe6b9e12852d46a8d629d5de4e6a5f063fd450d43c9d4d9ff2a967d88298abee"} Dec 06 04:16:58 crc kubenswrapper[4802]: I1206 04:16:58.017729 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5kq9w" podUID="87ffb33b-59aa-4aac-9f02-0fa126759dbc" containerName="registry-server" containerID="cri-o://6d52b2544e265338526c4ed36401c11a4c43b31826d3244c9c721a51ad836409" gracePeriod=2 Dec 06 04:16:58 crc kubenswrapper[4802]: I1206 04:16:58.057169 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-tddmb" podStartSLOduration=3.31167767 podStartE2EDuration="7.057137196s" podCreationTimestamp="2025-12-06 04:16:51 +0000 UTC" firstStartedPulling="2025-12-06 04:16:53.975993402 +0000 UTC m=+2206.847902554" lastFinishedPulling="2025-12-06 04:16:57.721452928 +0000 UTC m=+2210.593362080" observedRunningTime="2025-12-06 04:16:58.055685797 +0000 UTC m=+2210.927594959" watchObservedRunningTime="2025-12-06 04:16:58.057137196 +0000 UTC m=+2210.929046388" Dec 06 04:16:58 crc kubenswrapper[4802]: I1206 04:16:58.498844 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5kq9w" Dec 06 04:16:58 crc kubenswrapper[4802]: I1206 04:16:58.590716 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jtzwc\" (UniqueName: \"kubernetes.io/projected/87ffb33b-59aa-4aac-9f02-0fa126759dbc-kube-api-access-jtzwc\") pod \"87ffb33b-59aa-4aac-9f02-0fa126759dbc\" (UID: \"87ffb33b-59aa-4aac-9f02-0fa126759dbc\") " Dec 06 04:16:58 crc kubenswrapper[4802]: I1206 04:16:58.590887 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87ffb33b-59aa-4aac-9f02-0fa126759dbc-catalog-content\") pod \"87ffb33b-59aa-4aac-9f02-0fa126759dbc\" (UID: \"87ffb33b-59aa-4aac-9f02-0fa126759dbc\") " Dec 06 04:16:58 crc kubenswrapper[4802]: I1206 04:16:58.591057 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87ffb33b-59aa-4aac-9f02-0fa126759dbc-utilities\") pod \"87ffb33b-59aa-4aac-9f02-0fa126759dbc\" (UID: \"87ffb33b-59aa-4aac-9f02-0fa126759dbc\") " Dec 06 04:16:58 crc kubenswrapper[4802]: I1206 04:16:58.591583 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87ffb33b-59aa-4aac-9f02-0fa126759dbc-utilities" (OuterVolumeSpecName: "utilities") pod "87ffb33b-59aa-4aac-9f02-0fa126759dbc" (UID: "87ffb33b-59aa-4aac-9f02-0fa126759dbc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:16:58 crc kubenswrapper[4802]: I1206 04:16:58.597991 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87ffb33b-59aa-4aac-9f02-0fa126759dbc-kube-api-access-jtzwc" (OuterVolumeSpecName: "kube-api-access-jtzwc") pod "87ffb33b-59aa-4aac-9f02-0fa126759dbc" (UID: "87ffb33b-59aa-4aac-9f02-0fa126759dbc"). InnerVolumeSpecName "kube-api-access-jtzwc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:16:58 crc kubenswrapper[4802]: I1206 04:16:58.612233 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87ffb33b-59aa-4aac-9f02-0fa126759dbc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "87ffb33b-59aa-4aac-9f02-0fa126759dbc" (UID: "87ffb33b-59aa-4aac-9f02-0fa126759dbc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:16:58 crc kubenswrapper[4802]: I1206 04:16:58.693399 4802 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87ffb33b-59aa-4aac-9f02-0fa126759dbc-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 04:16:58 crc kubenswrapper[4802]: I1206 04:16:58.693429 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jtzwc\" (UniqueName: \"kubernetes.io/projected/87ffb33b-59aa-4aac-9f02-0fa126759dbc-kube-api-access-jtzwc\") on node \"crc\" DevicePath \"\"" Dec 06 04:16:58 crc kubenswrapper[4802]: I1206 04:16:58.693440 4802 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87ffb33b-59aa-4aac-9f02-0fa126759dbc-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 04:16:59 crc kubenswrapper[4802]: I1206 04:16:59.027338 4802 generic.go:334] "Generic (PLEG): container finished" podID="87ffb33b-59aa-4aac-9f02-0fa126759dbc" containerID="6d52b2544e265338526c4ed36401c11a4c43b31826d3244c9c721a51ad836409" exitCode=0 Dec 06 04:16:59 crc kubenswrapper[4802]: I1206 04:16:59.027391 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5kq9w" event={"ID":"87ffb33b-59aa-4aac-9f02-0fa126759dbc","Type":"ContainerDied","Data":"6d52b2544e265338526c4ed36401c11a4c43b31826d3244c9c721a51ad836409"} Dec 06 04:16:59 crc kubenswrapper[4802]: I1206 04:16:59.027422 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5kq9w" event={"ID":"87ffb33b-59aa-4aac-9f02-0fa126759dbc","Type":"ContainerDied","Data":"95d87ae6acf7cd73c47596473c0f4f6d678841e822dab06a02e3d94757989e08"} Dec 06 04:16:59 crc kubenswrapper[4802]: I1206 04:16:59.027443 4802 scope.go:117] "RemoveContainer" containerID="6d52b2544e265338526c4ed36401c11a4c43b31826d3244c9c721a51ad836409" Dec 06 04:16:59 crc kubenswrapper[4802]: I1206 04:16:59.027458 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5kq9w" Dec 06 04:16:59 crc kubenswrapper[4802]: I1206 04:16:59.052580 4802 scope.go:117] "RemoveContainer" containerID="cb8052c043f88229fb92096f8d90e7dfcfa988e4921a12ab0a65a2a0467cd197" Dec 06 04:16:59 crc kubenswrapper[4802]: I1206 04:16:59.071851 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5kq9w"] Dec 06 04:16:59 crc kubenswrapper[4802]: I1206 04:16:59.081442 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5kq9w"] Dec 06 04:16:59 crc kubenswrapper[4802]: I1206 04:16:59.120601 4802 scope.go:117] "RemoveContainer" containerID="d69d195ca5b0f276bde28b0cc7eb21eef2faa557845d07c57f877e579dff8079" Dec 06 04:16:59 crc kubenswrapper[4802]: I1206 04:16:59.150105 4802 scope.go:117] "RemoveContainer" containerID="6d52b2544e265338526c4ed36401c11a4c43b31826d3244c9c721a51ad836409" Dec 06 04:16:59 crc kubenswrapper[4802]: E1206 04:16:59.150594 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d52b2544e265338526c4ed36401c11a4c43b31826d3244c9c721a51ad836409\": container with ID starting with 6d52b2544e265338526c4ed36401c11a4c43b31826d3244c9c721a51ad836409 not found: ID does not exist" containerID="6d52b2544e265338526c4ed36401c11a4c43b31826d3244c9c721a51ad836409" Dec 06 04:16:59 crc kubenswrapper[4802]: I1206 04:16:59.150633 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d52b2544e265338526c4ed36401c11a4c43b31826d3244c9c721a51ad836409"} err="failed to get container status \"6d52b2544e265338526c4ed36401c11a4c43b31826d3244c9c721a51ad836409\": rpc error: code = NotFound desc = could not find container \"6d52b2544e265338526c4ed36401c11a4c43b31826d3244c9c721a51ad836409\": container with ID starting with 6d52b2544e265338526c4ed36401c11a4c43b31826d3244c9c721a51ad836409 not found: ID does not exist" Dec 06 04:16:59 crc kubenswrapper[4802]: I1206 04:16:59.150661 4802 scope.go:117] "RemoveContainer" containerID="cb8052c043f88229fb92096f8d90e7dfcfa988e4921a12ab0a65a2a0467cd197" Dec 06 04:16:59 crc kubenswrapper[4802]: E1206 04:16:59.150964 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb8052c043f88229fb92096f8d90e7dfcfa988e4921a12ab0a65a2a0467cd197\": container with ID starting with cb8052c043f88229fb92096f8d90e7dfcfa988e4921a12ab0a65a2a0467cd197 not found: ID does not exist" containerID="cb8052c043f88229fb92096f8d90e7dfcfa988e4921a12ab0a65a2a0467cd197" Dec 06 04:16:59 crc kubenswrapper[4802]: I1206 04:16:59.150987 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb8052c043f88229fb92096f8d90e7dfcfa988e4921a12ab0a65a2a0467cd197"} err="failed to get container status \"cb8052c043f88229fb92096f8d90e7dfcfa988e4921a12ab0a65a2a0467cd197\": rpc error: code = NotFound desc = could not find container \"cb8052c043f88229fb92096f8d90e7dfcfa988e4921a12ab0a65a2a0467cd197\": container with ID starting with cb8052c043f88229fb92096f8d90e7dfcfa988e4921a12ab0a65a2a0467cd197 not found: ID does not exist" Dec 06 04:16:59 crc kubenswrapper[4802]: I1206 04:16:59.151002 4802 scope.go:117] "RemoveContainer" containerID="d69d195ca5b0f276bde28b0cc7eb21eef2faa557845d07c57f877e579dff8079" Dec 06 04:16:59 crc kubenswrapper[4802]: E1206 04:16:59.151641 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d69d195ca5b0f276bde28b0cc7eb21eef2faa557845d07c57f877e579dff8079\": container with ID starting with d69d195ca5b0f276bde28b0cc7eb21eef2faa557845d07c57f877e579dff8079 not found: ID does not exist" containerID="d69d195ca5b0f276bde28b0cc7eb21eef2faa557845d07c57f877e579dff8079" Dec 06 04:16:59 crc kubenswrapper[4802]: I1206 04:16:59.151691 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d69d195ca5b0f276bde28b0cc7eb21eef2faa557845d07c57f877e579dff8079"} err="failed to get container status \"d69d195ca5b0f276bde28b0cc7eb21eef2faa557845d07c57f877e579dff8079\": rpc error: code = NotFound desc = could not find container \"d69d195ca5b0f276bde28b0cc7eb21eef2faa557845d07c57f877e579dff8079\": container with ID starting with d69d195ca5b0f276bde28b0cc7eb21eef2faa557845d07c57f877e579dff8079 not found: ID does not exist" Dec 06 04:16:59 crc kubenswrapper[4802]: I1206 04:16:59.464282 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87ffb33b-59aa-4aac-9f02-0fa126759dbc" path="/var/lib/kubelet/pods/87ffb33b-59aa-4aac-9f02-0fa126759dbc/volumes" Dec 06 04:17:02 crc kubenswrapper[4802]: I1206 04:17:02.157803 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-tddmb" Dec 06 04:17:02 crc kubenswrapper[4802]: I1206 04:17:02.158195 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-tddmb" Dec 06 04:17:03 crc kubenswrapper[4802]: I1206 04:17:03.224510 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-tddmb" podUID="71fbffa3-ba3d-4318-9403-cbc002acd0f8" containerName="registry-server" probeResult="failure" output=< Dec 06 04:17:03 crc kubenswrapper[4802]: timeout: failed to connect service ":50051" within 1s Dec 06 04:17:03 crc kubenswrapper[4802]: > Dec 06 04:17:12 crc kubenswrapper[4802]: I1206 04:17:12.214369 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-tddmb" Dec 06 04:17:12 crc kubenswrapper[4802]: I1206 04:17:12.274509 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-tddmb" Dec 06 04:17:12 crc kubenswrapper[4802]: I1206 04:17:12.455712 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tddmb"] Dec 06 04:17:13 crc kubenswrapper[4802]: I1206 04:17:13.283375 4802 patch_prober.go:28] interesting pod/machine-config-daemon-zpxxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:17:13 crc kubenswrapper[4802]: I1206 04:17:13.283455 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:17:13 crc kubenswrapper[4802]: I1206 04:17:13.283528 4802 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" Dec 06 04:17:13 crc kubenswrapper[4802]: I1206 04:17:13.284292 4802 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f8a17abbaa167592a628d9fdb3771d61462952618f6382d447c431c49f7c5690"} pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 04:17:13 crc kubenswrapper[4802]: I1206 04:17:13.284367 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" containerID="cri-o://f8a17abbaa167592a628d9fdb3771d61462952618f6382d447c431c49f7c5690" gracePeriod=600 Dec 06 04:17:13 crc kubenswrapper[4802]: E1206 04:17:13.417295 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:17:14 crc kubenswrapper[4802]: I1206 04:17:14.199843 4802 generic.go:334] "Generic (PLEG): container finished" podID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerID="f8a17abbaa167592a628d9fdb3771d61462952618f6382d447c431c49f7c5690" exitCode=0 Dec 06 04:17:14 crc kubenswrapper[4802]: I1206 04:17:14.199942 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" event={"ID":"6df38316-e0d3-4018-8d27-3620eba3a68d","Type":"ContainerDied","Data":"f8a17abbaa167592a628d9fdb3771d61462952618f6382d447c431c49f7c5690"} Dec 06 04:17:14 crc kubenswrapper[4802]: I1206 04:17:14.200022 4802 scope.go:117] "RemoveContainer" containerID="baa707a57be38104aa835e7ecacc0175c718b272c82894ed9c2693dc0041453b" Dec 06 04:17:14 crc kubenswrapper[4802]: I1206 04:17:14.200069 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-tddmb" podUID="71fbffa3-ba3d-4318-9403-cbc002acd0f8" containerName="registry-server" containerID="cri-o://fe6b9e12852d46a8d629d5de4e6a5f063fd450d43c9d4d9ff2a967d88298abee" gracePeriod=2 Dec 06 04:17:14 crc kubenswrapper[4802]: I1206 04:17:14.200915 4802 scope.go:117] "RemoveContainer" containerID="f8a17abbaa167592a628d9fdb3771d61462952618f6382d447c431c49f7c5690" Dec 06 04:17:14 crc kubenswrapper[4802]: E1206 04:17:14.201330 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:17:14 crc kubenswrapper[4802]: I1206 04:17:14.653377 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tddmb" Dec 06 04:17:14 crc kubenswrapper[4802]: I1206 04:17:14.798148 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71fbffa3-ba3d-4318-9403-cbc002acd0f8-catalog-content\") pod \"71fbffa3-ba3d-4318-9403-cbc002acd0f8\" (UID: \"71fbffa3-ba3d-4318-9403-cbc002acd0f8\") " Dec 06 04:17:14 crc kubenswrapper[4802]: I1206 04:17:14.798821 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71fbffa3-ba3d-4318-9403-cbc002acd0f8-utilities\") pod \"71fbffa3-ba3d-4318-9403-cbc002acd0f8\" (UID: \"71fbffa3-ba3d-4318-9403-cbc002acd0f8\") " Dec 06 04:17:14 crc kubenswrapper[4802]: I1206 04:17:14.798858 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4tjwc\" (UniqueName: \"kubernetes.io/projected/71fbffa3-ba3d-4318-9403-cbc002acd0f8-kube-api-access-4tjwc\") pod \"71fbffa3-ba3d-4318-9403-cbc002acd0f8\" (UID: \"71fbffa3-ba3d-4318-9403-cbc002acd0f8\") " Dec 06 04:17:14 crc kubenswrapper[4802]: I1206 04:17:14.799382 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71fbffa3-ba3d-4318-9403-cbc002acd0f8-utilities" (OuterVolumeSpecName: "utilities") pod "71fbffa3-ba3d-4318-9403-cbc002acd0f8" (UID: "71fbffa3-ba3d-4318-9403-cbc002acd0f8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:17:14 crc kubenswrapper[4802]: I1206 04:17:14.804415 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71fbffa3-ba3d-4318-9403-cbc002acd0f8-kube-api-access-4tjwc" (OuterVolumeSpecName: "kube-api-access-4tjwc") pod "71fbffa3-ba3d-4318-9403-cbc002acd0f8" (UID: "71fbffa3-ba3d-4318-9403-cbc002acd0f8"). InnerVolumeSpecName "kube-api-access-4tjwc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:17:14 crc kubenswrapper[4802]: I1206 04:17:14.900586 4802 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71fbffa3-ba3d-4318-9403-cbc002acd0f8-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 04:17:14 crc kubenswrapper[4802]: I1206 04:17:14.900646 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4tjwc\" (UniqueName: \"kubernetes.io/projected/71fbffa3-ba3d-4318-9403-cbc002acd0f8-kube-api-access-4tjwc\") on node \"crc\" DevicePath \"\"" Dec 06 04:17:14 crc kubenswrapper[4802]: I1206 04:17:14.918007 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71fbffa3-ba3d-4318-9403-cbc002acd0f8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "71fbffa3-ba3d-4318-9403-cbc002acd0f8" (UID: "71fbffa3-ba3d-4318-9403-cbc002acd0f8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:17:15 crc kubenswrapper[4802]: I1206 04:17:15.001891 4802 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71fbffa3-ba3d-4318-9403-cbc002acd0f8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 04:17:15 crc kubenswrapper[4802]: I1206 04:17:15.212620 4802 generic.go:334] "Generic (PLEG): container finished" podID="71fbffa3-ba3d-4318-9403-cbc002acd0f8" containerID="fe6b9e12852d46a8d629d5de4e6a5f063fd450d43c9d4d9ff2a967d88298abee" exitCode=0 Dec 06 04:17:15 crc kubenswrapper[4802]: I1206 04:17:15.212677 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tddmb" event={"ID":"71fbffa3-ba3d-4318-9403-cbc002acd0f8","Type":"ContainerDied","Data":"fe6b9e12852d46a8d629d5de4e6a5f063fd450d43c9d4d9ff2a967d88298abee"} Dec 06 04:17:15 crc kubenswrapper[4802]: I1206 04:17:15.212703 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tddmb" event={"ID":"71fbffa3-ba3d-4318-9403-cbc002acd0f8","Type":"ContainerDied","Data":"406770fd18d8bc742f4810c2fdaa39189a4a9a26b058201abc25e6d05da55191"} Dec 06 04:17:15 crc kubenswrapper[4802]: I1206 04:17:15.212719 4802 scope.go:117] "RemoveContainer" containerID="fe6b9e12852d46a8d629d5de4e6a5f063fd450d43c9d4d9ff2a967d88298abee" Dec 06 04:17:15 crc kubenswrapper[4802]: I1206 04:17:15.212813 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tddmb" Dec 06 04:17:15 crc kubenswrapper[4802]: I1206 04:17:15.252579 4802 scope.go:117] "RemoveContainer" containerID="ce2036f823d663aa27a4e8bc5b3acf351cace0d1856e260bf9bea6d541c11b2b" Dec 06 04:17:15 crc kubenswrapper[4802]: I1206 04:17:15.255234 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tddmb"] Dec 06 04:17:15 crc kubenswrapper[4802]: I1206 04:17:15.265156 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-tddmb"] Dec 06 04:17:15 crc kubenswrapper[4802]: I1206 04:17:15.277716 4802 scope.go:117] "RemoveContainer" containerID="5ef85a835b633bca739e3caad4f81de8d04e02fd52b702a3633a767f9f61a6cf" Dec 06 04:17:15 crc kubenswrapper[4802]: I1206 04:17:15.321451 4802 scope.go:117] "RemoveContainer" containerID="fe6b9e12852d46a8d629d5de4e6a5f063fd450d43c9d4d9ff2a967d88298abee" Dec 06 04:17:15 crc kubenswrapper[4802]: E1206 04:17:15.321989 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe6b9e12852d46a8d629d5de4e6a5f063fd450d43c9d4d9ff2a967d88298abee\": container with ID starting with fe6b9e12852d46a8d629d5de4e6a5f063fd450d43c9d4d9ff2a967d88298abee not found: ID does not exist" containerID="fe6b9e12852d46a8d629d5de4e6a5f063fd450d43c9d4d9ff2a967d88298abee" Dec 06 04:17:15 crc kubenswrapper[4802]: I1206 04:17:15.322020 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe6b9e12852d46a8d629d5de4e6a5f063fd450d43c9d4d9ff2a967d88298abee"} err="failed to get container status \"fe6b9e12852d46a8d629d5de4e6a5f063fd450d43c9d4d9ff2a967d88298abee\": rpc error: code = NotFound desc = could not find container \"fe6b9e12852d46a8d629d5de4e6a5f063fd450d43c9d4d9ff2a967d88298abee\": container with ID starting with fe6b9e12852d46a8d629d5de4e6a5f063fd450d43c9d4d9ff2a967d88298abee not found: ID does not exist" Dec 06 04:17:15 crc kubenswrapper[4802]: I1206 04:17:15.322042 4802 scope.go:117] "RemoveContainer" containerID="ce2036f823d663aa27a4e8bc5b3acf351cace0d1856e260bf9bea6d541c11b2b" Dec 06 04:17:15 crc kubenswrapper[4802]: E1206 04:17:15.322350 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce2036f823d663aa27a4e8bc5b3acf351cace0d1856e260bf9bea6d541c11b2b\": container with ID starting with ce2036f823d663aa27a4e8bc5b3acf351cace0d1856e260bf9bea6d541c11b2b not found: ID does not exist" containerID="ce2036f823d663aa27a4e8bc5b3acf351cace0d1856e260bf9bea6d541c11b2b" Dec 06 04:17:15 crc kubenswrapper[4802]: I1206 04:17:15.322420 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce2036f823d663aa27a4e8bc5b3acf351cace0d1856e260bf9bea6d541c11b2b"} err="failed to get container status \"ce2036f823d663aa27a4e8bc5b3acf351cace0d1856e260bf9bea6d541c11b2b\": rpc error: code = NotFound desc = could not find container \"ce2036f823d663aa27a4e8bc5b3acf351cace0d1856e260bf9bea6d541c11b2b\": container with ID starting with ce2036f823d663aa27a4e8bc5b3acf351cace0d1856e260bf9bea6d541c11b2b not found: ID does not exist" Dec 06 04:17:15 crc kubenswrapper[4802]: I1206 04:17:15.322453 4802 scope.go:117] "RemoveContainer" containerID="5ef85a835b633bca739e3caad4f81de8d04e02fd52b702a3633a767f9f61a6cf" Dec 06 04:17:15 crc kubenswrapper[4802]: E1206 04:17:15.322892 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ef85a835b633bca739e3caad4f81de8d04e02fd52b702a3633a767f9f61a6cf\": container with ID starting with 5ef85a835b633bca739e3caad4f81de8d04e02fd52b702a3633a767f9f61a6cf not found: ID does not exist" containerID="5ef85a835b633bca739e3caad4f81de8d04e02fd52b702a3633a767f9f61a6cf" Dec 06 04:17:15 crc kubenswrapper[4802]: I1206 04:17:15.322945 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ef85a835b633bca739e3caad4f81de8d04e02fd52b702a3633a767f9f61a6cf"} err="failed to get container status \"5ef85a835b633bca739e3caad4f81de8d04e02fd52b702a3633a767f9f61a6cf\": rpc error: code = NotFound desc = could not find container \"5ef85a835b633bca739e3caad4f81de8d04e02fd52b702a3633a767f9f61a6cf\": container with ID starting with 5ef85a835b633bca739e3caad4f81de8d04e02fd52b702a3633a767f9f61a6cf not found: ID does not exist" Dec 06 04:17:15 crc kubenswrapper[4802]: I1206 04:17:15.461406 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71fbffa3-ba3d-4318-9403-cbc002acd0f8" path="/var/lib/kubelet/pods/71fbffa3-ba3d-4318-9403-cbc002acd0f8/volumes" Dec 06 04:17:26 crc kubenswrapper[4802]: I1206 04:17:26.451226 4802 scope.go:117] "RemoveContainer" containerID="f8a17abbaa167592a628d9fdb3771d61462952618f6382d447c431c49f7c5690" Dec 06 04:17:26 crc kubenswrapper[4802]: E1206 04:17:26.452030 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:17:31 crc kubenswrapper[4802]: I1206 04:17:31.997567 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-psfqf"] Dec 06 04:17:31 crc kubenswrapper[4802]: E1206 04:17:31.999075 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71fbffa3-ba3d-4318-9403-cbc002acd0f8" containerName="extract-utilities" Dec 06 04:17:31 crc kubenswrapper[4802]: I1206 04:17:31.999107 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="71fbffa3-ba3d-4318-9403-cbc002acd0f8" containerName="extract-utilities" Dec 06 04:17:31 crc kubenswrapper[4802]: E1206 04:17:31.999132 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87ffb33b-59aa-4aac-9f02-0fa126759dbc" containerName="registry-server" Dec 06 04:17:31 crc kubenswrapper[4802]: I1206 04:17:31.999150 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="87ffb33b-59aa-4aac-9f02-0fa126759dbc" containerName="registry-server" Dec 06 04:17:31 crc kubenswrapper[4802]: E1206 04:17:31.999175 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71fbffa3-ba3d-4318-9403-cbc002acd0f8" containerName="extract-content" Dec 06 04:17:31 crc kubenswrapper[4802]: I1206 04:17:31.999192 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="71fbffa3-ba3d-4318-9403-cbc002acd0f8" containerName="extract-content" Dec 06 04:17:31 crc kubenswrapper[4802]: E1206 04:17:31.999240 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87ffb33b-59aa-4aac-9f02-0fa126759dbc" containerName="extract-utilities" Dec 06 04:17:31 crc kubenswrapper[4802]: I1206 04:17:31.999258 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="87ffb33b-59aa-4aac-9f02-0fa126759dbc" containerName="extract-utilities" Dec 06 04:17:31 crc kubenswrapper[4802]: E1206 04:17:31.999280 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71fbffa3-ba3d-4318-9403-cbc002acd0f8" containerName="registry-server" Dec 06 04:17:31 crc kubenswrapper[4802]: I1206 04:17:31.999297 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="71fbffa3-ba3d-4318-9403-cbc002acd0f8" containerName="registry-server" Dec 06 04:17:31 crc kubenswrapper[4802]: E1206 04:17:31.999323 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87ffb33b-59aa-4aac-9f02-0fa126759dbc" containerName="extract-content" Dec 06 04:17:31 crc kubenswrapper[4802]: I1206 04:17:31.999339 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="87ffb33b-59aa-4aac-9f02-0fa126759dbc" containerName="extract-content" Dec 06 04:17:32 crc kubenswrapper[4802]: I1206 04:17:31.999819 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="71fbffa3-ba3d-4318-9403-cbc002acd0f8" containerName="registry-server" Dec 06 04:17:32 crc kubenswrapper[4802]: I1206 04:17:31.999885 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="87ffb33b-59aa-4aac-9f02-0fa126759dbc" containerName="registry-server" Dec 06 04:17:32 crc kubenswrapper[4802]: I1206 04:17:32.003317 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-psfqf" Dec 06 04:17:32 crc kubenswrapper[4802]: I1206 04:17:32.013820 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-psfqf"] Dec 06 04:17:32 crc kubenswrapper[4802]: I1206 04:17:32.110404 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45d96a62-52e1-4e87-9f5e-2ccbebc57fab-catalog-content\") pod \"community-operators-psfqf\" (UID: \"45d96a62-52e1-4e87-9f5e-2ccbebc57fab\") " pod="openshift-marketplace/community-operators-psfqf" Dec 06 04:17:32 crc kubenswrapper[4802]: I1206 04:17:32.110562 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkffh\" (UniqueName: \"kubernetes.io/projected/45d96a62-52e1-4e87-9f5e-2ccbebc57fab-kube-api-access-zkffh\") pod \"community-operators-psfqf\" (UID: \"45d96a62-52e1-4e87-9f5e-2ccbebc57fab\") " pod="openshift-marketplace/community-operators-psfqf" Dec 06 04:17:32 crc kubenswrapper[4802]: I1206 04:17:32.110636 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45d96a62-52e1-4e87-9f5e-2ccbebc57fab-utilities\") pod \"community-operators-psfqf\" (UID: \"45d96a62-52e1-4e87-9f5e-2ccbebc57fab\") " pod="openshift-marketplace/community-operators-psfqf" Dec 06 04:17:32 crc kubenswrapper[4802]: I1206 04:17:32.212485 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkffh\" (UniqueName: \"kubernetes.io/projected/45d96a62-52e1-4e87-9f5e-2ccbebc57fab-kube-api-access-zkffh\") pod \"community-operators-psfqf\" (UID: \"45d96a62-52e1-4e87-9f5e-2ccbebc57fab\") " pod="openshift-marketplace/community-operators-psfqf" Dec 06 04:17:32 crc kubenswrapper[4802]: I1206 04:17:32.212589 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45d96a62-52e1-4e87-9f5e-2ccbebc57fab-utilities\") pod \"community-operators-psfqf\" (UID: \"45d96a62-52e1-4e87-9f5e-2ccbebc57fab\") " pod="openshift-marketplace/community-operators-psfqf" Dec 06 04:17:32 crc kubenswrapper[4802]: I1206 04:17:32.212654 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45d96a62-52e1-4e87-9f5e-2ccbebc57fab-catalog-content\") pod \"community-operators-psfqf\" (UID: \"45d96a62-52e1-4e87-9f5e-2ccbebc57fab\") " pod="openshift-marketplace/community-operators-psfqf" Dec 06 04:17:32 crc kubenswrapper[4802]: I1206 04:17:32.213201 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45d96a62-52e1-4e87-9f5e-2ccbebc57fab-catalog-content\") pod \"community-operators-psfqf\" (UID: \"45d96a62-52e1-4e87-9f5e-2ccbebc57fab\") " pod="openshift-marketplace/community-operators-psfqf" Dec 06 04:17:32 crc kubenswrapper[4802]: I1206 04:17:32.213734 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45d96a62-52e1-4e87-9f5e-2ccbebc57fab-utilities\") pod \"community-operators-psfqf\" (UID: \"45d96a62-52e1-4e87-9f5e-2ccbebc57fab\") " pod="openshift-marketplace/community-operators-psfqf" Dec 06 04:17:32 crc kubenswrapper[4802]: I1206 04:17:32.233112 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkffh\" (UniqueName: \"kubernetes.io/projected/45d96a62-52e1-4e87-9f5e-2ccbebc57fab-kube-api-access-zkffh\") pod \"community-operators-psfqf\" (UID: \"45d96a62-52e1-4e87-9f5e-2ccbebc57fab\") " pod="openshift-marketplace/community-operators-psfqf" Dec 06 04:17:32 crc kubenswrapper[4802]: I1206 04:17:32.339949 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-psfqf" Dec 06 04:17:32 crc kubenswrapper[4802]: I1206 04:17:32.851737 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-psfqf"] Dec 06 04:17:33 crc kubenswrapper[4802]: I1206 04:17:33.429936 4802 generic.go:334] "Generic (PLEG): container finished" podID="45d96a62-52e1-4e87-9f5e-2ccbebc57fab" containerID="779523f166ebd178e5a4eaaa2feb30fe325c4e35769036eb8b809a8dc535d443" exitCode=0 Dec 06 04:17:33 crc kubenswrapper[4802]: I1206 04:17:33.430071 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-psfqf" event={"ID":"45d96a62-52e1-4e87-9f5e-2ccbebc57fab","Type":"ContainerDied","Data":"779523f166ebd178e5a4eaaa2feb30fe325c4e35769036eb8b809a8dc535d443"} Dec 06 04:17:33 crc kubenswrapper[4802]: I1206 04:17:33.430432 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-psfqf" event={"ID":"45d96a62-52e1-4e87-9f5e-2ccbebc57fab","Type":"ContainerStarted","Data":"6797fa76bb0d4fbcc7a8b555d672bd41d956bcf1d87ac80e5591d19eecc53709"} Dec 06 04:17:34 crc kubenswrapper[4802]: I1206 04:17:34.446108 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-psfqf" event={"ID":"45d96a62-52e1-4e87-9f5e-2ccbebc57fab","Type":"ContainerStarted","Data":"f9da2a317a2f69036797530a457baf4951426c6598cedd22eebe93ae338dd44b"} Dec 06 04:17:35 crc kubenswrapper[4802]: I1206 04:17:35.459667 4802 generic.go:334] "Generic (PLEG): container finished" podID="45d96a62-52e1-4e87-9f5e-2ccbebc57fab" containerID="f9da2a317a2f69036797530a457baf4951426c6598cedd22eebe93ae338dd44b" exitCode=0 Dec 06 04:17:35 crc kubenswrapper[4802]: I1206 04:17:35.467555 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-psfqf" event={"ID":"45d96a62-52e1-4e87-9f5e-2ccbebc57fab","Type":"ContainerDied","Data":"f9da2a317a2f69036797530a457baf4951426c6598cedd22eebe93ae338dd44b"} Dec 06 04:17:36 crc kubenswrapper[4802]: I1206 04:17:36.469511 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-psfqf" event={"ID":"45d96a62-52e1-4e87-9f5e-2ccbebc57fab","Type":"ContainerStarted","Data":"1f2d3040f5c5744619340f81e2eae25306225679caa2ed1af61df402ff023f39"} Dec 06 04:17:36 crc kubenswrapper[4802]: I1206 04:17:36.493954 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-psfqf" podStartSLOduration=2.993971052 podStartE2EDuration="5.493935061s" podCreationTimestamp="2025-12-06 04:17:31 +0000 UTC" firstStartedPulling="2025-12-06 04:17:33.43444392 +0000 UTC m=+2246.306353072" lastFinishedPulling="2025-12-06 04:17:35.934407909 +0000 UTC m=+2248.806317081" observedRunningTime="2025-12-06 04:17:36.4843512 +0000 UTC m=+2249.356260352" watchObservedRunningTime="2025-12-06 04:17:36.493935061 +0000 UTC m=+2249.365844213" Dec 06 04:17:41 crc kubenswrapper[4802]: I1206 04:17:41.450326 4802 scope.go:117] "RemoveContainer" containerID="f8a17abbaa167592a628d9fdb3771d61462952618f6382d447c431c49f7c5690" Dec 06 04:17:41 crc kubenswrapper[4802]: E1206 04:17:41.451307 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:17:42 crc kubenswrapper[4802]: I1206 04:17:42.341098 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-psfqf" Dec 06 04:17:42 crc kubenswrapper[4802]: I1206 04:17:42.341476 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-psfqf" Dec 06 04:17:42 crc kubenswrapper[4802]: I1206 04:17:42.412739 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-psfqf" Dec 06 04:17:42 crc kubenswrapper[4802]: I1206 04:17:42.620896 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-psfqf" Dec 06 04:17:42 crc kubenswrapper[4802]: I1206 04:17:42.687975 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-psfqf"] Dec 06 04:17:44 crc kubenswrapper[4802]: I1206 04:17:44.557956 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-psfqf" podUID="45d96a62-52e1-4e87-9f5e-2ccbebc57fab" containerName="registry-server" containerID="cri-o://1f2d3040f5c5744619340f81e2eae25306225679caa2ed1af61df402ff023f39" gracePeriod=2 Dec 06 04:17:45 crc kubenswrapper[4802]: I1206 04:17:45.043791 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-psfqf" Dec 06 04:17:45 crc kubenswrapper[4802]: I1206 04:17:45.195104 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkffh\" (UniqueName: \"kubernetes.io/projected/45d96a62-52e1-4e87-9f5e-2ccbebc57fab-kube-api-access-zkffh\") pod \"45d96a62-52e1-4e87-9f5e-2ccbebc57fab\" (UID: \"45d96a62-52e1-4e87-9f5e-2ccbebc57fab\") " Dec 06 04:17:45 crc kubenswrapper[4802]: I1206 04:17:45.195217 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45d96a62-52e1-4e87-9f5e-2ccbebc57fab-utilities\") pod \"45d96a62-52e1-4e87-9f5e-2ccbebc57fab\" (UID: \"45d96a62-52e1-4e87-9f5e-2ccbebc57fab\") " Dec 06 04:17:45 crc kubenswrapper[4802]: I1206 04:17:45.195974 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45d96a62-52e1-4e87-9f5e-2ccbebc57fab-catalog-content\") pod \"45d96a62-52e1-4e87-9f5e-2ccbebc57fab\" (UID: \"45d96a62-52e1-4e87-9f5e-2ccbebc57fab\") " Dec 06 04:17:45 crc kubenswrapper[4802]: I1206 04:17:45.196174 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45d96a62-52e1-4e87-9f5e-2ccbebc57fab-utilities" (OuterVolumeSpecName: "utilities") pod "45d96a62-52e1-4e87-9f5e-2ccbebc57fab" (UID: "45d96a62-52e1-4e87-9f5e-2ccbebc57fab"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:17:45 crc kubenswrapper[4802]: I1206 04:17:45.196570 4802 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45d96a62-52e1-4e87-9f5e-2ccbebc57fab-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 04:17:45 crc kubenswrapper[4802]: I1206 04:17:45.201529 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45d96a62-52e1-4e87-9f5e-2ccbebc57fab-kube-api-access-zkffh" (OuterVolumeSpecName: "kube-api-access-zkffh") pod "45d96a62-52e1-4e87-9f5e-2ccbebc57fab" (UID: "45d96a62-52e1-4e87-9f5e-2ccbebc57fab"). InnerVolumeSpecName "kube-api-access-zkffh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:17:45 crc kubenswrapper[4802]: I1206 04:17:45.298107 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkffh\" (UniqueName: \"kubernetes.io/projected/45d96a62-52e1-4e87-9f5e-2ccbebc57fab-kube-api-access-zkffh\") on node \"crc\" DevicePath \"\"" Dec 06 04:17:45 crc kubenswrapper[4802]: I1206 04:17:45.344661 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45d96a62-52e1-4e87-9f5e-2ccbebc57fab-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "45d96a62-52e1-4e87-9f5e-2ccbebc57fab" (UID: "45d96a62-52e1-4e87-9f5e-2ccbebc57fab"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:17:45 crc kubenswrapper[4802]: I1206 04:17:45.400709 4802 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45d96a62-52e1-4e87-9f5e-2ccbebc57fab-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 04:17:45 crc kubenswrapper[4802]: I1206 04:17:45.569717 4802 generic.go:334] "Generic (PLEG): container finished" podID="45d96a62-52e1-4e87-9f5e-2ccbebc57fab" containerID="1f2d3040f5c5744619340f81e2eae25306225679caa2ed1af61df402ff023f39" exitCode=0 Dec 06 04:17:45 crc kubenswrapper[4802]: I1206 04:17:45.569783 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-psfqf" event={"ID":"45d96a62-52e1-4e87-9f5e-2ccbebc57fab","Type":"ContainerDied","Data":"1f2d3040f5c5744619340f81e2eae25306225679caa2ed1af61df402ff023f39"} Dec 06 04:17:45 crc kubenswrapper[4802]: I1206 04:17:45.569815 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-psfqf" event={"ID":"45d96a62-52e1-4e87-9f5e-2ccbebc57fab","Type":"ContainerDied","Data":"6797fa76bb0d4fbcc7a8b555d672bd41d956bcf1d87ac80e5591d19eecc53709"} Dec 06 04:17:45 crc kubenswrapper[4802]: I1206 04:17:45.569839 4802 scope.go:117] "RemoveContainer" containerID="1f2d3040f5c5744619340f81e2eae25306225679caa2ed1af61df402ff023f39" Dec 06 04:17:45 crc kubenswrapper[4802]: I1206 04:17:45.569974 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-psfqf" Dec 06 04:17:45 crc kubenswrapper[4802]: I1206 04:17:45.596671 4802 scope.go:117] "RemoveContainer" containerID="f9da2a317a2f69036797530a457baf4951426c6598cedd22eebe93ae338dd44b" Dec 06 04:17:45 crc kubenswrapper[4802]: I1206 04:17:45.601798 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-psfqf"] Dec 06 04:17:45 crc kubenswrapper[4802]: I1206 04:17:45.621922 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-psfqf"] Dec 06 04:17:45 crc kubenswrapper[4802]: I1206 04:17:45.652481 4802 scope.go:117] "RemoveContainer" containerID="779523f166ebd178e5a4eaaa2feb30fe325c4e35769036eb8b809a8dc535d443" Dec 06 04:17:45 crc kubenswrapper[4802]: I1206 04:17:45.686708 4802 scope.go:117] "RemoveContainer" containerID="1f2d3040f5c5744619340f81e2eae25306225679caa2ed1af61df402ff023f39" Dec 06 04:17:45 crc kubenswrapper[4802]: E1206 04:17:45.687397 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f2d3040f5c5744619340f81e2eae25306225679caa2ed1af61df402ff023f39\": container with ID starting with 1f2d3040f5c5744619340f81e2eae25306225679caa2ed1af61df402ff023f39 not found: ID does not exist" containerID="1f2d3040f5c5744619340f81e2eae25306225679caa2ed1af61df402ff023f39" Dec 06 04:17:45 crc kubenswrapper[4802]: I1206 04:17:45.687450 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f2d3040f5c5744619340f81e2eae25306225679caa2ed1af61df402ff023f39"} err="failed to get container status \"1f2d3040f5c5744619340f81e2eae25306225679caa2ed1af61df402ff023f39\": rpc error: code = NotFound desc = could not find container \"1f2d3040f5c5744619340f81e2eae25306225679caa2ed1af61df402ff023f39\": container with ID starting with 1f2d3040f5c5744619340f81e2eae25306225679caa2ed1af61df402ff023f39 not found: ID does not exist" Dec 06 04:17:45 crc kubenswrapper[4802]: I1206 04:17:45.687480 4802 scope.go:117] "RemoveContainer" containerID="f9da2a317a2f69036797530a457baf4951426c6598cedd22eebe93ae338dd44b" Dec 06 04:17:45 crc kubenswrapper[4802]: E1206 04:17:45.688264 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f9da2a317a2f69036797530a457baf4951426c6598cedd22eebe93ae338dd44b\": container with ID starting with f9da2a317a2f69036797530a457baf4951426c6598cedd22eebe93ae338dd44b not found: ID does not exist" containerID="f9da2a317a2f69036797530a457baf4951426c6598cedd22eebe93ae338dd44b" Dec 06 04:17:45 crc kubenswrapper[4802]: I1206 04:17:45.688286 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9da2a317a2f69036797530a457baf4951426c6598cedd22eebe93ae338dd44b"} err="failed to get container status \"f9da2a317a2f69036797530a457baf4951426c6598cedd22eebe93ae338dd44b\": rpc error: code = NotFound desc = could not find container \"f9da2a317a2f69036797530a457baf4951426c6598cedd22eebe93ae338dd44b\": container with ID starting with f9da2a317a2f69036797530a457baf4951426c6598cedd22eebe93ae338dd44b not found: ID does not exist" Dec 06 04:17:45 crc kubenswrapper[4802]: I1206 04:17:45.688302 4802 scope.go:117] "RemoveContainer" containerID="779523f166ebd178e5a4eaaa2feb30fe325c4e35769036eb8b809a8dc535d443" Dec 06 04:17:45 crc kubenswrapper[4802]: E1206 04:17:45.688631 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"779523f166ebd178e5a4eaaa2feb30fe325c4e35769036eb8b809a8dc535d443\": container with ID starting with 779523f166ebd178e5a4eaaa2feb30fe325c4e35769036eb8b809a8dc535d443 not found: ID does not exist" containerID="779523f166ebd178e5a4eaaa2feb30fe325c4e35769036eb8b809a8dc535d443" Dec 06 04:17:45 crc kubenswrapper[4802]: I1206 04:17:45.688708 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"779523f166ebd178e5a4eaaa2feb30fe325c4e35769036eb8b809a8dc535d443"} err="failed to get container status \"779523f166ebd178e5a4eaaa2feb30fe325c4e35769036eb8b809a8dc535d443\": rpc error: code = NotFound desc = could not find container \"779523f166ebd178e5a4eaaa2feb30fe325c4e35769036eb8b809a8dc535d443\": container with ID starting with 779523f166ebd178e5a4eaaa2feb30fe325c4e35769036eb8b809a8dc535d443 not found: ID does not exist" Dec 06 04:17:47 crc kubenswrapper[4802]: I1206 04:17:47.459894 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45d96a62-52e1-4e87-9f5e-2ccbebc57fab" path="/var/lib/kubelet/pods/45d96a62-52e1-4e87-9f5e-2ccbebc57fab/volumes" Dec 06 04:17:52 crc kubenswrapper[4802]: I1206 04:17:52.450994 4802 scope.go:117] "RemoveContainer" containerID="f8a17abbaa167592a628d9fdb3771d61462952618f6382d447c431c49f7c5690" Dec 06 04:17:52 crc kubenswrapper[4802]: E1206 04:17:52.451810 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:18:04 crc kubenswrapper[4802]: I1206 04:18:04.450406 4802 scope.go:117] "RemoveContainer" containerID="f8a17abbaa167592a628d9fdb3771d61462952618f6382d447c431c49f7c5690" Dec 06 04:18:04 crc kubenswrapper[4802]: E1206 04:18:04.451290 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:18:19 crc kubenswrapper[4802]: I1206 04:18:19.451776 4802 scope.go:117] "RemoveContainer" containerID="f8a17abbaa167592a628d9fdb3771d61462952618f6382d447c431c49f7c5690" Dec 06 04:18:19 crc kubenswrapper[4802]: E1206 04:18:19.453167 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:18:30 crc kubenswrapper[4802]: I1206 04:18:30.450211 4802 scope.go:117] "RemoveContainer" containerID="f8a17abbaa167592a628d9fdb3771d61462952618f6382d447c431c49f7c5690" Dec 06 04:18:30 crc kubenswrapper[4802]: E1206 04:18:30.451316 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:18:41 crc kubenswrapper[4802]: I1206 04:18:41.450527 4802 scope.go:117] "RemoveContainer" containerID="f8a17abbaa167592a628d9fdb3771d61462952618f6382d447c431c49f7c5690" Dec 06 04:18:41 crc kubenswrapper[4802]: E1206 04:18:41.451515 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:18:55 crc kubenswrapper[4802]: I1206 04:18:55.450102 4802 scope.go:117] "RemoveContainer" containerID="f8a17abbaa167592a628d9fdb3771d61462952618f6382d447c431c49f7c5690" Dec 06 04:18:55 crc kubenswrapper[4802]: E1206 04:18:55.450951 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:19:08 crc kubenswrapper[4802]: I1206 04:19:08.450916 4802 scope.go:117] "RemoveContainer" containerID="f8a17abbaa167592a628d9fdb3771d61462952618f6382d447c431c49f7c5690" Dec 06 04:19:08 crc kubenswrapper[4802]: E1206 04:19:08.451670 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:19:20 crc kubenswrapper[4802]: I1206 04:19:20.450176 4802 scope.go:117] "RemoveContainer" containerID="f8a17abbaa167592a628d9fdb3771d61462952618f6382d447c431c49f7c5690" Dec 06 04:19:20 crc kubenswrapper[4802]: E1206 04:19:20.450790 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:19:32 crc kubenswrapper[4802]: I1206 04:19:32.450799 4802 scope.go:117] "RemoveContainer" containerID="f8a17abbaa167592a628d9fdb3771d61462952618f6382d447c431c49f7c5690" Dec 06 04:19:32 crc kubenswrapper[4802]: E1206 04:19:32.451715 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:19:47 crc kubenswrapper[4802]: I1206 04:19:47.458190 4802 scope.go:117] "RemoveContainer" containerID="f8a17abbaa167592a628d9fdb3771d61462952618f6382d447c431c49f7c5690" Dec 06 04:19:47 crc kubenswrapper[4802]: E1206 04:19:47.459009 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:19:59 crc kubenswrapper[4802]: I1206 04:19:59.450722 4802 scope.go:117] "RemoveContainer" containerID="f8a17abbaa167592a628d9fdb3771d61462952618f6382d447c431c49f7c5690" Dec 06 04:19:59 crc kubenswrapper[4802]: E1206 04:19:59.451407 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:20:14 crc kubenswrapper[4802]: I1206 04:20:14.451031 4802 scope.go:117] "RemoveContainer" containerID="f8a17abbaa167592a628d9fdb3771d61462952618f6382d447c431c49f7c5690" Dec 06 04:20:14 crc kubenswrapper[4802]: E1206 04:20:14.452047 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:20:26 crc kubenswrapper[4802]: I1206 04:20:26.450017 4802 scope.go:117] "RemoveContainer" containerID="f8a17abbaa167592a628d9fdb3771d61462952618f6382d447c431c49f7c5690" Dec 06 04:20:26 crc kubenswrapper[4802]: E1206 04:20:26.450720 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:20:38 crc kubenswrapper[4802]: I1206 04:20:38.450167 4802 scope.go:117] "RemoveContainer" containerID="f8a17abbaa167592a628d9fdb3771d61462952618f6382d447c431c49f7c5690" Dec 06 04:20:38 crc kubenswrapper[4802]: E1206 04:20:38.451188 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:20:39 crc kubenswrapper[4802]: I1206 04:20:39.271435 4802 generic.go:334] "Generic (PLEG): container finished" podID="18179f2f-334f-48c2-a6be-6d1c1792e600" containerID="c4b2ce0306615d6526357d3ee8392af4a43bf9e7c8579cbaf02c2c03100e3ab4" exitCode=0 Dec 06 04:20:39 crc kubenswrapper[4802]: I1206 04:20:39.271720 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tckw2" event={"ID":"18179f2f-334f-48c2-a6be-6d1c1792e600","Type":"ContainerDied","Data":"c4b2ce0306615d6526357d3ee8392af4a43bf9e7c8579cbaf02c2c03100e3ab4"} Dec 06 04:20:40 crc kubenswrapper[4802]: I1206 04:20:40.684995 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tckw2" Dec 06 04:20:40 crc kubenswrapper[4802]: I1206 04:20:40.837611 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18179f2f-334f-48c2-a6be-6d1c1792e600-inventory\") pod \"18179f2f-334f-48c2-a6be-6d1c1792e600\" (UID: \"18179f2f-334f-48c2-a6be-6d1c1792e600\") " Dec 06 04:20:40 crc kubenswrapper[4802]: I1206 04:20:40.837869 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18179f2f-334f-48c2-a6be-6d1c1792e600-libvirt-combined-ca-bundle\") pod \"18179f2f-334f-48c2-a6be-6d1c1792e600\" (UID: \"18179f2f-334f-48c2-a6be-6d1c1792e600\") " Dec 06 04:20:40 crc kubenswrapper[4802]: I1206 04:20:40.837900 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f4j6b\" (UniqueName: \"kubernetes.io/projected/18179f2f-334f-48c2-a6be-6d1c1792e600-kube-api-access-f4j6b\") pod \"18179f2f-334f-48c2-a6be-6d1c1792e600\" (UID: \"18179f2f-334f-48c2-a6be-6d1c1792e600\") " Dec 06 04:20:40 crc kubenswrapper[4802]: I1206 04:20:40.837965 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/18179f2f-334f-48c2-a6be-6d1c1792e600-ssh-key\") pod \"18179f2f-334f-48c2-a6be-6d1c1792e600\" (UID: \"18179f2f-334f-48c2-a6be-6d1c1792e600\") " Dec 06 04:20:40 crc kubenswrapper[4802]: I1206 04:20:40.838570 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/18179f2f-334f-48c2-a6be-6d1c1792e600-libvirt-secret-0\") pod \"18179f2f-334f-48c2-a6be-6d1c1792e600\" (UID: \"18179f2f-334f-48c2-a6be-6d1c1792e600\") " Dec 06 04:20:40 crc kubenswrapper[4802]: I1206 04:20:40.843218 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18179f2f-334f-48c2-a6be-6d1c1792e600-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "18179f2f-334f-48c2-a6be-6d1c1792e600" (UID: "18179f2f-334f-48c2-a6be-6d1c1792e600"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:20:40 crc kubenswrapper[4802]: I1206 04:20:40.843905 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18179f2f-334f-48c2-a6be-6d1c1792e600-kube-api-access-f4j6b" (OuterVolumeSpecName: "kube-api-access-f4j6b") pod "18179f2f-334f-48c2-a6be-6d1c1792e600" (UID: "18179f2f-334f-48c2-a6be-6d1c1792e600"). InnerVolumeSpecName "kube-api-access-f4j6b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:20:40 crc kubenswrapper[4802]: I1206 04:20:40.864235 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18179f2f-334f-48c2-a6be-6d1c1792e600-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "18179f2f-334f-48c2-a6be-6d1c1792e600" (UID: "18179f2f-334f-48c2-a6be-6d1c1792e600"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:20:40 crc kubenswrapper[4802]: I1206 04:20:40.864511 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18179f2f-334f-48c2-a6be-6d1c1792e600-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "18179f2f-334f-48c2-a6be-6d1c1792e600" (UID: "18179f2f-334f-48c2-a6be-6d1c1792e600"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:20:40 crc kubenswrapper[4802]: I1206 04:20:40.870329 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18179f2f-334f-48c2-a6be-6d1c1792e600-inventory" (OuterVolumeSpecName: "inventory") pod "18179f2f-334f-48c2-a6be-6d1c1792e600" (UID: "18179f2f-334f-48c2-a6be-6d1c1792e600"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:20:40 crc kubenswrapper[4802]: I1206 04:20:40.940528 4802 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/18179f2f-334f-48c2-a6be-6d1c1792e600-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 04:20:40 crc kubenswrapper[4802]: I1206 04:20:40.940728 4802 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/18179f2f-334f-48c2-a6be-6d1c1792e600-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Dec 06 04:20:40 crc kubenswrapper[4802]: I1206 04:20:40.940822 4802 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18179f2f-334f-48c2-a6be-6d1c1792e600-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 04:20:40 crc kubenswrapper[4802]: I1206 04:20:40.940909 4802 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18179f2f-334f-48c2-a6be-6d1c1792e600-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:20:40 crc kubenswrapper[4802]: I1206 04:20:40.940965 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f4j6b\" (UniqueName: \"kubernetes.io/projected/18179f2f-334f-48c2-a6be-6d1c1792e600-kube-api-access-f4j6b\") on node \"crc\" DevicePath \"\"" Dec 06 04:20:41 crc kubenswrapper[4802]: I1206 04:20:41.289624 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tckw2" event={"ID":"18179f2f-334f-48c2-a6be-6d1c1792e600","Type":"ContainerDied","Data":"e07fc280ce32e557c76f79d7eefeceb843334dfea6d5e0a0a40571cfb091499c"} Dec 06 04:20:41 crc kubenswrapper[4802]: I1206 04:20:41.289975 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e07fc280ce32e557c76f79d7eefeceb843334dfea6d5e0a0a40571cfb091499c" Dec 06 04:20:41 crc kubenswrapper[4802]: I1206 04:20:41.289668 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tckw2" Dec 06 04:20:41 crc kubenswrapper[4802]: I1206 04:20:41.392688 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-5plkw"] Dec 06 04:20:41 crc kubenswrapper[4802]: E1206 04:20:41.393159 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45d96a62-52e1-4e87-9f5e-2ccbebc57fab" containerName="extract-content" Dec 06 04:20:41 crc kubenswrapper[4802]: I1206 04:20:41.393182 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="45d96a62-52e1-4e87-9f5e-2ccbebc57fab" containerName="extract-content" Dec 06 04:20:41 crc kubenswrapper[4802]: E1206 04:20:41.393202 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18179f2f-334f-48c2-a6be-6d1c1792e600" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 06 04:20:41 crc kubenswrapper[4802]: I1206 04:20:41.393211 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="18179f2f-334f-48c2-a6be-6d1c1792e600" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 06 04:20:41 crc kubenswrapper[4802]: E1206 04:20:41.393249 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45d96a62-52e1-4e87-9f5e-2ccbebc57fab" containerName="extract-utilities" Dec 06 04:20:41 crc kubenswrapper[4802]: I1206 04:20:41.393256 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="45d96a62-52e1-4e87-9f5e-2ccbebc57fab" containerName="extract-utilities" Dec 06 04:20:41 crc kubenswrapper[4802]: E1206 04:20:41.393273 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45d96a62-52e1-4e87-9f5e-2ccbebc57fab" containerName="registry-server" Dec 06 04:20:41 crc kubenswrapper[4802]: I1206 04:20:41.393280 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="45d96a62-52e1-4e87-9f5e-2ccbebc57fab" containerName="registry-server" Dec 06 04:20:41 crc kubenswrapper[4802]: I1206 04:20:41.393510 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="45d96a62-52e1-4e87-9f5e-2ccbebc57fab" containerName="registry-server" Dec 06 04:20:41 crc kubenswrapper[4802]: I1206 04:20:41.393534 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="18179f2f-334f-48c2-a6be-6d1c1792e600" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 06 04:20:41 crc kubenswrapper[4802]: I1206 04:20:41.394361 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5plkw" Dec 06 04:20:41 crc kubenswrapper[4802]: I1206 04:20:41.396708 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-rgvjc" Dec 06 04:20:41 crc kubenswrapper[4802]: I1206 04:20:41.401117 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 04:20:41 crc kubenswrapper[4802]: I1206 04:20:41.401265 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Dec 06 04:20:41 crc kubenswrapper[4802]: I1206 04:20:41.401366 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 04:20:41 crc kubenswrapper[4802]: I1206 04:20:41.401534 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 06 04:20:41 crc kubenswrapper[4802]: I1206 04:20:41.401599 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 06 04:20:41 crc kubenswrapper[4802]: I1206 04:20:41.401719 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 04:20:41 crc kubenswrapper[4802]: I1206 04:20:41.424871 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-5plkw"] Dec 06 04:20:41 crc kubenswrapper[4802]: I1206 04:20:41.554063 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/b8e8984b-b60b-48e5-9586-1a1a8d883143-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5plkw\" (UID: \"b8e8984b-b60b-48e5-9586-1a1a8d883143\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5plkw" Dec 06 04:20:41 crc kubenswrapper[4802]: I1206 04:20:41.554117 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/b8e8984b-b60b-48e5-9586-1a1a8d883143-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5plkw\" (UID: \"b8e8984b-b60b-48e5-9586-1a1a8d883143\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5plkw" Dec 06 04:20:41 crc kubenswrapper[4802]: I1206 04:20:41.554147 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8e8984b-b60b-48e5-9586-1a1a8d883143-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5plkw\" (UID: \"b8e8984b-b60b-48e5-9586-1a1a8d883143\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5plkw" Dec 06 04:20:41 crc kubenswrapper[4802]: I1206 04:20:41.554189 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/b8e8984b-b60b-48e5-9586-1a1a8d883143-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5plkw\" (UID: \"b8e8984b-b60b-48e5-9586-1a1a8d883143\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5plkw" Dec 06 04:20:41 crc kubenswrapper[4802]: I1206 04:20:41.554215 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjsh2\" (UniqueName: \"kubernetes.io/projected/b8e8984b-b60b-48e5-9586-1a1a8d883143-kube-api-access-hjsh2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5plkw\" (UID: \"b8e8984b-b60b-48e5-9586-1a1a8d883143\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5plkw" Dec 06 04:20:41 crc kubenswrapper[4802]: I1206 04:20:41.554367 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/b8e8984b-b60b-48e5-9586-1a1a8d883143-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5plkw\" (UID: \"b8e8984b-b60b-48e5-9586-1a1a8d883143\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5plkw" Dec 06 04:20:41 crc kubenswrapper[4802]: I1206 04:20:41.554623 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b8e8984b-b60b-48e5-9586-1a1a8d883143-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5plkw\" (UID: \"b8e8984b-b60b-48e5-9586-1a1a8d883143\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5plkw" Dec 06 04:20:41 crc kubenswrapper[4802]: I1206 04:20:41.555497 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/b8e8984b-b60b-48e5-9586-1a1a8d883143-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5plkw\" (UID: \"b8e8984b-b60b-48e5-9586-1a1a8d883143\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5plkw" Dec 06 04:20:41 crc kubenswrapper[4802]: I1206 04:20:41.555583 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8e8984b-b60b-48e5-9586-1a1a8d883143-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5plkw\" (UID: \"b8e8984b-b60b-48e5-9586-1a1a8d883143\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5plkw" Dec 06 04:20:41 crc kubenswrapper[4802]: I1206 04:20:41.658649 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b8e8984b-b60b-48e5-9586-1a1a8d883143-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5plkw\" (UID: \"b8e8984b-b60b-48e5-9586-1a1a8d883143\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5plkw" Dec 06 04:20:41 crc kubenswrapper[4802]: I1206 04:20:41.658954 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/b8e8984b-b60b-48e5-9586-1a1a8d883143-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5plkw\" (UID: \"b8e8984b-b60b-48e5-9586-1a1a8d883143\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5plkw" Dec 06 04:20:41 crc kubenswrapper[4802]: I1206 04:20:41.659047 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8e8984b-b60b-48e5-9586-1a1a8d883143-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5plkw\" (UID: \"b8e8984b-b60b-48e5-9586-1a1a8d883143\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5plkw" Dec 06 04:20:41 crc kubenswrapper[4802]: I1206 04:20:41.659163 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/b8e8984b-b60b-48e5-9586-1a1a8d883143-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5plkw\" (UID: \"b8e8984b-b60b-48e5-9586-1a1a8d883143\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5plkw" Dec 06 04:20:41 crc kubenswrapper[4802]: I1206 04:20:41.659228 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/b8e8984b-b60b-48e5-9586-1a1a8d883143-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5plkw\" (UID: \"b8e8984b-b60b-48e5-9586-1a1a8d883143\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5plkw" Dec 06 04:20:41 crc kubenswrapper[4802]: I1206 04:20:41.659289 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8e8984b-b60b-48e5-9586-1a1a8d883143-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5plkw\" (UID: \"b8e8984b-b60b-48e5-9586-1a1a8d883143\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5plkw" Dec 06 04:20:41 crc kubenswrapper[4802]: I1206 04:20:41.659366 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/b8e8984b-b60b-48e5-9586-1a1a8d883143-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5plkw\" (UID: \"b8e8984b-b60b-48e5-9586-1a1a8d883143\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5plkw" Dec 06 04:20:41 crc kubenswrapper[4802]: I1206 04:20:41.659433 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjsh2\" (UniqueName: \"kubernetes.io/projected/b8e8984b-b60b-48e5-9586-1a1a8d883143-kube-api-access-hjsh2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5plkw\" (UID: \"b8e8984b-b60b-48e5-9586-1a1a8d883143\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5plkw" Dec 06 04:20:41 crc kubenswrapper[4802]: I1206 04:20:41.659504 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/b8e8984b-b60b-48e5-9586-1a1a8d883143-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5plkw\" (UID: \"b8e8984b-b60b-48e5-9586-1a1a8d883143\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5plkw" Dec 06 04:20:41 crc kubenswrapper[4802]: I1206 04:20:41.661667 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/b8e8984b-b60b-48e5-9586-1a1a8d883143-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5plkw\" (UID: \"b8e8984b-b60b-48e5-9586-1a1a8d883143\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5plkw" Dec 06 04:20:41 crc kubenswrapper[4802]: I1206 04:20:41.664302 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/b8e8984b-b60b-48e5-9586-1a1a8d883143-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5plkw\" (UID: \"b8e8984b-b60b-48e5-9586-1a1a8d883143\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5plkw" Dec 06 04:20:41 crc kubenswrapper[4802]: I1206 04:20:41.664494 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8e8984b-b60b-48e5-9586-1a1a8d883143-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5plkw\" (UID: \"b8e8984b-b60b-48e5-9586-1a1a8d883143\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5plkw" Dec 06 04:20:41 crc kubenswrapper[4802]: I1206 04:20:41.664617 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b8e8984b-b60b-48e5-9586-1a1a8d883143-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5plkw\" (UID: \"b8e8984b-b60b-48e5-9586-1a1a8d883143\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5plkw" Dec 06 04:20:41 crc kubenswrapper[4802]: I1206 04:20:41.664658 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8e8984b-b60b-48e5-9586-1a1a8d883143-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5plkw\" (UID: \"b8e8984b-b60b-48e5-9586-1a1a8d883143\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5plkw" Dec 06 04:20:41 crc kubenswrapper[4802]: I1206 04:20:41.664961 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/b8e8984b-b60b-48e5-9586-1a1a8d883143-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5plkw\" (UID: \"b8e8984b-b60b-48e5-9586-1a1a8d883143\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5plkw" Dec 06 04:20:41 crc kubenswrapper[4802]: I1206 04:20:41.669347 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/b8e8984b-b60b-48e5-9586-1a1a8d883143-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5plkw\" (UID: \"b8e8984b-b60b-48e5-9586-1a1a8d883143\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5plkw" Dec 06 04:20:41 crc kubenswrapper[4802]: I1206 04:20:41.669939 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/b8e8984b-b60b-48e5-9586-1a1a8d883143-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5plkw\" (UID: \"b8e8984b-b60b-48e5-9586-1a1a8d883143\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5plkw" Dec 06 04:20:41 crc kubenswrapper[4802]: I1206 04:20:41.686294 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjsh2\" (UniqueName: \"kubernetes.io/projected/b8e8984b-b60b-48e5-9586-1a1a8d883143-kube-api-access-hjsh2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5plkw\" (UID: \"b8e8984b-b60b-48e5-9586-1a1a8d883143\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5plkw" Dec 06 04:20:41 crc kubenswrapper[4802]: I1206 04:20:41.713418 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5plkw" Dec 06 04:20:42 crc kubenswrapper[4802]: I1206 04:20:42.231342 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-5plkw"] Dec 06 04:20:42 crc kubenswrapper[4802]: I1206 04:20:42.298509 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5plkw" event={"ID":"b8e8984b-b60b-48e5-9586-1a1a8d883143","Type":"ContainerStarted","Data":"f38b3c2d3cfc059e92d01f68369fcf02c0115261b1d1197b4b12dbeae10989bb"} Dec 06 04:20:43 crc kubenswrapper[4802]: I1206 04:20:43.309239 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5plkw" event={"ID":"b8e8984b-b60b-48e5-9586-1a1a8d883143","Type":"ContainerStarted","Data":"03134fb4d2e174c7247dee7d8a7029a0156ead392c9c09acb27ebedd57f12c93"} Dec 06 04:20:43 crc kubenswrapper[4802]: I1206 04:20:43.327007 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5plkw" podStartSLOduration=1.897810338 podStartE2EDuration="2.326985541s" podCreationTimestamp="2025-12-06 04:20:41 +0000 UTC" firstStartedPulling="2025-12-06 04:20:42.233732312 +0000 UTC m=+2435.105641464" lastFinishedPulling="2025-12-06 04:20:42.662907515 +0000 UTC m=+2435.534816667" observedRunningTime="2025-12-06 04:20:43.324818401 +0000 UTC m=+2436.196727563" watchObservedRunningTime="2025-12-06 04:20:43.326985541 +0000 UTC m=+2436.198894693" Dec 06 04:20:51 crc kubenswrapper[4802]: I1206 04:20:51.451303 4802 scope.go:117] "RemoveContainer" containerID="f8a17abbaa167592a628d9fdb3771d61462952618f6382d447c431c49f7c5690" Dec 06 04:20:51 crc kubenswrapper[4802]: E1206 04:20:51.452323 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:21:02 crc kubenswrapper[4802]: I1206 04:21:02.450502 4802 scope.go:117] "RemoveContainer" containerID="f8a17abbaa167592a628d9fdb3771d61462952618f6382d447c431c49f7c5690" Dec 06 04:21:02 crc kubenswrapper[4802]: E1206 04:21:02.451126 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:21:13 crc kubenswrapper[4802]: I1206 04:21:13.450364 4802 scope.go:117] "RemoveContainer" containerID="f8a17abbaa167592a628d9fdb3771d61462952618f6382d447c431c49f7c5690" Dec 06 04:21:13 crc kubenswrapper[4802]: E1206 04:21:13.451273 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:21:24 crc kubenswrapper[4802]: I1206 04:21:24.450334 4802 scope.go:117] "RemoveContainer" containerID="f8a17abbaa167592a628d9fdb3771d61462952618f6382d447c431c49f7c5690" Dec 06 04:21:24 crc kubenswrapper[4802]: E1206 04:21:24.451331 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:21:37 crc kubenswrapper[4802]: I1206 04:21:37.450444 4802 scope.go:117] "RemoveContainer" containerID="f8a17abbaa167592a628d9fdb3771d61462952618f6382d447c431c49f7c5690" Dec 06 04:21:37 crc kubenswrapper[4802]: E1206 04:21:37.451382 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:21:48 crc kubenswrapper[4802]: I1206 04:21:48.450027 4802 scope.go:117] "RemoveContainer" containerID="f8a17abbaa167592a628d9fdb3771d61462952618f6382d447c431c49f7c5690" Dec 06 04:21:48 crc kubenswrapper[4802]: E1206 04:21:48.450805 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:22:02 crc kubenswrapper[4802]: I1206 04:22:02.451090 4802 scope.go:117] "RemoveContainer" containerID="f8a17abbaa167592a628d9fdb3771d61462952618f6382d447c431c49f7c5690" Dec 06 04:22:02 crc kubenswrapper[4802]: E1206 04:22:02.451946 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:22:16 crc kubenswrapper[4802]: I1206 04:22:16.451251 4802 scope.go:117] "RemoveContainer" containerID="f8a17abbaa167592a628d9fdb3771d61462952618f6382d447c431c49f7c5690" Dec 06 04:22:18 crc kubenswrapper[4802]: I1206 04:22:18.224501 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" event={"ID":"6df38316-e0d3-4018-8d27-3620eba3a68d","Type":"ContainerStarted","Data":"77efcf0d5716fddf110919e082917d423162e98a637ed45489f128e2cf499066"} Dec 06 04:23:25 crc kubenswrapper[4802]: I1206 04:23:25.902325 4802 generic.go:334] "Generic (PLEG): container finished" podID="b8e8984b-b60b-48e5-9586-1a1a8d883143" containerID="03134fb4d2e174c7247dee7d8a7029a0156ead392c9c09acb27ebedd57f12c93" exitCode=0 Dec 06 04:23:25 crc kubenswrapper[4802]: I1206 04:23:25.902425 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5plkw" event={"ID":"b8e8984b-b60b-48e5-9586-1a1a8d883143","Type":"ContainerDied","Data":"03134fb4d2e174c7247dee7d8a7029a0156ead392c9c09acb27ebedd57f12c93"} Dec 06 04:23:27 crc kubenswrapper[4802]: I1206 04:23:27.456547 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5plkw" Dec 06 04:23:27 crc kubenswrapper[4802]: I1206 04:23:27.556654 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/b8e8984b-b60b-48e5-9586-1a1a8d883143-nova-cell1-compute-config-0\") pod \"b8e8984b-b60b-48e5-9586-1a1a8d883143\" (UID: \"b8e8984b-b60b-48e5-9586-1a1a8d883143\") " Dec 06 04:23:27 crc kubenswrapper[4802]: I1206 04:23:27.556691 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/b8e8984b-b60b-48e5-9586-1a1a8d883143-nova-extra-config-0\") pod \"b8e8984b-b60b-48e5-9586-1a1a8d883143\" (UID: \"b8e8984b-b60b-48e5-9586-1a1a8d883143\") " Dec 06 04:23:27 crc kubenswrapper[4802]: I1206 04:23:27.556717 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjsh2\" (UniqueName: \"kubernetes.io/projected/b8e8984b-b60b-48e5-9586-1a1a8d883143-kube-api-access-hjsh2\") pod \"b8e8984b-b60b-48e5-9586-1a1a8d883143\" (UID: \"b8e8984b-b60b-48e5-9586-1a1a8d883143\") " Dec 06 04:23:27 crc kubenswrapper[4802]: I1206 04:23:27.556809 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8e8984b-b60b-48e5-9586-1a1a8d883143-inventory\") pod \"b8e8984b-b60b-48e5-9586-1a1a8d883143\" (UID: \"b8e8984b-b60b-48e5-9586-1a1a8d883143\") " Dec 06 04:23:27 crc kubenswrapper[4802]: I1206 04:23:27.556876 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b8e8984b-b60b-48e5-9586-1a1a8d883143-ssh-key\") pod \"b8e8984b-b60b-48e5-9586-1a1a8d883143\" (UID: \"b8e8984b-b60b-48e5-9586-1a1a8d883143\") " Dec 06 04:23:27 crc kubenswrapper[4802]: I1206 04:23:27.556929 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/b8e8984b-b60b-48e5-9586-1a1a8d883143-nova-migration-ssh-key-1\") pod \"b8e8984b-b60b-48e5-9586-1a1a8d883143\" (UID: \"b8e8984b-b60b-48e5-9586-1a1a8d883143\") " Dec 06 04:23:27 crc kubenswrapper[4802]: I1206 04:23:27.556959 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8e8984b-b60b-48e5-9586-1a1a8d883143-nova-combined-ca-bundle\") pod \"b8e8984b-b60b-48e5-9586-1a1a8d883143\" (UID: \"b8e8984b-b60b-48e5-9586-1a1a8d883143\") " Dec 06 04:23:27 crc kubenswrapper[4802]: I1206 04:23:27.557003 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/b8e8984b-b60b-48e5-9586-1a1a8d883143-nova-cell1-compute-config-1\") pod \"b8e8984b-b60b-48e5-9586-1a1a8d883143\" (UID: \"b8e8984b-b60b-48e5-9586-1a1a8d883143\") " Dec 06 04:23:27 crc kubenswrapper[4802]: I1206 04:23:27.557025 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/b8e8984b-b60b-48e5-9586-1a1a8d883143-nova-migration-ssh-key-0\") pod \"b8e8984b-b60b-48e5-9586-1a1a8d883143\" (UID: \"b8e8984b-b60b-48e5-9586-1a1a8d883143\") " Dec 06 04:23:27 crc kubenswrapper[4802]: I1206 04:23:27.563053 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8e8984b-b60b-48e5-9586-1a1a8d883143-kube-api-access-hjsh2" (OuterVolumeSpecName: "kube-api-access-hjsh2") pod "b8e8984b-b60b-48e5-9586-1a1a8d883143" (UID: "b8e8984b-b60b-48e5-9586-1a1a8d883143"). InnerVolumeSpecName "kube-api-access-hjsh2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:23:27 crc kubenswrapper[4802]: I1206 04:23:27.569151 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8e8984b-b60b-48e5-9586-1a1a8d883143-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "b8e8984b-b60b-48e5-9586-1a1a8d883143" (UID: "b8e8984b-b60b-48e5-9586-1a1a8d883143"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:23:27 crc kubenswrapper[4802]: I1206 04:23:27.585863 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8e8984b-b60b-48e5-9586-1a1a8d883143-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "b8e8984b-b60b-48e5-9586-1a1a8d883143" (UID: "b8e8984b-b60b-48e5-9586-1a1a8d883143"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:23:27 crc kubenswrapper[4802]: I1206 04:23:27.586602 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8e8984b-b60b-48e5-9586-1a1a8d883143-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "b8e8984b-b60b-48e5-9586-1a1a8d883143" (UID: "b8e8984b-b60b-48e5-9586-1a1a8d883143"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:23:27 crc kubenswrapper[4802]: I1206 04:23:27.591022 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b8e8984b-b60b-48e5-9586-1a1a8d883143-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "b8e8984b-b60b-48e5-9586-1a1a8d883143" (UID: "b8e8984b-b60b-48e5-9586-1a1a8d883143"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:23:27 crc kubenswrapper[4802]: I1206 04:23:27.592866 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8e8984b-b60b-48e5-9586-1a1a8d883143-inventory" (OuterVolumeSpecName: "inventory") pod "b8e8984b-b60b-48e5-9586-1a1a8d883143" (UID: "b8e8984b-b60b-48e5-9586-1a1a8d883143"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:23:27 crc kubenswrapper[4802]: I1206 04:23:27.594934 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8e8984b-b60b-48e5-9586-1a1a8d883143-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b8e8984b-b60b-48e5-9586-1a1a8d883143" (UID: "b8e8984b-b60b-48e5-9586-1a1a8d883143"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:23:27 crc kubenswrapper[4802]: I1206 04:23:27.596579 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8e8984b-b60b-48e5-9586-1a1a8d883143-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "b8e8984b-b60b-48e5-9586-1a1a8d883143" (UID: "b8e8984b-b60b-48e5-9586-1a1a8d883143"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:23:27 crc kubenswrapper[4802]: I1206 04:23:27.610916 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8e8984b-b60b-48e5-9586-1a1a8d883143-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "b8e8984b-b60b-48e5-9586-1a1a8d883143" (UID: "b8e8984b-b60b-48e5-9586-1a1a8d883143"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:23:27 crc kubenswrapper[4802]: I1206 04:23:27.658231 4802 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/b8e8984b-b60b-48e5-9586-1a1a8d883143-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 06 04:23:27 crc kubenswrapper[4802]: I1206 04:23:27.658265 4802 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/b8e8984b-b60b-48e5-9586-1a1a8d883143-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Dec 06 04:23:27 crc kubenswrapper[4802]: I1206 04:23:27.658277 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjsh2\" (UniqueName: \"kubernetes.io/projected/b8e8984b-b60b-48e5-9586-1a1a8d883143-kube-api-access-hjsh2\") on node \"crc\" DevicePath \"\"" Dec 06 04:23:27 crc kubenswrapper[4802]: I1206 04:23:27.658288 4802 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8e8984b-b60b-48e5-9586-1a1a8d883143-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 04:23:27 crc kubenswrapper[4802]: I1206 04:23:27.658300 4802 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b8e8984b-b60b-48e5-9586-1a1a8d883143-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 04:23:27 crc kubenswrapper[4802]: I1206 04:23:27.658311 4802 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/b8e8984b-b60b-48e5-9586-1a1a8d883143-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 06 04:23:27 crc kubenswrapper[4802]: I1206 04:23:27.658320 4802 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8e8984b-b60b-48e5-9586-1a1a8d883143-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:23:27 crc kubenswrapper[4802]: I1206 04:23:27.658331 4802 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/b8e8984b-b60b-48e5-9586-1a1a8d883143-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 06 04:23:27 crc kubenswrapper[4802]: I1206 04:23:27.658340 4802 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/b8e8984b-b60b-48e5-9586-1a1a8d883143-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 06 04:23:27 crc kubenswrapper[4802]: I1206 04:23:27.932586 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5plkw" event={"ID":"b8e8984b-b60b-48e5-9586-1a1a8d883143","Type":"ContainerDied","Data":"f38b3c2d3cfc059e92d01f68369fcf02c0115261b1d1197b4b12dbeae10989bb"} Dec 06 04:23:27 crc kubenswrapper[4802]: I1206 04:23:27.932855 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f38b3c2d3cfc059e92d01f68369fcf02c0115261b1d1197b4b12dbeae10989bb" Dec 06 04:23:27 crc kubenswrapper[4802]: I1206 04:23:27.932869 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5plkw" Dec 06 04:23:28 crc kubenswrapper[4802]: I1206 04:23:28.090928 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ns78h"] Dec 06 04:23:28 crc kubenswrapper[4802]: E1206 04:23:28.091339 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8e8984b-b60b-48e5-9586-1a1a8d883143" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 06 04:23:28 crc kubenswrapper[4802]: I1206 04:23:28.091359 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8e8984b-b60b-48e5-9586-1a1a8d883143" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 06 04:23:28 crc kubenswrapper[4802]: I1206 04:23:28.091537 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8e8984b-b60b-48e5-9586-1a1a8d883143" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 06 04:23:28 crc kubenswrapper[4802]: I1206 04:23:28.092216 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ns78h" Dec 06 04:23:28 crc kubenswrapper[4802]: I1206 04:23:28.095030 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-rgvjc" Dec 06 04:23:28 crc kubenswrapper[4802]: I1206 04:23:28.095109 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Dec 06 04:23:28 crc kubenswrapper[4802]: I1206 04:23:28.095717 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 04:23:28 crc kubenswrapper[4802]: I1206 04:23:28.095746 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 04:23:28 crc kubenswrapper[4802]: I1206 04:23:28.095742 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 04:23:28 crc kubenswrapper[4802]: I1206 04:23:28.107434 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ns78h"] Dec 06 04:23:28 crc kubenswrapper[4802]: I1206 04:23:28.165670 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/565b62ff-4b9a-4e53-b61c-d6c492d7b253-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-ns78h\" (UID: \"565b62ff-4b9a-4e53-b61c-d6c492d7b253\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ns78h" Dec 06 04:23:28 crc kubenswrapper[4802]: I1206 04:23:28.165811 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/565b62ff-4b9a-4e53-b61c-d6c492d7b253-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-ns78h\" (UID: \"565b62ff-4b9a-4e53-b61c-d6c492d7b253\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ns78h" Dec 06 04:23:28 crc kubenswrapper[4802]: I1206 04:23:28.165841 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/565b62ff-4b9a-4e53-b61c-d6c492d7b253-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-ns78h\" (UID: \"565b62ff-4b9a-4e53-b61c-d6c492d7b253\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ns78h" Dec 06 04:23:28 crc kubenswrapper[4802]: I1206 04:23:28.165891 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvv4q\" (UniqueName: \"kubernetes.io/projected/565b62ff-4b9a-4e53-b61c-d6c492d7b253-kube-api-access-gvv4q\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-ns78h\" (UID: \"565b62ff-4b9a-4e53-b61c-d6c492d7b253\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ns78h" Dec 06 04:23:28 crc kubenswrapper[4802]: I1206 04:23:28.165960 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/565b62ff-4b9a-4e53-b61c-d6c492d7b253-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-ns78h\" (UID: \"565b62ff-4b9a-4e53-b61c-d6c492d7b253\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ns78h" Dec 06 04:23:28 crc kubenswrapper[4802]: I1206 04:23:28.166009 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/565b62ff-4b9a-4e53-b61c-d6c492d7b253-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-ns78h\" (UID: \"565b62ff-4b9a-4e53-b61c-d6c492d7b253\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ns78h" Dec 06 04:23:28 crc kubenswrapper[4802]: I1206 04:23:28.166237 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/565b62ff-4b9a-4e53-b61c-d6c492d7b253-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-ns78h\" (UID: \"565b62ff-4b9a-4e53-b61c-d6c492d7b253\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ns78h" Dec 06 04:23:28 crc kubenswrapper[4802]: I1206 04:23:28.267570 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/565b62ff-4b9a-4e53-b61c-d6c492d7b253-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-ns78h\" (UID: \"565b62ff-4b9a-4e53-b61c-d6c492d7b253\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ns78h" Dec 06 04:23:28 crc kubenswrapper[4802]: I1206 04:23:28.267694 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/565b62ff-4b9a-4e53-b61c-d6c492d7b253-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-ns78h\" (UID: \"565b62ff-4b9a-4e53-b61c-d6c492d7b253\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ns78h" Dec 06 04:23:28 crc kubenswrapper[4802]: I1206 04:23:28.267808 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/565b62ff-4b9a-4e53-b61c-d6c492d7b253-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-ns78h\" (UID: \"565b62ff-4b9a-4e53-b61c-d6c492d7b253\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ns78h" Dec 06 04:23:28 crc kubenswrapper[4802]: I1206 04:23:28.267924 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/565b62ff-4b9a-4e53-b61c-d6c492d7b253-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-ns78h\" (UID: \"565b62ff-4b9a-4e53-b61c-d6c492d7b253\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ns78h" Dec 06 04:23:28 crc kubenswrapper[4802]: I1206 04:23:28.267990 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/565b62ff-4b9a-4e53-b61c-d6c492d7b253-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-ns78h\" (UID: \"565b62ff-4b9a-4e53-b61c-d6c492d7b253\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ns78h" Dec 06 04:23:28 crc kubenswrapper[4802]: I1206 04:23:28.268095 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvv4q\" (UniqueName: \"kubernetes.io/projected/565b62ff-4b9a-4e53-b61c-d6c492d7b253-kube-api-access-gvv4q\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-ns78h\" (UID: \"565b62ff-4b9a-4e53-b61c-d6c492d7b253\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ns78h" Dec 06 04:23:28 crc kubenswrapper[4802]: I1206 04:23:28.268175 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/565b62ff-4b9a-4e53-b61c-d6c492d7b253-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-ns78h\" (UID: \"565b62ff-4b9a-4e53-b61c-d6c492d7b253\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ns78h" Dec 06 04:23:28 crc kubenswrapper[4802]: I1206 04:23:28.273037 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/565b62ff-4b9a-4e53-b61c-d6c492d7b253-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-ns78h\" (UID: \"565b62ff-4b9a-4e53-b61c-d6c492d7b253\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ns78h" Dec 06 04:23:28 crc kubenswrapper[4802]: I1206 04:23:28.273179 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/565b62ff-4b9a-4e53-b61c-d6c492d7b253-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-ns78h\" (UID: \"565b62ff-4b9a-4e53-b61c-d6c492d7b253\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ns78h" Dec 06 04:23:28 crc kubenswrapper[4802]: I1206 04:23:28.273576 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/565b62ff-4b9a-4e53-b61c-d6c492d7b253-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-ns78h\" (UID: \"565b62ff-4b9a-4e53-b61c-d6c492d7b253\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ns78h" Dec 06 04:23:28 crc kubenswrapper[4802]: I1206 04:23:28.274178 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/565b62ff-4b9a-4e53-b61c-d6c492d7b253-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-ns78h\" (UID: \"565b62ff-4b9a-4e53-b61c-d6c492d7b253\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ns78h" Dec 06 04:23:28 crc kubenswrapper[4802]: I1206 04:23:28.274707 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/565b62ff-4b9a-4e53-b61c-d6c492d7b253-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-ns78h\" (UID: \"565b62ff-4b9a-4e53-b61c-d6c492d7b253\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ns78h" Dec 06 04:23:28 crc kubenswrapper[4802]: I1206 04:23:28.277629 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/565b62ff-4b9a-4e53-b61c-d6c492d7b253-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-ns78h\" (UID: \"565b62ff-4b9a-4e53-b61c-d6c492d7b253\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ns78h" Dec 06 04:23:28 crc kubenswrapper[4802]: I1206 04:23:28.287706 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvv4q\" (UniqueName: \"kubernetes.io/projected/565b62ff-4b9a-4e53-b61c-d6c492d7b253-kube-api-access-gvv4q\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-ns78h\" (UID: \"565b62ff-4b9a-4e53-b61c-d6c492d7b253\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ns78h" Dec 06 04:23:28 crc kubenswrapper[4802]: I1206 04:23:28.417595 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ns78h" Dec 06 04:23:28 crc kubenswrapper[4802]: I1206 04:23:28.978581 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ns78h"] Dec 06 04:23:28 crc kubenswrapper[4802]: I1206 04:23:28.991173 4802 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 04:23:29 crc kubenswrapper[4802]: I1206 04:23:29.948573 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ns78h" event={"ID":"565b62ff-4b9a-4e53-b61c-d6c492d7b253","Type":"ContainerStarted","Data":"b2107bf5f1212117700cb36a53f37ef79e35882f425cc9008a038d4381bd8515"} Dec 06 04:23:29 crc kubenswrapper[4802]: I1206 04:23:29.949024 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ns78h" event={"ID":"565b62ff-4b9a-4e53-b61c-d6c492d7b253","Type":"ContainerStarted","Data":"46da03655e11de09611ce762d03d205ed467c6f6103562ea43ecac2a92c3d91f"} Dec 06 04:24:43 crc kubenswrapper[4802]: I1206 04:24:43.283124 4802 patch_prober.go:28] interesting pod/machine-config-daemon-zpxxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:24:43 crc kubenswrapper[4802]: I1206 04:24:43.283657 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:25:13 crc kubenswrapper[4802]: I1206 04:25:13.283614 4802 patch_prober.go:28] interesting pod/machine-config-daemon-zpxxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:25:13 crc kubenswrapper[4802]: I1206 04:25:13.284317 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:25:43 crc kubenswrapper[4802]: I1206 04:25:43.283958 4802 patch_prober.go:28] interesting pod/machine-config-daemon-zpxxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:25:43 crc kubenswrapper[4802]: I1206 04:25:43.284893 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:25:43 crc kubenswrapper[4802]: I1206 04:25:43.284948 4802 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" Dec 06 04:25:43 crc kubenswrapper[4802]: I1206 04:25:43.285736 4802 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"77efcf0d5716fddf110919e082917d423162e98a637ed45489f128e2cf499066"} pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 04:25:43 crc kubenswrapper[4802]: I1206 04:25:43.285823 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" containerID="cri-o://77efcf0d5716fddf110919e082917d423162e98a637ed45489f128e2cf499066" gracePeriod=600 Dec 06 04:25:43 crc kubenswrapper[4802]: I1206 04:25:43.398031 4802 generic.go:334] "Generic (PLEG): container finished" podID="565b62ff-4b9a-4e53-b61c-d6c492d7b253" containerID="b2107bf5f1212117700cb36a53f37ef79e35882f425cc9008a038d4381bd8515" exitCode=0 Dec 06 04:25:43 crc kubenswrapper[4802]: I1206 04:25:43.398115 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ns78h" event={"ID":"565b62ff-4b9a-4e53-b61c-d6c492d7b253","Type":"ContainerDied","Data":"b2107bf5f1212117700cb36a53f37ef79e35882f425cc9008a038d4381bd8515"} Dec 06 04:25:44 crc kubenswrapper[4802]: I1206 04:25:44.413736 4802 generic.go:334] "Generic (PLEG): container finished" podID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerID="77efcf0d5716fddf110919e082917d423162e98a637ed45489f128e2cf499066" exitCode=0 Dec 06 04:25:44 crc kubenswrapper[4802]: I1206 04:25:44.413778 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" event={"ID":"6df38316-e0d3-4018-8d27-3620eba3a68d","Type":"ContainerDied","Data":"77efcf0d5716fddf110919e082917d423162e98a637ed45489f128e2cf499066"} Dec 06 04:25:44 crc kubenswrapper[4802]: I1206 04:25:44.414269 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" event={"ID":"6df38316-e0d3-4018-8d27-3620eba3a68d","Type":"ContainerStarted","Data":"ce5c9338b4ba535e43ce4e2eb6bfc97dd2ebe59199fb4ec623134cb1b6016b9c"} Dec 06 04:25:44 crc kubenswrapper[4802]: I1206 04:25:44.414311 4802 scope.go:117] "RemoveContainer" containerID="f8a17abbaa167592a628d9fdb3771d61462952618f6382d447c431c49f7c5690" Dec 06 04:25:44 crc kubenswrapper[4802]: I1206 04:25:44.865457 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ns78h" Dec 06 04:25:44 crc kubenswrapper[4802]: I1206 04:25:44.967341 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/565b62ff-4b9a-4e53-b61c-d6c492d7b253-telemetry-combined-ca-bundle\") pod \"565b62ff-4b9a-4e53-b61c-d6c492d7b253\" (UID: \"565b62ff-4b9a-4e53-b61c-d6c492d7b253\") " Dec 06 04:25:44 crc kubenswrapper[4802]: I1206 04:25:44.967428 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/565b62ff-4b9a-4e53-b61c-d6c492d7b253-ssh-key\") pod \"565b62ff-4b9a-4e53-b61c-d6c492d7b253\" (UID: \"565b62ff-4b9a-4e53-b61c-d6c492d7b253\") " Dec 06 04:25:44 crc kubenswrapper[4802]: I1206 04:25:44.967522 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/565b62ff-4b9a-4e53-b61c-d6c492d7b253-ceilometer-compute-config-data-1\") pod \"565b62ff-4b9a-4e53-b61c-d6c492d7b253\" (UID: \"565b62ff-4b9a-4e53-b61c-d6c492d7b253\") " Dec 06 04:25:44 crc kubenswrapper[4802]: I1206 04:25:44.967548 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/565b62ff-4b9a-4e53-b61c-d6c492d7b253-inventory\") pod \"565b62ff-4b9a-4e53-b61c-d6c492d7b253\" (UID: \"565b62ff-4b9a-4e53-b61c-d6c492d7b253\") " Dec 06 04:25:44 crc kubenswrapper[4802]: I1206 04:25:44.967577 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/565b62ff-4b9a-4e53-b61c-d6c492d7b253-ceilometer-compute-config-data-2\") pod \"565b62ff-4b9a-4e53-b61c-d6c492d7b253\" (UID: \"565b62ff-4b9a-4e53-b61c-d6c492d7b253\") " Dec 06 04:25:44 crc kubenswrapper[4802]: I1206 04:25:44.967737 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/565b62ff-4b9a-4e53-b61c-d6c492d7b253-ceilometer-compute-config-data-0\") pod \"565b62ff-4b9a-4e53-b61c-d6c492d7b253\" (UID: \"565b62ff-4b9a-4e53-b61c-d6c492d7b253\") " Dec 06 04:25:44 crc kubenswrapper[4802]: I1206 04:25:44.967835 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gvv4q\" (UniqueName: \"kubernetes.io/projected/565b62ff-4b9a-4e53-b61c-d6c492d7b253-kube-api-access-gvv4q\") pod \"565b62ff-4b9a-4e53-b61c-d6c492d7b253\" (UID: \"565b62ff-4b9a-4e53-b61c-d6c492d7b253\") " Dec 06 04:25:44 crc kubenswrapper[4802]: I1206 04:25:44.992593 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/565b62ff-4b9a-4e53-b61c-d6c492d7b253-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "565b62ff-4b9a-4e53-b61c-d6c492d7b253" (UID: "565b62ff-4b9a-4e53-b61c-d6c492d7b253"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:25:44 crc kubenswrapper[4802]: I1206 04:25:44.992606 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/565b62ff-4b9a-4e53-b61c-d6c492d7b253-kube-api-access-gvv4q" (OuterVolumeSpecName: "kube-api-access-gvv4q") pod "565b62ff-4b9a-4e53-b61c-d6c492d7b253" (UID: "565b62ff-4b9a-4e53-b61c-d6c492d7b253"). InnerVolumeSpecName "kube-api-access-gvv4q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:25:44 crc kubenswrapper[4802]: I1206 04:25:44.998014 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/565b62ff-4b9a-4e53-b61c-d6c492d7b253-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "565b62ff-4b9a-4e53-b61c-d6c492d7b253" (UID: "565b62ff-4b9a-4e53-b61c-d6c492d7b253"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:25:44 crc kubenswrapper[4802]: I1206 04:25:44.998399 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/565b62ff-4b9a-4e53-b61c-d6c492d7b253-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "565b62ff-4b9a-4e53-b61c-d6c492d7b253" (UID: "565b62ff-4b9a-4e53-b61c-d6c492d7b253"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:25:45 crc kubenswrapper[4802]: I1206 04:25:45.007999 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/565b62ff-4b9a-4e53-b61c-d6c492d7b253-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "565b62ff-4b9a-4e53-b61c-d6c492d7b253" (UID: "565b62ff-4b9a-4e53-b61c-d6c492d7b253"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:25:45 crc kubenswrapper[4802]: I1206 04:25:45.012122 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/565b62ff-4b9a-4e53-b61c-d6c492d7b253-inventory" (OuterVolumeSpecName: "inventory") pod "565b62ff-4b9a-4e53-b61c-d6c492d7b253" (UID: "565b62ff-4b9a-4e53-b61c-d6c492d7b253"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:25:45 crc kubenswrapper[4802]: I1206 04:25:45.017548 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/565b62ff-4b9a-4e53-b61c-d6c492d7b253-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "565b62ff-4b9a-4e53-b61c-d6c492d7b253" (UID: "565b62ff-4b9a-4e53-b61c-d6c492d7b253"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:25:45 crc kubenswrapper[4802]: I1206 04:25:45.071723 4802 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/565b62ff-4b9a-4e53-b61c-d6c492d7b253-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 06 04:25:45 crc kubenswrapper[4802]: I1206 04:25:45.071784 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gvv4q\" (UniqueName: \"kubernetes.io/projected/565b62ff-4b9a-4e53-b61c-d6c492d7b253-kube-api-access-gvv4q\") on node \"crc\" DevicePath \"\"" Dec 06 04:25:45 crc kubenswrapper[4802]: I1206 04:25:45.071800 4802 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/565b62ff-4b9a-4e53-b61c-d6c492d7b253-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:25:45 crc kubenswrapper[4802]: I1206 04:25:45.071816 4802 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/565b62ff-4b9a-4e53-b61c-d6c492d7b253-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 04:25:45 crc kubenswrapper[4802]: I1206 04:25:45.071830 4802 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/565b62ff-4b9a-4e53-b61c-d6c492d7b253-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 04:25:45 crc kubenswrapper[4802]: I1206 04:25:45.071843 4802 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/565b62ff-4b9a-4e53-b61c-d6c492d7b253-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 06 04:25:45 crc kubenswrapper[4802]: I1206 04:25:45.071855 4802 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/565b62ff-4b9a-4e53-b61c-d6c492d7b253-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 06 04:25:45 crc kubenswrapper[4802]: I1206 04:25:45.426547 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ns78h" event={"ID":"565b62ff-4b9a-4e53-b61c-d6c492d7b253","Type":"ContainerDied","Data":"46da03655e11de09611ce762d03d205ed467c6f6103562ea43ecac2a92c3d91f"} Dec 06 04:25:45 crc kubenswrapper[4802]: I1206 04:25:45.426596 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-ns78h" Dec 06 04:25:45 crc kubenswrapper[4802]: I1206 04:25:45.426610 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="46da03655e11de09611ce762d03d205ed467c6f6103562ea43ecac2a92c3d91f" Dec 06 04:26:40 crc kubenswrapper[4802]: I1206 04:26:40.061840 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Dec 06 04:26:40 crc kubenswrapper[4802]: E1206 04:26:40.062691 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="565b62ff-4b9a-4e53-b61c-d6c492d7b253" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 06 04:26:40 crc kubenswrapper[4802]: I1206 04:26:40.062704 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="565b62ff-4b9a-4e53-b61c-d6c492d7b253" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 06 04:26:40 crc kubenswrapper[4802]: I1206 04:26:40.063002 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="565b62ff-4b9a-4e53-b61c-d6c492d7b253" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 06 04:26:40 crc kubenswrapper[4802]: I1206 04:26:40.063618 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 06 04:26:40 crc kubenswrapper[4802]: I1206 04:26:40.065424 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Dec 06 04:26:40 crc kubenswrapper[4802]: I1206 04:26:40.065776 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-grnp2" Dec 06 04:26:40 crc kubenswrapper[4802]: I1206 04:26:40.065801 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Dec 06 04:26:40 crc kubenswrapper[4802]: I1206 04:26:40.065867 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 06 04:26:40 crc kubenswrapper[4802]: I1206 04:26:40.080572 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 06 04:26:40 crc kubenswrapper[4802]: I1206 04:26:40.202044 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/63c9a1aa-a443-4071-b60e-317a7b81395e-config-data\") pod \"tempest-tests-tempest\" (UID: \"63c9a1aa-a443-4071-b60e-317a7b81395e\") " pod="openstack/tempest-tests-tempest" Dec 06 04:26:40 crc kubenswrapper[4802]: I1206 04:26:40.202133 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/63c9a1aa-a443-4071-b60e-317a7b81395e-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"63c9a1aa-a443-4071-b60e-317a7b81395e\") " pod="openstack/tempest-tests-tempest" Dec 06 04:26:40 crc kubenswrapper[4802]: I1206 04:26:40.202162 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/63c9a1aa-a443-4071-b60e-317a7b81395e-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"63c9a1aa-a443-4071-b60e-317a7b81395e\") " pod="openstack/tempest-tests-tempest" Dec 06 04:26:40 crc kubenswrapper[4802]: I1206 04:26:40.202237 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/63c9a1aa-a443-4071-b60e-317a7b81395e-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"63c9a1aa-a443-4071-b60e-317a7b81395e\") " pod="openstack/tempest-tests-tempest" Dec 06 04:26:40 crc kubenswrapper[4802]: I1206 04:26:40.202306 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/63c9a1aa-a443-4071-b60e-317a7b81395e-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"63c9a1aa-a443-4071-b60e-317a7b81395e\") " pod="openstack/tempest-tests-tempest" Dec 06 04:26:40 crc kubenswrapper[4802]: I1206 04:26:40.202389 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/63c9a1aa-a443-4071-b60e-317a7b81395e-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"63c9a1aa-a443-4071-b60e-317a7b81395e\") " pod="openstack/tempest-tests-tempest" Dec 06 04:26:40 crc kubenswrapper[4802]: I1206 04:26:40.202462 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srnml\" (UniqueName: \"kubernetes.io/projected/63c9a1aa-a443-4071-b60e-317a7b81395e-kube-api-access-srnml\") pod \"tempest-tests-tempest\" (UID: \"63c9a1aa-a443-4071-b60e-317a7b81395e\") " pod="openstack/tempest-tests-tempest" Dec 06 04:26:40 crc kubenswrapper[4802]: I1206 04:26:40.202494 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"tempest-tests-tempest\" (UID: \"63c9a1aa-a443-4071-b60e-317a7b81395e\") " pod="openstack/tempest-tests-tempest" Dec 06 04:26:40 crc kubenswrapper[4802]: I1206 04:26:40.202898 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/63c9a1aa-a443-4071-b60e-317a7b81395e-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"63c9a1aa-a443-4071-b60e-317a7b81395e\") " pod="openstack/tempest-tests-tempest" Dec 06 04:26:40 crc kubenswrapper[4802]: I1206 04:26:40.304474 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/63c9a1aa-a443-4071-b60e-317a7b81395e-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"63c9a1aa-a443-4071-b60e-317a7b81395e\") " pod="openstack/tempest-tests-tempest" Dec 06 04:26:40 crc kubenswrapper[4802]: I1206 04:26:40.304527 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/63c9a1aa-a443-4071-b60e-317a7b81395e-config-data\") pod \"tempest-tests-tempest\" (UID: \"63c9a1aa-a443-4071-b60e-317a7b81395e\") " pod="openstack/tempest-tests-tempest" Dec 06 04:26:40 crc kubenswrapper[4802]: I1206 04:26:40.304553 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/63c9a1aa-a443-4071-b60e-317a7b81395e-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"63c9a1aa-a443-4071-b60e-317a7b81395e\") " pod="openstack/tempest-tests-tempest" Dec 06 04:26:40 crc kubenswrapper[4802]: I1206 04:26:40.304577 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/63c9a1aa-a443-4071-b60e-317a7b81395e-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"63c9a1aa-a443-4071-b60e-317a7b81395e\") " pod="openstack/tempest-tests-tempest" Dec 06 04:26:40 crc kubenswrapper[4802]: I1206 04:26:40.304650 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/63c9a1aa-a443-4071-b60e-317a7b81395e-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"63c9a1aa-a443-4071-b60e-317a7b81395e\") " pod="openstack/tempest-tests-tempest" Dec 06 04:26:40 crc kubenswrapper[4802]: I1206 04:26:40.304692 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/63c9a1aa-a443-4071-b60e-317a7b81395e-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"63c9a1aa-a443-4071-b60e-317a7b81395e\") " pod="openstack/tempest-tests-tempest" Dec 06 04:26:40 crc kubenswrapper[4802]: I1206 04:26:40.304713 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/63c9a1aa-a443-4071-b60e-317a7b81395e-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"63c9a1aa-a443-4071-b60e-317a7b81395e\") " pod="openstack/tempest-tests-tempest" Dec 06 04:26:40 crc kubenswrapper[4802]: I1206 04:26:40.304774 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srnml\" (UniqueName: \"kubernetes.io/projected/63c9a1aa-a443-4071-b60e-317a7b81395e-kube-api-access-srnml\") pod \"tempest-tests-tempest\" (UID: \"63c9a1aa-a443-4071-b60e-317a7b81395e\") " pod="openstack/tempest-tests-tempest" Dec 06 04:26:40 crc kubenswrapper[4802]: I1206 04:26:40.304803 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"tempest-tests-tempest\" (UID: \"63c9a1aa-a443-4071-b60e-317a7b81395e\") " pod="openstack/tempest-tests-tempest" Dec 06 04:26:40 crc kubenswrapper[4802]: I1206 04:26:40.305199 4802 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"tempest-tests-tempest\" (UID: \"63c9a1aa-a443-4071-b60e-317a7b81395e\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/tempest-tests-tempest" Dec 06 04:26:40 crc kubenswrapper[4802]: I1206 04:26:40.305614 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/63c9a1aa-a443-4071-b60e-317a7b81395e-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"63c9a1aa-a443-4071-b60e-317a7b81395e\") " pod="openstack/tempest-tests-tempest" Dec 06 04:26:40 crc kubenswrapper[4802]: I1206 04:26:40.305805 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/63c9a1aa-a443-4071-b60e-317a7b81395e-config-data\") pod \"tempest-tests-tempest\" (UID: \"63c9a1aa-a443-4071-b60e-317a7b81395e\") " pod="openstack/tempest-tests-tempest" Dec 06 04:26:40 crc kubenswrapper[4802]: I1206 04:26:40.305876 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/63c9a1aa-a443-4071-b60e-317a7b81395e-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"63c9a1aa-a443-4071-b60e-317a7b81395e\") " pod="openstack/tempest-tests-tempest" Dec 06 04:26:40 crc kubenswrapper[4802]: I1206 04:26:40.306576 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/63c9a1aa-a443-4071-b60e-317a7b81395e-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"63c9a1aa-a443-4071-b60e-317a7b81395e\") " pod="openstack/tempest-tests-tempest" Dec 06 04:26:40 crc kubenswrapper[4802]: I1206 04:26:40.311495 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/63c9a1aa-a443-4071-b60e-317a7b81395e-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"63c9a1aa-a443-4071-b60e-317a7b81395e\") " pod="openstack/tempest-tests-tempest" Dec 06 04:26:40 crc kubenswrapper[4802]: I1206 04:26:40.311583 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/63c9a1aa-a443-4071-b60e-317a7b81395e-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"63c9a1aa-a443-4071-b60e-317a7b81395e\") " pod="openstack/tempest-tests-tempest" Dec 06 04:26:40 crc kubenswrapper[4802]: I1206 04:26:40.324640 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/63c9a1aa-a443-4071-b60e-317a7b81395e-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"63c9a1aa-a443-4071-b60e-317a7b81395e\") " pod="openstack/tempest-tests-tempest" Dec 06 04:26:40 crc kubenswrapper[4802]: I1206 04:26:40.335293 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srnml\" (UniqueName: \"kubernetes.io/projected/63c9a1aa-a443-4071-b60e-317a7b81395e-kube-api-access-srnml\") pod \"tempest-tests-tempest\" (UID: \"63c9a1aa-a443-4071-b60e-317a7b81395e\") " pod="openstack/tempest-tests-tempest" Dec 06 04:26:40 crc kubenswrapper[4802]: I1206 04:26:40.352685 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"tempest-tests-tempest\" (UID: \"63c9a1aa-a443-4071-b60e-317a7b81395e\") " pod="openstack/tempest-tests-tempest" Dec 06 04:26:40 crc kubenswrapper[4802]: I1206 04:26:40.390947 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 06 04:26:40 crc kubenswrapper[4802]: I1206 04:26:40.944730 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 06 04:26:40 crc kubenswrapper[4802]: I1206 04:26:40.955992 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"63c9a1aa-a443-4071-b60e-317a7b81395e","Type":"ContainerStarted","Data":"07b4a4f8606f4f4fb91971ba1cbdfe0e279ec2deaa25ec341c8ea507ecc7f035"} Dec 06 04:27:14 crc kubenswrapper[4802]: E1206 04:27:14.702057 4802 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Dec 06 04:27:14 crc kubenswrapper[4802]: E1206 04:27:14.702660 4802 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-srnml,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(63c9a1aa-a443-4071-b60e-317a7b81395e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 04:27:14 crc kubenswrapper[4802]: E1206 04:27:14.703952 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="63c9a1aa-a443-4071-b60e-317a7b81395e" Dec 06 04:27:15 crc kubenswrapper[4802]: E1206 04:27:15.348082 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="63c9a1aa-a443-4071-b60e-317a7b81395e" Dec 06 04:27:31 crc kubenswrapper[4802]: I1206 04:27:31.295583 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 06 04:27:33 crc kubenswrapper[4802]: I1206 04:27:33.530506 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"63c9a1aa-a443-4071-b60e-317a7b81395e","Type":"ContainerStarted","Data":"6d5db2b1a0c0fa5c3abd518f7e7502c7095b137577f7b3ca876dd132cbfe3ad6"} Dec 06 04:27:33 crc kubenswrapper[4802]: I1206 04:27:33.559789 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=4.215661757 podStartE2EDuration="54.559772542s" podCreationTimestamp="2025-12-06 04:26:39 +0000 UTC" firstStartedPulling="2025-12-06 04:26:40.948738415 +0000 UTC m=+2793.820647567" lastFinishedPulling="2025-12-06 04:27:31.29284916 +0000 UTC m=+2844.164758352" observedRunningTime="2025-12-06 04:27:33.554367084 +0000 UTC m=+2846.426276246" watchObservedRunningTime="2025-12-06 04:27:33.559772542 +0000 UTC m=+2846.431681694" Dec 06 04:27:43 crc kubenswrapper[4802]: I1206 04:27:43.283599 4802 patch_prober.go:28] interesting pod/machine-config-daemon-zpxxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:27:43 crc kubenswrapper[4802]: I1206 04:27:43.284309 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:27:56 crc kubenswrapper[4802]: I1206 04:27:56.822609 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nlvvm"] Dec 06 04:27:56 crc kubenswrapper[4802]: I1206 04:27:56.825613 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nlvvm" Dec 06 04:27:56 crc kubenswrapper[4802]: I1206 04:27:56.835503 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nlvvm"] Dec 06 04:27:56 crc kubenswrapper[4802]: I1206 04:27:56.872558 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfp4w\" (UniqueName: \"kubernetes.io/projected/6eb93440-f1c7-4430-8e8d-770d5405de0c-kube-api-access-bfp4w\") pod \"community-operators-nlvvm\" (UID: \"6eb93440-f1c7-4430-8e8d-770d5405de0c\") " pod="openshift-marketplace/community-operators-nlvvm" Dec 06 04:27:56 crc kubenswrapper[4802]: I1206 04:27:56.872659 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6eb93440-f1c7-4430-8e8d-770d5405de0c-utilities\") pod \"community-operators-nlvvm\" (UID: \"6eb93440-f1c7-4430-8e8d-770d5405de0c\") " pod="openshift-marketplace/community-operators-nlvvm" Dec 06 04:27:56 crc kubenswrapper[4802]: I1206 04:27:56.872681 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6eb93440-f1c7-4430-8e8d-770d5405de0c-catalog-content\") pod \"community-operators-nlvvm\" (UID: \"6eb93440-f1c7-4430-8e8d-770d5405de0c\") " pod="openshift-marketplace/community-operators-nlvvm" Dec 06 04:27:56 crc kubenswrapper[4802]: I1206 04:27:56.974006 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfp4w\" (UniqueName: \"kubernetes.io/projected/6eb93440-f1c7-4430-8e8d-770d5405de0c-kube-api-access-bfp4w\") pod \"community-operators-nlvvm\" (UID: \"6eb93440-f1c7-4430-8e8d-770d5405de0c\") " pod="openshift-marketplace/community-operators-nlvvm" Dec 06 04:27:56 crc kubenswrapper[4802]: I1206 04:27:56.974092 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6eb93440-f1c7-4430-8e8d-770d5405de0c-utilities\") pod \"community-operators-nlvvm\" (UID: \"6eb93440-f1c7-4430-8e8d-770d5405de0c\") " pod="openshift-marketplace/community-operators-nlvvm" Dec 06 04:27:56 crc kubenswrapper[4802]: I1206 04:27:56.974110 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6eb93440-f1c7-4430-8e8d-770d5405de0c-catalog-content\") pod \"community-operators-nlvvm\" (UID: \"6eb93440-f1c7-4430-8e8d-770d5405de0c\") " pod="openshift-marketplace/community-operators-nlvvm" Dec 06 04:27:56 crc kubenswrapper[4802]: I1206 04:27:56.974513 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6eb93440-f1c7-4430-8e8d-770d5405de0c-utilities\") pod \"community-operators-nlvvm\" (UID: \"6eb93440-f1c7-4430-8e8d-770d5405de0c\") " pod="openshift-marketplace/community-operators-nlvvm" Dec 06 04:27:56 crc kubenswrapper[4802]: I1206 04:27:56.974592 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6eb93440-f1c7-4430-8e8d-770d5405de0c-catalog-content\") pod \"community-operators-nlvvm\" (UID: \"6eb93440-f1c7-4430-8e8d-770d5405de0c\") " pod="openshift-marketplace/community-operators-nlvvm" Dec 06 04:27:56 crc kubenswrapper[4802]: I1206 04:27:56.995473 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfp4w\" (UniqueName: \"kubernetes.io/projected/6eb93440-f1c7-4430-8e8d-770d5405de0c-kube-api-access-bfp4w\") pod \"community-operators-nlvvm\" (UID: \"6eb93440-f1c7-4430-8e8d-770d5405de0c\") " pod="openshift-marketplace/community-operators-nlvvm" Dec 06 04:27:57 crc kubenswrapper[4802]: I1206 04:27:57.144631 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nlvvm" Dec 06 04:27:57 crc kubenswrapper[4802]: I1206 04:27:57.778340 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nlvvm"] Dec 06 04:27:57 crc kubenswrapper[4802]: I1206 04:27:57.821586 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nlvvm" event={"ID":"6eb93440-f1c7-4430-8e8d-770d5405de0c","Type":"ContainerStarted","Data":"1377c9048ce57b8e1fa164dd35c2dd5fa40e8c77987b911467f210385fd34fcc"} Dec 06 04:27:58 crc kubenswrapper[4802]: I1206 04:27:58.832662 4802 generic.go:334] "Generic (PLEG): container finished" podID="6eb93440-f1c7-4430-8e8d-770d5405de0c" containerID="72f209633f83d6be9223a8ec919db2e01903263ed3de5493a05099f09aff27c4" exitCode=0 Dec 06 04:27:58 crc kubenswrapper[4802]: I1206 04:27:58.832964 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nlvvm" event={"ID":"6eb93440-f1c7-4430-8e8d-770d5405de0c","Type":"ContainerDied","Data":"72f209633f83d6be9223a8ec919db2e01903263ed3de5493a05099f09aff27c4"} Dec 06 04:27:59 crc kubenswrapper[4802]: I1206 04:27:59.841930 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nlvvm" event={"ID":"6eb93440-f1c7-4430-8e8d-770d5405de0c","Type":"ContainerStarted","Data":"2c4ddc9976dda8e2e0ed2fc763dedf5062d639111e558f36cfac43a0edbab42d"} Dec 06 04:28:00 crc kubenswrapper[4802]: I1206 04:28:00.854072 4802 generic.go:334] "Generic (PLEG): container finished" podID="6eb93440-f1c7-4430-8e8d-770d5405de0c" containerID="2c4ddc9976dda8e2e0ed2fc763dedf5062d639111e558f36cfac43a0edbab42d" exitCode=0 Dec 06 04:28:00 crc kubenswrapper[4802]: I1206 04:28:00.854151 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nlvvm" event={"ID":"6eb93440-f1c7-4430-8e8d-770d5405de0c","Type":"ContainerDied","Data":"2c4ddc9976dda8e2e0ed2fc763dedf5062d639111e558f36cfac43a0edbab42d"} Dec 06 04:28:01 crc kubenswrapper[4802]: I1206 04:28:01.863913 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nlvvm" event={"ID":"6eb93440-f1c7-4430-8e8d-770d5405de0c","Type":"ContainerStarted","Data":"056fe3ef1b4af6c5dd1f9396b3380b90db54ddef5eafdc6bead668598a939ce0"} Dec 06 04:28:01 crc kubenswrapper[4802]: I1206 04:28:01.894789 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nlvvm" podStartSLOduration=3.450624307 podStartE2EDuration="5.894745302s" podCreationTimestamp="2025-12-06 04:27:56 +0000 UTC" firstStartedPulling="2025-12-06 04:27:58.836698474 +0000 UTC m=+2871.708607636" lastFinishedPulling="2025-12-06 04:28:01.280819459 +0000 UTC m=+2874.152728631" observedRunningTime="2025-12-06 04:28:01.883326459 +0000 UTC m=+2874.755235611" watchObservedRunningTime="2025-12-06 04:28:01.894745302 +0000 UTC m=+2874.766654454" Dec 06 04:28:07 crc kubenswrapper[4802]: I1206 04:28:07.145392 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nlvvm" Dec 06 04:28:07 crc kubenswrapper[4802]: I1206 04:28:07.146022 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nlvvm" Dec 06 04:28:07 crc kubenswrapper[4802]: I1206 04:28:07.191152 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nlvvm" Dec 06 04:28:07 crc kubenswrapper[4802]: I1206 04:28:07.998348 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nlvvm" Dec 06 04:28:08 crc kubenswrapper[4802]: I1206 04:28:08.052246 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nlvvm"] Dec 06 04:28:09 crc kubenswrapper[4802]: I1206 04:28:09.951719 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-nlvvm" podUID="6eb93440-f1c7-4430-8e8d-770d5405de0c" containerName="registry-server" containerID="cri-o://056fe3ef1b4af6c5dd1f9396b3380b90db54ddef5eafdc6bead668598a939ce0" gracePeriod=2 Dec 06 04:28:10 crc kubenswrapper[4802]: I1206 04:28:10.493104 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nlvvm" Dec 06 04:28:10 crc kubenswrapper[4802]: I1206 04:28:10.599331 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bfp4w\" (UniqueName: \"kubernetes.io/projected/6eb93440-f1c7-4430-8e8d-770d5405de0c-kube-api-access-bfp4w\") pod \"6eb93440-f1c7-4430-8e8d-770d5405de0c\" (UID: \"6eb93440-f1c7-4430-8e8d-770d5405de0c\") " Dec 06 04:28:10 crc kubenswrapper[4802]: I1206 04:28:10.599441 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6eb93440-f1c7-4430-8e8d-770d5405de0c-catalog-content\") pod \"6eb93440-f1c7-4430-8e8d-770d5405de0c\" (UID: \"6eb93440-f1c7-4430-8e8d-770d5405de0c\") " Dec 06 04:28:10 crc kubenswrapper[4802]: I1206 04:28:10.599522 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6eb93440-f1c7-4430-8e8d-770d5405de0c-utilities\") pod \"6eb93440-f1c7-4430-8e8d-770d5405de0c\" (UID: \"6eb93440-f1c7-4430-8e8d-770d5405de0c\") " Dec 06 04:28:10 crc kubenswrapper[4802]: I1206 04:28:10.601657 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6eb93440-f1c7-4430-8e8d-770d5405de0c-utilities" (OuterVolumeSpecName: "utilities") pod "6eb93440-f1c7-4430-8e8d-770d5405de0c" (UID: "6eb93440-f1c7-4430-8e8d-770d5405de0c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:28:10 crc kubenswrapper[4802]: I1206 04:28:10.607799 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6eb93440-f1c7-4430-8e8d-770d5405de0c-kube-api-access-bfp4w" (OuterVolumeSpecName: "kube-api-access-bfp4w") pod "6eb93440-f1c7-4430-8e8d-770d5405de0c" (UID: "6eb93440-f1c7-4430-8e8d-770d5405de0c"). InnerVolumeSpecName "kube-api-access-bfp4w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:28:10 crc kubenswrapper[4802]: I1206 04:28:10.660248 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6eb93440-f1c7-4430-8e8d-770d5405de0c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6eb93440-f1c7-4430-8e8d-770d5405de0c" (UID: "6eb93440-f1c7-4430-8e8d-770d5405de0c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:28:10 crc kubenswrapper[4802]: I1206 04:28:10.702537 4802 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6eb93440-f1c7-4430-8e8d-770d5405de0c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 04:28:10 crc kubenswrapper[4802]: I1206 04:28:10.702591 4802 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6eb93440-f1c7-4430-8e8d-770d5405de0c-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 04:28:10 crc kubenswrapper[4802]: I1206 04:28:10.702604 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bfp4w\" (UniqueName: \"kubernetes.io/projected/6eb93440-f1c7-4430-8e8d-770d5405de0c-kube-api-access-bfp4w\") on node \"crc\" DevicePath \"\"" Dec 06 04:28:10 crc kubenswrapper[4802]: I1206 04:28:10.963739 4802 generic.go:334] "Generic (PLEG): container finished" podID="6eb93440-f1c7-4430-8e8d-770d5405de0c" containerID="056fe3ef1b4af6c5dd1f9396b3380b90db54ddef5eafdc6bead668598a939ce0" exitCode=0 Dec 06 04:28:10 crc kubenswrapper[4802]: I1206 04:28:10.963797 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nlvvm" event={"ID":"6eb93440-f1c7-4430-8e8d-770d5405de0c","Type":"ContainerDied","Data":"056fe3ef1b4af6c5dd1f9396b3380b90db54ddef5eafdc6bead668598a939ce0"} Dec 06 04:28:10 crc kubenswrapper[4802]: I1206 04:28:10.963825 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nlvvm" event={"ID":"6eb93440-f1c7-4430-8e8d-770d5405de0c","Type":"ContainerDied","Data":"1377c9048ce57b8e1fa164dd35c2dd5fa40e8c77987b911467f210385fd34fcc"} Dec 06 04:28:10 crc kubenswrapper[4802]: I1206 04:28:10.963853 4802 scope.go:117] "RemoveContainer" containerID="056fe3ef1b4af6c5dd1f9396b3380b90db54ddef5eafdc6bead668598a939ce0" Dec 06 04:28:10 crc kubenswrapper[4802]: I1206 04:28:10.963870 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nlvvm" Dec 06 04:28:10 crc kubenswrapper[4802]: I1206 04:28:10.987995 4802 scope.go:117] "RemoveContainer" containerID="2c4ddc9976dda8e2e0ed2fc763dedf5062d639111e558f36cfac43a0edbab42d" Dec 06 04:28:11 crc kubenswrapper[4802]: I1206 04:28:11.007593 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nlvvm"] Dec 06 04:28:11 crc kubenswrapper[4802]: I1206 04:28:11.024450 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-nlvvm"] Dec 06 04:28:11 crc kubenswrapper[4802]: I1206 04:28:11.034182 4802 scope.go:117] "RemoveContainer" containerID="72f209633f83d6be9223a8ec919db2e01903263ed3de5493a05099f09aff27c4" Dec 06 04:28:11 crc kubenswrapper[4802]: I1206 04:28:11.077950 4802 scope.go:117] "RemoveContainer" containerID="056fe3ef1b4af6c5dd1f9396b3380b90db54ddef5eafdc6bead668598a939ce0" Dec 06 04:28:11 crc kubenswrapper[4802]: E1206 04:28:11.078663 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"056fe3ef1b4af6c5dd1f9396b3380b90db54ddef5eafdc6bead668598a939ce0\": container with ID starting with 056fe3ef1b4af6c5dd1f9396b3380b90db54ddef5eafdc6bead668598a939ce0 not found: ID does not exist" containerID="056fe3ef1b4af6c5dd1f9396b3380b90db54ddef5eafdc6bead668598a939ce0" Dec 06 04:28:11 crc kubenswrapper[4802]: I1206 04:28:11.078804 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"056fe3ef1b4af6c5dd1f9396b3380b90db54ddef5eafdc6bead668598a939ce0"} err="failed to get container status \"056fe3ef1b4af6c5dd1f9396b3380b90db54ddef5eafdc6bead668598a939ce0\": rpc error: code = NotFound desc = could not find container \"056fe3ef1b4af6c5dd1f9396b3380b90db54ddef5eafdc6bead668598a939ce0\": container with ID starting with 056fe3ef1b4af6c5dd1f9396b3380b90db54ddef5eafdc6bead668598a939ce0 not found: ID does not exist" Dec 06 04:28:11 crc kubenswrapper[4802]: I1206 04:28:11.078857 4802 scope.go:117] "RemoveContainer" containerID="2c4ddc9976dda8e2e0ed2fc763dedf5062d639111e558f36cfac43a0edbab42d" Dec 06 04:28:11 crc kubenswrapper[4802]: E1206 04:28:11.079548 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c4ddc9976dda8e2e0ed2fc763dedf5062d639111e558f36cfac43a0edbab42d\": container with ID starting with 2c4ddc9976dda8e2e0ed2fc763dedf5062d639111e558f36cfac43a0edbab42d not found: ID does not exist" containerID="2c4ddc9976dda8e2e0ed2fc763dedf5062d639111e558f36cfac43a0edbab42d" Dec 06 04:28:11 crc kubenswrapper[4802]: I1206 04:28:11.079579 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c4ddc9976dda8e2e0ed2fc763dedf5062d639111e558f36cfac43a0edbab42d"} err="failed to get container status \"2c4ddc9976dda8e2e0ed2fc763dedf5062d639111e558f36cfac43a0edbab42d\": rpc error: code = NotFound desc = could not find container \"2c4ddc9976dda8e2e0ed2fc763dedf5062d639111e558f36cfac43a0edbab42d\": container with ID starting with 2c4ddc9976dda8e2e0ed2fc763dedf5062d639111e558f36cfac43a0edbab42d not found: ID does not exist" Dec 06 04:28:11 crc kubenswrapper[4802]: I1206 04:28:11.079597 4802 scope.go:117] "RemoveContainer" containerID="72f209633f83d6be9223a8ec919db2e01903263ed3de5493a05099f09aff27c4" Dec 06 04:28:11 crc kubenswrapper[4802]: E1206 04:28:11.080739 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72f209633f83d6be9223a8ec919db2e01903263ed3de5493a05099f09aff27c4\": container with ID starting with 72f209633f83d6be9223a8ec919db2e01903263ed3de5493a05099f09aff27c4 not found: ID does not exist" containerID="72f209633f83d6be9223a8ec919db2e01903263ed3de5493a05099f09aff27c4" Dec 06 04:28:11 crc kubenswrapper[4802]: I1206 04:28:11.080819 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72f209633f83d6be9223a8ec919db2e01903263ed3de5493a05099f09aff27c4"} err="failed to get container status \"72f209633f83d6be9223a8ec919db2e01903263ed3de5493a05099f09aff27c4\": rpc error: code = NotFound desc = could not find container \"72f209633f83d6be9223a8ec919db2e01903263ed3de5493a05099f09aff27c4\": container with ID starting with 72f209633f83d6be9223a8ec919db2e01903263ed3de5493a05099f09aff27c4 not found: ID does not exist" Dec 06 04:28:11 crc kubenswrapper[4802]: I1206 04:28:11.463621 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6eb93440-f1c7-4430-8e8d-770d5405de0c" path="/var/lib/kubelet/pods/6eb93440-f1c7-4430-8e8d-770d5405de0c/volumes" Dec 06 04:28:13 crc kubenswrapper[4802]: I1206 04:28:13.283475 4802 patch_prober.go:28] interesting pod/machine-config-daemon-zpxxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:28:13 crc kubenswrapper[4802]: I1206 04:28:13.284483 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:28:43 crc kubenswrapper[4802]: I1206 04:28:43.284235 4802 patch_prober.go:28] interesting pod/machine-config-daemon-zpxxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:28:43 crc kubenswrapper[4802]: I1206 04:28:43.284860 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:28:43 crc kubenswrapper[4802]: I1206 04:28:43.284925 4802 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" Dec 06 04:28:43 crc kubenswrapper[4802]: I1206 04:28:43.285959 4802 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ce5c9338b4ba535e43ce4e2eb6bfc97dd2ebe59199fb4ec623134cb1b6016b9c"} pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 04:28:43 crc kubenswrapper[4802]: I1206 04:28:43.286053 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" containerID="cri-o://ce5c9338b4ba535e43ce4e2eb6bfc97dd2ebe59199fb4ec623134cb1b6016b9c" gracePeriod=600 Dec 06 04:28:43 crc kubenswrapper[4802]: E1206 04:28:43.413205 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:28:44 crc kubenswrapper[4802]: I1206 04:28:44.270085 4802 generic.go:334] "Generic (PLEG): container finished" podID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerID="ce5c9338b4ba535e43ce4e2eb6bfc97dd2ebe59199fb4ec623134cb1b6016b9c" exitCode=0 Dec 06 04:28:44 crc kubenswrapper[4802]: I1206 04:28:44.270194 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" event={"ID":"6df38316-e0d3-4018-8d27-3620eba3a68d","Type":"ContainerDied","Data":"ce5c9338b4ba535e43ce4e2eb6bfc97dd2ebe59199fb4ec623134cb1b6016b9c"} Dec 06 04:28:44 crc kubenswrapper[4802]: I1206 04:28:44.270427 4802 scope.go:117] "RemoveContainer" containerID="77efcf0d5716fddf110919e082917d423162e98a637ed45489f128e2cf499066" Dec 06 04:28:44 crc kubenswrapper[4802]: I1206 04:28:44.271195 4802 scope.go:117] "RemoveContainer" containerID="ce5c9338b4ba535e43ce4e2eb6bfc97dd2ebe59199fb4ec623134cb1b6016b9c" Dec 06 04:28:44 crc kubenswrapper[4802]: E1206 04:28:44.271701 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:28:55 crc kubenswrapper[4802]: I1206 04:28:55.450581 4802 scope.go:117] "RemoveContainer" containerID="ce5c9338b4ba535e43ce4e2eb6bfc97dd2ebe59199fb4ec623134cb1b6016b9c" Dec 06 04:28:55 crc kubenswrapper[4802]: E1206 04:28:55.452111 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:29:09 crc kubenswrapper[4802]: I1206 04:29:09.454219 4802 scope.go:117] "RemoveContainer" containerID="ce5c9338b4ba535e43ce4e2eb6bfc97dd2ebe59199fb4ec623134cb1b6016b9c" Dec 06 04:29:09 crc kubenswrapper[4802]: E1206 04:29:09.455141 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:29:21 crc kubenswrapper[4802]: I1206 04:29:21.452115 4802 scope.go:117] "RemoveContainer" containerID="ce5c9338b4ba535e43ce4e2eb6bfc97dd2ebe59199fb4ec623134cb1b6016b9c" Dec 06 04:29:21 crc kubenswrapper[4802]: E1206 04:29:21.452923 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:29:32 crc kubenswrapper[4802]: I1206 04:29:32.450384 4802 scope.go:117] "RemoveContainer" containerID="ce5c9338b4ba535e43ce4e2eb6bfc97dd2ebe59199fb4ec623134cb1b6016b9c" Dec 06 04:29:32 crc kubenswrapper[4802]: E1206 04:29:32.452120 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:29:43 crc kubenswrapper[4802]: I1206 04:29:43.450585 4802 scope.go:117] "RemoveContainer" containerID="ce5c9338b4ba535e43ce4e2eb6bfc97dd2ebe59199fb4ec623134cb1b6016b9c" Dec 06 04:29:43 crc kubenswrapper[4802]: E1206 04:29:43.451519 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:29:54 crc kubenswrapper[4802]: I1206 04:29:54.451487 4802 scope.go:117] "RemoveContainer" containerID="ce5c9338b4ba535e43ce4e2eb6bfc97dd2ebe59199fb4ec623134cb1b6016b9c" Dec 06 04:29:54 crc kubenswrapper[4802]: E1206 04:29:54.452506 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:30:00 crc kubenswrapper[4802]: I1206 04:30:00.146880 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416590-vr5ld"] Dec 06 04:30:00 crc kubenswrapper[4802]: E1206 04:30:00.147907 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6eb93440-f1c7-4430-8e8d-770d5405de0c" containerName="extract-content" Dec 06 04:30:00 crc kubenswrapper[4802]: I1206 04:30:00.147926 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="6eb93440-f1c7-4430-8e8d-770d5405de0c" containerName="extract-content" Dec 06 04:30:00 crc kubenswrapper[4802]: E1206 04:30:00.147939 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6eb93440-f1c7-4430-8e8d-770d5405de0c" containerName="extract-utilities" Dec 06 04:30:00 crc kubenswrapper[4802]: I1206 04:30:00.147947 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="6eb93440-f1c7-4430-8e8d-770d5405de0c" containerName="extract-utilities" Dec 06 04:30:00 crc kubenswrapper[4802]: E1206 04:30:00.147972 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6eb93440-f1c7-4430-8e8d-770d5405de0c" containerName="registry-server" Dec 06 04:30:00 crc kubenswrapper[4802]: I1206 04:30:00.147979 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="6eb93440-f1c7-4430-8e8d-770d5405de0c" containerName="registry-server" Dec 06 04:30:00 crc kubenswrapper[4802]: I1206 04:30:00.148187 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="6eb93440-f1c7-4430-8e8d-770d5405de0c" containerName="registry-server" Dec 06 04:30:00 crc kubenswrapper[4802]: I1206 04:30:00.149238 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416590-vr5ld" Dec 06 04:30:00 crc kubenswrapper[4802]: I1206 04:30:00.155571 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 04:30:00 crc kubenswrapper[4802]: I1206 04:30:00.155991 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 04:30:00 crc kubenswrapper[4802]: I1206 04:30:00.176947 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416590-vr5ld"] Dec 06 04:30:00 crc kubenswrapper[4802]: I1206 04:30:00.208546 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8b7e5bdf-8bf6-4a75-8fe7-bd74584596cf-config-volume\") pod \"collect-profiles-29416590-vr5ld\" (UID: \"8b7e5bdf-8bf6-4a75-8fe7-bd74584596cf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416590-vr5ld" Dec 06 04:30:00 crc kubenswrapper[4802]: I1206 04:30:00.208630 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kp5hn\" (UniqueName: \"kubernetes.io/projected/8b7e5bdf-8bf6-4a75-8fe7-bd74584596cf-kube-api-access-kp5hn\") pod \"collect-profiles-29416590-vr5ld\" (UID: \"8b7e5bdf-8bf6-4a75-8fe7-bd74584596cf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416590-vr5ld" Dec 06 04:30:00 crc kubenswrapper[4802]: I1206 04:30:00.208762 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8b7e5bdf-8bf6-4a75-8fe7-bd74584596cf-secret-volume\") pod \"collect-profiles-29416590-vr5ld\" (UID: \"8b7e5bdf-8bf6-4a75-8fe7-bd74584596cf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416590-vr5ld" Dec 06 04:30:00 crc kubenswrapper[4802]: I1206 04:30:00.310120 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8b7e5bdf-8bf6-4a75-8fe7-bd74584596cf-config-volume\") pod \"collect-profiles-29416590-vr5ld\" (UID: \"8b7e5bdf-8bf6-4a75-8fe7-bd74584596cf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416590-vr5ld" Dec 06 04:30:00 crc kubenswrapper[4802]: I1206 04:30:00.310183 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kp5hn\" (UniqueName: \"kubernetes.io/projected/8b7e5bdf-8bf6-4a75-8fe7-bd74584596cf-kube-api-access-kp5hn\") pod \"collect-profiles-29416590-vr5ld\" (UID: \"8b7e5bdf-8bf6-4a75-8fe7-bd74584596cf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416590-vr5ld" Dec 06 04:30:00 crc kubenswrapper[4802]: I1206 04:30:00.310256 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8b7e5bdf-8bf6-4a75-8fe7-bd74584596cf-secret-volume\") pod \"collect-profiles-29416590-vr5ld\" (UID: \"8b7e5bdf-8bf6-4a75-8fe7-bd74584596cf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416590-vr5ld" Dec 06 04:30:00 crc kubenswrapper[4802]: I1206 04:30:00.311148 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8b7e5bdf-8bf6-4a75-8fe7-bd74584596cf-config-volume\") pod \"collect-profiles-29416590-vr5ld\" (UID: \"8b7e5bdf-8bf6-4a75-8fe7-bd74584596cf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416590-vr5ld" Dec 06 04:30:00 crc kubenswrapper[4802]: I1206 04:30:00.316419 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8b7e5bdf-8bf6-4a75-8fe7-bd74584596cf-secret-volume\") pod \"collect-profiles-29416590-vr5ld\" (UID: \"8b7e5bdf-8bf6-4a75-8fe7-bd74584596cf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416590-vr5ld" Dec 06 04:30:00 crc kubenswrapper[4802]: I1206 04:30:00.326869 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kp5hn\" (UniqueName: \"kubernetes.io/projected/8b7e5bdf-8bf6-4a75-8fe7-bd74584596cf-kube-api-access-kp5hn\") pod \"collect-profiles-29416590-vr5ld\" (UID: \"8b7e5bdf-8bf6-4a75-8fe7-bd74584596cf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416590-vr5ld" Dec 06 04:30:00 crc kubenswrapper[4802]: I1206 04:30:00.493219 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416590-vr5ld" Dec 06 04:30:00 crc kubenswrapper[4802]: I1206 04:30:00.967164 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416590-vr5ld"] Dec 06 04:30:00 crc kubenswrapper[4802]: I1206 04:30:00.991049 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416590-vr5ld" event={"ID":"8b7e5bdf-8bf6-4a75-8fe7-bd74584596cf","Type":"ContainerStarted","Data":"705f6f7de67f746830b67bf38307e9974ae58fd96821480728ae58d7078033b1"} Dec 06 04:30:02 crc kubenswrapper[4802]: I1206 04:30:02.004376 4802 generic.go:334] "Generic (PLEG): container finished" podID="8b7e5bdf-8bf6-4a75-8fe7-bd74584596cf" containerID="6dffc29abffe2de19cb1a235774be1d05601076ecc213a0b9dd5d2cea15729b0" exitCode=0 Dec 06 04:30:02 crc kubenswrapper[4802]: I1206 04:30:02.004475 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416590-vr5ld" event={"ID":"8b7e5bdf-8bf6-4a75-8fe7-bd74584596cf","Type":"ContainerDied","Data":"6dffc29abffe2de19cb1a235774be1d05601076ecc213a0b9dd5d2cea15729b0"} Dec 06 04:30:03 crc kubenswrapper[4802]: I1206 04:30:03.460614 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416590-vr5ld" Dec 06 04:30:03 crc kubenswrapper[4802]: I1206 04:30:03.580638 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8b7e5bdf-8bf6-4a75-8fe7-bd74584596cf-config-volume\") pod \"8b7e5bdf-8bf6-4a75-8fe7-bd74584596cf\" (UID: \"8b7e5bdf-8bf6-4a75-8fe7-bd74584596cf\") " Dec 06 04:30:03 crc kubenswrapper[4802]: I1206 04:30:03.580797 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8b7e5bdf-8bf6-4a75-8fe7-bd74584596cf-secret-volume\") pod \"8b7e5bdf-8bf6-4a75-8fe7-bd74584596cf\" (UID: \"8b7e5bdf-8bf6-4a75-8fe7-bd74584596cf\") " Dec 06 04:30:03 crc kubenswrapper[4802]: I1206 04:30:03.580981 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kp5hn\" (UniqueName: \"kubernetes.io/projected/8b7e5bdf-8bf6-4a75-8fe7-bd74584596cf-kube-api-access-kp5hn\") pod \"8b7e5bdf-8bf6-4a75-8fe7-bd74584596cf\" (UID: \"8b7e5bdf-8bf6-4a75-8fe7-bd74584596cf\") " Dec 06 04:30:03 crc kubenswrapper[4802]: I1206 04:30:03.581668 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b7e5bdf-8bf6-4a75-8fe7-bd74584596cf-config-volume" (OuterVolumeSpecName: "config-volume") pod "8b7e5bdf-8bf6-4a75-8fe7-bd74584596cf" (UID: "8b7e5bdf-8bf6-4a75-8fe7-bd74584596cf"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:30:03 crc kubenswrapper[4802]: I1206 04:30:03.588045 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b7e5bdf-8bf6-4a75-8fe7-bd74584596cf-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "8b7e5bdf-8bf6-4a75-8fe7-bd74584596cf" (UID: "8b7e5bdf-8bf6-4a75-8fe7-bd74584596cf"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:30:03 crc kubenswrapper[4802]: I1206 04:30:03.588372 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b7e5bdf-8bf6-4a75-8fe7-bd74584596cf-kube-api-access-kp5hn" (OuterVolumeSpecName: "kube-api-access-kp5hn") pod "8b7e5bdf-8bf6-4a75-8fe7-bd74584596cf" (UID: "8b7e5bdf-8bf6-4a75-8fe7-bd74584596cf"). InnerVolumeSpecName "kube-api-access-kp5hn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:30:03 crc kubenswrapper[4802]: I1206 04:30:03.683847 4802 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8b7e5bdf-8bf6-4a75-8fe7-bd74584596cf-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 04:30:03 crc kubenswrapper[4802]: I1206 04:30:03.683882 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kp5hn\" (UniqueName: \"kubernetes.io/projected/8b7e5bdf-8bf6-4a75-8fe7-bd74584596cf-kube-api-access-kp5hn\") on node \"crc\" DevicePath \"\"" Dec 06 04:30:03 crc kubenswrapper[4802]: I1206 04:30:03.683893 4802 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8b7e5bdf-8bf6-4a75-8fe7-bd74584596cf-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 04:30:04 crc kubenswrapper[4802]: I1206 04:30:04.025921 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416590-vr5ld" event={"ID":"8b7e5bdf-8bf6-4a75-8fe7-bd74584596cf","Type":"ContainerDied","Data":"705f6f7de67f746830b67bf38307e9974ae58fd96821480728ae58d7078033b1"} Dec 06 04:30:04 crc kubenswrapper[4802]: I1206 04:30:04.026287 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="705f6f7de67f746830b67bf38307e9974ae58fd96821480728ae58d7078033b1" Dec 06 04:30:04 crc kubenswrapper[4802]: I1206 04:30:04.025987 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416590-vr5ld" Dec 06 04:30:04 crc kubenswrapper[4802]: I1206 04:30:04.544974 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416545-9z9qr"] Dec 06 04:30:04 crc kubenswrapper[4802]: I1206 04:30:04.552628 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416545-9z9qr"] Dec 06 04:30:05 crc kubenswrapper[4802]: I1206 04:30:05.451232 4802 scope.go:117] "RemoveContainer" containerID="ce5c9338b4ba535e43ce4e2eb6bfc97dd2ebe59199fb4ec623134cb1b6016b9c" Dec 06 04:30:05 crc kubenswrapper[4802]: E1206 04:30:05.451915 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:30:05 crc kubenswrapper[4802]: I1206 04:30:05.493160 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db461f4d-dd3b-48ff-9d07-f686dce48122" path="/var/lib/kubelet/pods/db461f4d-dd3b-48ff-9d07-f686dce48122/volumes" Dec 06 04:30:16 crc kubenswrapper[4802]: I1206 04:30:16.450527 4802 scope.go:117] "RemoveContainer" containerID="ce5c9338b4ba535e43ce4e2eb6bfc97dd2ebe59199fb4ec623134cb1b6016b9c" Dec 06 04:30:16 crc kubenswrapper[4802]: E1206 04:30:16.451390 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:30:22 crc kubenswrapper[4802]: I1206 04:30:22.724882 4802 scope.go:117] "RemoveContainer" containerID="5e5b040efcfd42d0060dedb8ab5af64340411a989af338fdb19b891bf7d03056" Dec 06 04:30:29 crc kubenswrapper[4802]: I1206 04:30:29.450781 4802 scope.go:117] "RemoveContainer" containerID="ce5c9338b4ba535e43ce4e2eb6bfc97dd2ebe59199fb4ec623134cb1b6016b9c" Dec 06 04:30:29 crc kubenswrapper[4802]: E1206 04:30:29.451612 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:30:40 crc kubenswrapper[4802]: I1206 04:30:40.450623 4802 scope.go:117] "RemoveContainer" containerID="ce5c9338b4ba535e43ce4e2eb6bfc97dd2ebe59199fb4ec623134cb1b6016b9c" Dec 06 04:30:40 crc kubenswrapper[4802]: E1206 04:30:40.451525 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:30:52 crc kubenswrapper[4802]: I1206 04:30:52.450891 4802 scope.go:117] "RemoveContainer" containerID="ce5c9338b4ba535e43ce4e2eb6bfc97dd2ebe59199fb4ec623134cb1b6016b9c" Dec 06 04:30:52 crc kubenswrapper[4802]: E1206 04:30:52.451627 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:31:07 crc kubenswrapper[4802]: I1206 04:31:07.455639 4802 scope.go:117] "RemoveContainer" containerID="ce5c9338b4ba535e43ce4e2eb6bfc97dd2ebe59199fb4ec623134cb1b6016b9c" Dec 06 04:31:07 crc kubenswrapper[4802]: E1206 04:31:07.456379 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:31:22 crc kubenswrapper[4802]: I1206 04:31:22.450612 4802 scope.go:117] "RemoveContainer" containerID="ce5c9338b4ba535e43ce4e2eb6bfc97dd2ebe59199fb4ec623134cb1b6016b9c" Dec 06 04:31:22 crc kubenswrapper[4802]: E1206 04:31:22.451323 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:31:37 crc kubenswrapper[4802]: I1206 04:31:37.455718 4802 scope.go:117] "RemoveContainer" containerID="ce5c9338b4ba535e43ce4e2eb6bfc97dd2ebe59199fb4ec623134cb1b6016b9c" Dec 06 04:31:37 crc kubenswrapper[4802]: E1206 04:31:37.456550 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:31:51 crc kubenswrapper[4802]: I1206 04:31:51.449843 4802 scope.go:117] "RemoveContainer" containerID="ce5c9338b4ba535e43ce4e2eb6bfc97dd2ebe59199fb4ec623134cb1b6016b9c" Dec 06 04:31:51 crc kubenswrapper[4802]: E1206 04:31:51.450607 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:32:02 crc kubenswrapper[4802]: I1206 04:32:02.450335 4802 scope.go:117] "RemoveContainer" containerID="ce5c9338b4ba535e43ce4e2eb6bfc97dd2ebe59199fb4ec623134cb1b6016b9c" Dec 06 04:32:02 crc kubenswrapper[4802]: E1206 04:32:02.451044 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:32:15 crc kubenswrapper[4802]: I1206 04:32:15.450269 4802 scope.go:117] "RemoveContainer" containerID="ce5c9338b4ba535e43ce4e2eb6bfc97dd2ebe59199fb4ec623134cb1b6016b9c" Dec 06 04:32:15 crc kubenswrapper[4802]: E1206 04:32:15.450923 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:32:28 crc kubenswrapper[4802]: I1206 04:32:28.451246 4802 scope.go:117] "RemoveContainer" containerID="ce5c9338b4ba535e43ce4e2eb6bfc97dd2ebe59199fb4ec623134cb1b6016b9c" Dec 06 04:32:28 crc kubenswrapper[4802]: E1206 04:32:28.452108 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:32:40 crc kubenswrapper[4802]: I1206 04:32:40.450262 4802 scope.go:117] "RemoveContainer" containerID="ce5c9338b4ba535e43ce4e2eb6bfc97dd2ebe59199fb4ec623134cb1b6016b9c" Dec 06 04:32:40 crc kubenswrapper[4802]: E1206 04:32:40.450946 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:32:53 crc kubenswrapper[4802]: I1206 04:32:53.450783 4802 scope.go:117] "RemoveContainer" containerID="ce5c9338b4ba535e43ce4e2eb6bfc97dd2ebe59199fb4ec623134cb1b6016b9c" Dec 06 04:32:53 crc kubenswrapper[4802]: E1206 04:32:53.452092 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:33:06 crc kubenswrapper[4802]: I1206 04:33:06.449948 4802 scope.go:117] "RemoveContainer" containerID="ce5c9338b4ba535e43ce4e2eb6bfc97dd2ebe59199fb4ec623134cb1b6016b9c" Dec 06 04:33:06 crc kubenswrapper[4802]: E1206 04:33:06.450700 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:33:21 crc kubenswrapper[4802]: I1206 04:33:21.450448 4802 scope.go:117] "RemoveContainer" containerID="ce5c9338b4ba535e43ce4e2eb6bfc97dd2ebe59199fb4ec623134cb1b6016b9c" Dec 06 04:33:21 crc kubenswrapper[4802]: E1206 04:33:21.451368 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:33:27 crc kubenswrapper[4802]: I1206 04:33:27.059231 4802 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-686fbdfd7f-sfrn5" podUID="4985762a-1d53-40ad-85da-c6728bdbaa3d" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Dec 06 04:33:33 crc kubenswrapper[4802]: I1206 04:33:33.451718 4802 scope.go:117] "RemoveContainer" containerID="ce5c9338b4ba535e43ce4e2eb6bfc97dd2ebe59199fb4ec623134cb1b6016b9c" Dec 06 04:33:33 crc kubenswrapper[4802]: E1206 04:33:33.453008 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:33:44 crc kubenswrapper[4802]: I1206 04:33:44.451097 4802 scope.go:117] "RemoveContainer" containerID="ce5c9338b4ba535e43ce4e2eb6bfc97dd2ebe59199fb4ec623134cb1b6016b9c" Dec 06 04:33:45 crc kubenswrapper[4802]: I1206 04:33:45.077441 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" event={"ID":"6df38316-e0d3-4018-8d27-3620eba3a68d","Type":"ContainerStarted","Data":"549c98317cf5beda6d72e30c3f31e49c51734f812384e19e7ca9ae6871d665af"} Dec 06 04:35:14 crc kubenswrapper[4802]: I1206 04:35:14.229381 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2h8z5"] Dec 06 04:35:14 crc kubenswrapper[4802]: E1206 04:35:14.230563 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b7e5bdf-8bf6-4a75-8fe7-bd74584596cf" containerName="collect-profiles" Dec 06 04:35:14 crc kubenswrapper[4802]: I1206 04:35:14.230586 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b7e5bdf-8bf6-4a75-8fe7-bd74584596cf" containerName="collect-profiles" Dec 06 04:35:14 crc kubenswrapper[4802]: I1206 04:35:14.230951 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b7e5bdf-8bf6-4a75-8fe7-bd74584596cf" containerName="collect-profiles" Dec 06 04:35:14 crc kubenswrapper[4802]: I1206 04:35:14.234444 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2h8z5" Dec 06 04:35:14 crc kubenswrapper[4802]: I1206 04:35:14.245235 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2h8z5"] Dec 06 04:35:14 crc kubenswrapper[4802]: I1206 04:35:14.309046 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc3f763e-de8d-43bd-86e2-abceb87a373a-utilities\") pod \"certified-operators-2h8z5\" (UID: \"bc3f763e-de8d-43bd-86e2-abceb87a373a\") " pod="openshift-marketplace/certified-operators-2h8z5" Dec 06 04:35:14 crc kubenswrapper[4802]: I1206 04:35:14.309124 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdl8k\" (UniqueName: \"kubernetes.io/projected/bc3f763e-de8d-43bd-86e2-abceb87a373a-kube-api-access-xdl8k\") pod \"certified-operators-2h8z5\" (UID: \"bc3f763e-de8d-43bd-86e2-abceb87a373a\") " pod="openshift-marketplace/certified-operators-2h8z5" Dec 06 04:35:14 crc kubenswrapper[4802]: I1206 04:35:14.309192 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc3f763e-de8d-43bd-86e2-abceb87a373a-catalog-content\") pod \"certified-operators-2h8z5\" (UID: \"bc3f763e-de8d-43bd-86e2-abceb87a373a\") " pod="openshift-marketplace/certified-operators-2h8z5" Dec 06 04:35:14 crc kubenswrapper[4802]: I1206 04:35:14.410922 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdl8k\" (UniqueName: \"kubernetes.io/projected/bc3f763e-de8d-43bd-86e2-abceb87a373a-kube-api-access-xdl8k\") pod \"certified-operators-2h8z5\" (UID: \"bc3f763e-de8d-43bd-86e2-abceb87a373a\") " pod="openshift-marketplace/certified-operators-2h8z5" Dec 06 04:35:14 crc kubenswrapper[4802]: I1206 04:35:14.411021 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc3f763e-de8d-43bd-86e2-abceb87a373a-catalog-content\") pod \"certified-operators-2h8z5\" (UID: \"bc3f763e-de8d-43bd-86e2-abceb87a373a\") " pod="openshift-marketplace/certified-operators-2h8z5" Dec 06 04:35:14 crc kubenswrapper[4802]: I1206 04:35:14.411160 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc3f763e-de8d-43bd-86e2-abceb87a373a-utilities\") pod \"certified-operators-2h8z5\" (UID: \"bc3f763e-de8d-43bd-86e2-abceb87a373a\") " pod="openshift-marketplace/certified-operators-2h8z5" Dec 06 04:35:14 crc kubenswrapper[4802]: I1206 04:35:14.411679 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc3f763e-de8d-43bd-86e2-abceb87a373a-catalog-content\") pod \"certified-operators-2h8z5\" (UID: \"bc3f763e-de8d-43bd-86e2-abceb87a373a\") " pod="openshift-marketplace/certified-operators-2h8z5" Dec 06 04:35:14 crc kubenswrapper[4802]: I1206 04:35:14.411708 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc3f763e-de8d-43bd-86e2-abceb87a373a-utilities\") pod \"certified-operators-2h8z5\" (UID: \"bc3f763e-de8d-43bd-86e2-abceb87a373a\") " pod="openshift-marketplace/certified-operators-2h8z5" Dec 06 04:35:14 crc kubenswrapper[4802]: I1206 04:35:14.439294 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdl8k\" (UniqueName: \"kubernetes.io/projected/bc3f763e-de8d-43bd-86e2-abceb87a373a-kube-api-access-xdl8k\") pod \"certified-operators-2h8z5\" (UID: \"bc3f763e-de8d-43bd-86e2-abceb87a373a\") " pod="openshift-marketplace/certified-operators-2h8z5" Dec 06 04:35:14 crc kubenswrapper[4802]: I1206 04:35:14.573696 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2h8z5" Dec 06 04:35:15 crc kubenswrapper[4802]: I1206 04:35:15.122003 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2h8z5"] Dec 06 04:35:15 crc kubenswrapper[4802]: I1206 04:35:15.941109 4802 generic.go:334] "Generic (PLEG): container finished" podID="bc3f763e-de8d-43bd-86e2-abceb87a373a" containerID="e3adcff27252a8ef1f3290e5e7ce1967650cb0c56bc29c5332bfce6ecb6089a1" exitCode=0 Dec 06 04:35:15 crc kubenswrapper[4802]: I1206 04:35:15.941157 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2h8z5" event={"ID":"bc3f763e-de8d-43bd-86e2-abceb87a373a","Type":"ContainerDied","Data":"e3adcff27252a8ef1f3290e5e7ce1967650cb0c56bc29c5332bfce6ecb6089a1"} Dec 06 04:35:15 crc kubenswrapper[4802]: I1206 04:35:15.943110 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2h8z5" event={"ID":"bc3f763e-de8d-43bd-86e2-abceb87a373a","Type":"ContainerStarted","Data":"33483fc412e8b08044486833b481861bec497cdf9a5ef1db9745c66cc3267191"} Dec 06 04:35:15 crc kubenswrapper[4802]: I1206 04:35:15.945121 4802 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 04:35:16 crc kubenswrapper[4802]: I1206 04:35:16.953823 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2h8z5" event={"ID":"bc3f763e-de8d-43bd-86e2-abceb87a373a","Type":"ContainerStarted","Data":"76a730687b641ded76c1a205f8f420dfa1576ba402ec0610643d841f206faf89"} Dec 06 04:35:18 crc kubenswrapper[4802]: I1206 04:35:18.975069 4802 generic.go:334] "Generic (PLEG): container finished" podID="bc3f763e-de8d-43bd-86e2-abceb87a373a" containerID="76a730687b641ded76c1a205f8f420dfa1576ba402ec0610643d841f206faf89" exitCode=0 Dec 06 04:35:18 crc kubenswrapper[4802]: I1206 04:35:18.975119 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2h8z5" event={"ID":"bc3f763e-de8d-43bd-86e2-abceb87a373a","Type":"ContainerDied","Data":"76a730687b641ded76c1a205f8f420dfa1576ba402ec0610643d841f206faf89"} Dec 06 04:35:19 crc kubenswrapper[4802]: I1206 04:35:19.987620 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2h8z5" event={"ID":"bc3f763e-de8d-43bd-86e2-abceb87a373a","Type":"ContainerStarted","Data":"d2379b923f568e4ce4189958b91c013d2a1005997f0764309e9e50a8821b0db7"} Dec 06 04:35:20 crc kubenswrapper[4802]: I1206 04:35:20.014401 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2h8z5" podStartSLOduration=2.5531088459999998 podStartE2EDuration="6.014353531s" podCreationTimestamp="2025-12-06 04:35:14 +0000 UTC" firstStartedPulling="2025-12-06 04:35:15.94459624 +0000 UTC m=+3308.816505392" lastFinishedPulling="2025-12-06 04:35:19.405840925 +0000 UTC m=+3312.277750077" observedRunningTime="2025-12-06 04:35:20.004288248 +0000 UTC m=+3312.876197400" watchObservedRunningTime="2025-12-06 04:35:20.014353531 +0000 UTC m=+3312.886262693" Dec 06 04:35:24 crc kubenswrapper[4802]: I1206 04:35:24.574331 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2h8z5" Dec 06 04:35:24 crc kubenswrapper[4802]: I1206 04:35:24.574928 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2h8z5" Dec 06 04:35:24 crc kubenswrapper[4802]: I1206 04:35:24.620496 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2h8z5" Dec 06 04:35:25 crc kubenswrapper[4802]: I1206 04:35:25.109653 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2h8z5" Dec 06 04:35:25 crc kubenswrapper[4802]: I1206 04:35:25.167794 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2h8z5"] Dec 06 04:35:27 crc kubenswrapper[4802]: I1206 04:35:27.061530 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-2h8z5" podUID="bc3f763e-de8d-43bd-86e2-abceb87a373a" containerName="registry-server" containerID="cri-o://d2379b923f568e4ce4189958b91c013d2a1005997f0764309e9e50a8821b0db7" gracePeriod=2 Dec 06 04:35:27 crc kubenswrapper[4802]: I1206 04:35:27.692815 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2h8z5" Dec 06 04:35:27 crc kubenswrapper[4802]: I1206 04:35:27.777644 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc3f763e-de8d-43bd-86e2-abceb87a373a-utilities\") pod \"bc3f763e-de8d-43bd-86e2-abceb87a373a\" (UID: \"bc3f763e-de8d-43bd-86e2-abceb87a373a\") " Dec 06 04:35:27 crc kubenswrapper[4802]: I1206 04:35:27.777744 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc3f763e-de8d-43bd-86e2-abceb87a373a-catalog-content\") pod \"bc3f763e-de8d-43bd-86e2-abceb87a373a\" (UID: \"bc3f763e-de8d-43bd-86e2-abceb87a373a\") " Dec 06 04:35:27 crc kubenswrapper[4802]: I1206 04:35:27.777898 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xdl8k\" (UniqueName: \"kubernetes.io/projected/bc3f763e-de8d-43bd-86e2-abceb87a373a-kube-api-access-xdl8k\") pod \"bc3f763e-de8d-43bd-86e2-abceb87a373a\" (UID: \"bc3f763e-de8d-43bd-86e2-abceb87a373a\") " Dec 06 04:35:27 crc kubenswrapper[4802]: I1206 04:35:27.783888 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc3f763e-de8d-43bd-86e2-abceb87a373a-kube-api-access-xdl8k" (OuterVolumeSpecName: "kube-api-access-xdl8k") pod "bc3f763e-de8d-43bd-86e2-abceb87a373a" (UID: "bc3f763e-de8d-43bd-86e2-abceb87a373a"). InnerVolumeSpecName "kube-api-access-xdl8k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:35:27 crc kubenswrapper[4802]: I1206 04:35:27.790742 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc3f763e-de8d-43bd-86e2-abceb87a373a-utilities" (OuterVolumeSpecName: "utilities") pod "bc3f763e-de8d-43bd-86e2-abceb87a373a" (UID: "bc3f763e-de8d-43bd-86e2-abceb87a373a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:35:27 crc kubenswrapper[4802]: I1206 04:35:27.831023 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc3f763e-de8d-43bd-86e2-abceb87a373a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bc3f763e-de8d-43bd-86e2-abceb87a373a" (UID: "bc3f763e-de8d-43bd-86e2-abceb87a373a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:35:27 crc kubenswrapper[4802]: I1206 04:35:27.881340 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xdl8k\" (UniqueName: \"kubernetes.io/projected/bc3f763e-de8d-43bd-86e2-abceb87a373a-kube-api-access-xdl8k\") on node \"crc\" DevicePath \"\"" Dec 06 04:35:27 crc kubenswrapper[4802]: I1206 04:35:27.881450 4802 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc3f763e-de8d-43bd-86e2-abceb87a373a-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 04:35:27 crc kubenswrapper[4802]: I1206 04:35:27.881476 4802 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc3f763e-de8d-43bd-86e2-abceb87a373a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 04:35:28 crc kubenswrapper[4802]: I1206 04:35:28.072193 4802 generic.go:334] "Generic (PLEG): container finished" podID="bc3f763e-de8d-43bd-86e2-abceb87a373a" containerID="d2379b923f568e4ce4189958b91c013d2a1005997f0764309e9e50a8821b0db7" exitCode=0 Dec 06 04:35:28 crc kubenswrapper[4802]: I1206 04:35:28.072247 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2h8z5" Dec 06 04:35:28 crc kubenswrapper[4802]: I1206 04:35:28.072251 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2h8z5" event={"ID":"bc3f763e-de8d-43bd-86e2-abceb87a373a","Type":"ContainerDied","Data":"d2379b923f568e4ce4189958b91c013d2a1005997f0764309e9e50a8821b0db7"} Dec 06 04:35:28 crc kubenswrapper[4802]: I1206 04:35:28.072367 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2h8z5" event={"ID":"bc3f763e-de8d-43bd-86e2-abceb87a373a","Type":"ContainerDied","Data":"33483fc412e8b08044486833b481861bec497cdf9a5ef1db9745c66cc3267191"} Dec 06 04:35:28 crc kubenswrapper[4802]: I1206 04:35:28.072399 4802 scope.go:117] "RemoveContainer" containerID="d2379b923f568e4ce4189958b91c013d2a1005997f0764309e9e50a8821b0db7" Dec 06 04:35:28 crc kubenswrapper[4802]: I1206 04:35:28.109851 4802 scope.go:117] "RemoveContainer" containerID="76a730687b641ded76c1a205f8f420dfa1576ba402ec0610643d841f206faf89" Dec 06 04:35:28 crc kubenswrapper[4802]: I1206 04:35:28.112996 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2h8z5"] Dec 06 04:35:28 crc kubenswrapper[4802]: I1206 04:35:28.121802 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-2h8z5"] Dec 06 04:35:28 crc kubenswrapper[4802]: I1206 04:35:28.134986 4802 scope.go:117] "RemoveContainer" containerID="e3adcff27252a8ef1f3290e5e7ce1967650cb0c56bc29c5332bfce6ecb6089a1" Dec 06 04:35:28 crc kubenswrapper[4802]: I1206 04:35:28.193856 4802 scope.go:117] "RemoveContainer" containerID="d2379b923f568e4ce4189958b91c013d2a1005997f0764309e9e50a8821b0db7" Dec 06 04:35:28 crc kubenswrapper[4802]: E1206 04:35:28.194419 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2379b923f568e4ce4189958b91c013d2a1005997f0764309e9e50a8821b0db7\": container with ID starting with d2379b923f568e4ce4189958b91c013d2a1005997f0764309e9e50a8821b0db7 not found: ID does not exist" containerID="d2379b923f568e4ce4189958b91c013d2a1005997f0764309e9e50a8821b0db7" Dec 06 04:35:28 crc kubenswrapper[4802]: I1206 04:35:28.194461 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2379b923f568e4ce4189958b91c013d2a1005997f0764309e9e50a8821b0db7"} err="failed to get container status \"d2379b923f568e4ce4189958b91c013d2a1005997f0764309e9e50a8821b0db7\": rpc error: code = NotFound desc = could not find container \"d2379b923f568e4ce4189958b91c013d2a1005997f0764309e9e50a8821b0db7\": container with ID starting with d2379b923f568e4ce4189958b91c013d2a1005997f0764309e9e50a8821b0db7 not found: ID does not exist" Dec 06 04:35:28 crc kubenswrapper[4802]: I1206 04:35:28.194488 4802 scope.go:117] "RemoveContainer" containerID="76a730687b641ded76c1a205f8f420dfa1576ba402ec0610643d841f206faf89" Dec 06 04:35:28 crc kubenswrapper[4802]: E1206 04:35:28.194858 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76a730687b641ded76c1a205f8f420dfa1576ba402ec0610643d841f206faf89\": container with ID starting with 76a730687b641ded76c1a205f8f420dfa1576ba402ec0610643d841f206faf89 not found: ID does not exist" containerID="76a730687b641ded76c1a205f8f420dfa1576ba402ec0610643d841f206faf89" Dec 06 04:35:28 crc kubenswrapper[4802]: I1206 04:35:28.194891 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76a730687b641ded76c1a205f8f420dfa1576ba402ec0610643d841f206faf89"} err="failed to get container status \"76a730687b641ded76c1a205f8f420dfa1576ba402ec0610643d841f206faf89\": rpc error: code = NotFound desc = could not find container \"76a730687b641ded76c1a205f8f420dfa1576ba402ec0610643d841f206faf89\": container with ID starting with 76a730687b641ded76c1a205f8f420dfa1576ba402ec0610643d841f206faf89 not found: ID does not exist" Dec 06 04:35:28 crc kubenswrapper[4802]: I1206 04:35:28.194909 4802 scope.go:117] "RemoveContainer" containerID="e3adcff27252a8ef1f3290e5e7ce1967650cb0c56bc29c5332bfce6ecb6089a1" Dec 06 04:35:28 crc kubenswrapper[4802]: E1206 04:35:28.195216 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3adcff27252a8ef1f3290e5e7ce1967650cb0c56bc29c5332bfce6ecb6089a1\": container with ID starting with e3adcff27252a8ef1f3290e5e7ce1967650cb0c56bc29c5332bfce6ecb6089a1 not found: ID does not exist" containerID="e3adcff27252a8ef1f3290e5e7ce1967650cb0c56bc29c5332bfce6ecb6089a1" Dec 06 04:35:28 crc kubenswrapper[4802]: I1206 04:35:28.195255 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3adcff27252a8ef1f3290e5e7ce1967650cb0c56bc29c5332bfce6ecb6089a1"} err="failed to get container status \"e3adcff27252a8ef1f3290e5e7ce1967650cb0c56bc29c5332bfce6ecb6089a1\": rpc error: code = NotFound desc = could not find container \"e3adcff27252a8ef1f3290e5e7ce1967650cb0c56bc29c5332bfce6ecb6089a1\": container with ID starting with e3adcff27252a8ef1f3290e5e7ce1967650cb0c56bc29c5332bfce6ecb6089a1 not found: ID does not exist" Dec 06 04:35:29 crc kubenswrapper[4802]: I1206 04:35:29.468401 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc3f763e-de8d-43bd-86e2-abceb87a373a" path="/var/lib/kubelet/pods/bc3f763e-de8d-43bd-86e2-abceb87a373a/volumes" Dec 06 04:35:50 crc kubenswrapper[4802]: I1206 04:35:50.469586 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-ldgt4"] Dec 06 04:35:50 crc kubenswrapper[4802]: E1206 04:35:50.472314 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc3f763e-de8d-43bd-86e2-abceb87a373a" containerName="extract-utilities" Dec 06 04:35:50 crc kubenswrapper[4802]: I1206 04:35:50.472330 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc3f763e-de8d-43bd-86e2-abceb87a373a" containerName="extract-utilities" Dec 06 04:35:50 crc kubenswrapper[4802]: E1206 04:35:50.472347 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc3f763e-de8d-43bd-86e2-abceb87a373a" containerName="registry-server" Dec 06 04:35:50 crc kubenswrapper[4802]: I1206 04:35:50.472353 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc3f763e-de8d-43bd-86e2-abceb87a373a" containerName="registry-server" Dec 06 04:35:50 crc kubenswrapper[4802]: E1206 04:35:50.472396 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc3f763e-de8d-43bd-86e2-abceb87a373a" containerName="extract-content" Dec 06 04:35:50 crc kubenswrapper[4802]: I1206 04:35:50.472403 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc3f763e-de8d-43bd-86e2-abceb87a373a" containerName="extract-content" Dec 06 04:35:50 crc kubenswrapper[4802]: I1206 04:35:50.472605 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc3f763e-de8d-43bd-86e2-abceb87a373a" containerName="registry-server" Dec 06 04:35:50 crc kubenswrapper[4802]: I1206 04:35:50.474013 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ldgt4" Dec 06 04:35:50 crc kubenswrapper[4802]: I1206 04:35:50.490620 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ldgt4"] Dec 06 04:35:50 crc kubenswrapper[4802]: I1206 04:35:50.589598 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9d2p\" (UniqueName: \"kubernetes.io/projected/29b4b708-0ccc-4a02-b18d-ea90aef664b3-kube-api-access-g9d2p\") pod \"redhat-operators-ldgt4\" (UID: \"29b4b708-0ccc-4a02-b18d-ea90aef664b3\") " pod="openshift-marketplace/redhat-operators-ldgt4" Dec 06 04:35:50 crc kubenswrapper[4802]: I1206 04:35:50.589723 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29b4b708-0ccc-4a02-b18d-ea90aef664b3-catalog-content\") pod \"redhat-operators-ldgt4\" (UID: \"29b4b708-0ccc-4a02-b18d-ea90aef664b3\") " pod="openshift-marketplace/redhat-operators-ldgt4" Dec 06 04:35:50 crc kubenswrapper[4802]: I1206 04:35:50.590019 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29b4b708-0ccc-4a02-b18d-ea90aef664b3-utilities\") pod \"redhat-operators-ldgt4\" (UID: \"29b4b708-0ccc-4a02-b18d-ea90aef664b3\") " pod="openshift-marketplace/redhat-operators-ldgt4" Dec 06 04:35:50 crc kubenswrapper[4802]: I1206 04:35:50.673517 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bnk58"] Dec 06 04:35:50 crc kubenswrapper[4802]: I1206 04:35:50.676513 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bnk58" Dec 06 04:35:50 crc kubenswrapper[4802]: I1206 04:35:50.692116 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bnk58"] Dec 06 04:35:50 crc kubenswrapper[4802]: I1206 04:35:50.726852 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9d2p\" (UniqueName: \"kubernetes.io/projected/29b4b708-0ccc-4a02-b18d-ea90aef664b3-kube-api-access-g9d2p\") pod \"redhat-operators-ldgt4\" (UID: \"29b4b708-0ccc-4a02-b18d-ea90aef664b3\") " pod="openshift-marketplace/redhat-operators-ldgt4" Dec 06 04:35:50 crc kubenswrapper[4802]: I1206 04:35:50.726932 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29b4b708-0ccc-4a02-b18d-ea90aef664b3-catalog-content\") pod \"redhat-operators-ldgt4\" (UID: \"29b4b708-0ccc-4a02-b18d-ea90aef664b3\") " pod="openshift-marketplace/redhat-operators-ldgt4" Dec 06 04:35:50 crc kubenswrapper[4802]: I1206 04:35:50.727039 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29b4b708-0ccc-4a02-b18d-ea90aef664b3-utilities\") pod \"redhat-operators-ldgt4\" (UID: \"29b4b708-0ccc-4a02-b18d-ea90aef664b3\") " pod="openshift-marketplace/redhat-operators-ldgt4" Dec 06 04:35:50 crc kubenswrapper[4802]: I1206 04:35:50.727842 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29b4b708-0ccc-4a02-b18d-ea90aef664b3-utilities\") pod \"redhat-operators-ldgt4\" (UID: \"29b4b708-0ccc-4a02-b18d-ea90aef664b3\") " pod="openshift-marketplace/redhat-operators-ldgt4" Dec 06 04:35:50 crc kubenswrapper[4802]: I1206 04:35:50.727867 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29b4b708-0ccc-4a02-b18d-ea90aef664b3-catalog-content\") pod \"redhat-operators-ldgt4\" (UID: \"29b4b708-0ccc-4a02-b18d-ea90aef664b3\") " pod="openshift-marketplace/redhat-operators-ldgt4" Dec 06 04:35:50 crc kubenswrapper[4802]: I1206 04:35:50.758597 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9d2p\" (UniqueName: \"kubernetes.io/projected/29b4b708-0ccc-4a02-b18d-ea90aef664b3-kube-api-access-g9d2p\") pod \"redhat-operators-ldgt4\" (UID: \"29b4b708-0ccc-4a02-b18d-ea90aef664b3\") " pod="openshift-marketplace/redhat-operators-ldgt4" Dec 06 04:35:50 crc kubenswrapper[4802]: I1206 04:35:50.814712 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ldgt4" Dec 06 04:35:50 crc kubenswrapper[4802]: I1206 04:35:50.828802 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74r2r\" (UniqueName: \"kubernetes.io/projected/dd042f6c-9dcd-4786-8673-b8b932fdcaea-kube-api-access-74r2r\") pod \"redhat-marketplace-bnk58\" (UID: \"dd042f6c-9dcd-4786-8673-b8b932fdcaea\") " pod="openshift-marketplace/redhat-marketplace-bnk58" Dec 06 04:35:50 crc kubenswrapper[4802]: I1206 04:35:50.828914 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd042f6c-9dcd-4786-8673-b8b932fdcaea-utilities\") pod \"redhat-marketplace-bnk58\" (UID: \"dd042f6c-9dcd-4786-8673-b8b932fdcaea\") " pod="openshift-marketplace/redhat-marketplace-bnk58" Dec 06 04:35:50 crc kubenswrapper[4802]: I1206 04:35:50.829008 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd042f6c-9dcd-4786-8673-b8b932fdcaea-catalog-content\") pod \"redhat-marketplace-bnk58\" (UID: \"dd042f6c-9dcd-4786-8673-b8b932fdcaea\") " pod="openshift-marketplace/redhat-marketplace-bnk58" Dec 06 04:35:50 crc kubenswrapper[4802]: I1206 04:35:50.930433 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74r2r\" (UniqueName: \"kubernetes.io/projected/dd042f6c-9dcd-4786-8673-b8b932fdcaea-kube-api-access-74r2r\") pod \"redhat-marketplace-bnk58\" (UID: \"dd042f6c-9dcd-4786-8673-b8b932fdcaea\") " pod="openshift-marketplace/redhat-marketplace-bnk58" Dec 06 04:35:50 crc kubenswrapper[4802]: I1206 04:35:50.930497 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd042f6c-9dcd-4786-8673-b8b932fdcaea-utilities\") pod \"redhat-marketplace-bnk58\" (UID: \"dd042f6c-9dcd-4786-8673-b8b932fdcaea\") " pod="openshift-marketplace/redhat-marketplace-bnk58" Dec 06 04:35:50 crc kubenswrapper[4802]: I1206 04:35:50.930565 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd042f6c-9dcd-4786-8673-b8b932fdcaea-catalog-content\") pod \"redhat-marketplace-bnk58\" (UID: \"dd042f6c-9dcd-4786-8673-b8b932fdcaea\") " pod="openshift-marketplace/redhat-marketplace-bnk58" Dec 06 04:35:50 crc kubenswrapper[4802]: I1206 04:35:50.931273 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd042f6c-9dcd-4786-8673-b8b932fdcaea-catalog-content\") pod \"redhat-marketplace-bnk58\" (UID: \"dd042f6c-9dcd-4786-8673-b8b932fdcaea\") " pod="openshift-marketplace/redhat-marketplace-bnk58" Dec 06 04:35:50 crc kubenswrapper[4802]: I1206 04:35:50.932006 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd042f6c-9dcd-4786-8673-b8b932fdcaea-utilities\") pod \"redhat-marketplace-bnk58\" (UID: \"dd042f6c-9dcd-4786-8673-b8b932fdcaea\") " pod="openshift-marketplace/redhat-marketplace-bnk58" Dec 06 04:35:50 crc kubenswrapper[4802]: I1206 04:35:50.960484 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74r2r\" (UniqueName: \"kubernetes.io/projected/dd042f6c-9dcd-4786-8673-b8b932fdcaea-kube-api-access-74r2r\") pod \"redhat-marketplace-bnk58\" (UID: \"dd042f6c-9dcd-4786-8673-b8b932fdcaea\") " pod="openshift-marketplace/redhat-marketplace-bnk58" Dec 06 04:35:51 crc kubenswrapper[4802]: I1206 04:35:51.039877 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bnk58" Dec 06 04:35:51 crc kubenswrapper[4802]: I1206 04:35:51.293060 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ldgt4"] Dec 06 04:35:51 crc kubenswrapper[4802]: I1206 04:35:51.335878 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ldgt4" event={"ID":"29b4b708-0ccc-4a02-b18d-ea90aef664b3","Type":"ContainerStarted","Data":"728f1fbbaedaa0a95e56ce916f78a2140b00055d49cf0f366393453c35c76545"} Dec 06 04:35:51 crc kubenswrapper[4802]: I1206 04:35:51.528849 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bnk58"] Dec 06 04:35:51 crc kubenswrapper[4802]: W1206 04:35:51.531467 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddd042f6c_9dcd_4786_8673_b8b932fdcaea.slice/crio-45e70adbda1740382d06d7220ee9798787a343c02c9cd966cc5cd681098a0671 WatchSource:0}: Error finding container 45e70adbda1740382d06d7220ee9798787a343c02c9cd966cc5cd681098a0671: Status 404 returned error can't find the container with id 45e70adbda1740382d06d7220ee9798787a343c02c9cd966cc5cd681098a0671 Dec 06 04:35:52 crc kubenswrapper[4802]: I1206 04:35:52.346473 4802 generic.go:334] "Generic (PLEG): container finished" podID="29b4b708-0ccc-4a02-b18d-ea90aef664b3" containerID="16b5767f1944f4ee90b288665bf0776adbb40c696639d7d3b7ea18c3e6c0ffc4" exitCode=0 Dec 06 04:35:52 crc kubenswrapper[4802]: I1206 04:35:52.346639 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ldgt4" event={"ID":"29b4b708-0ccc-4a02-b18d-ea90aef664b3","Type":"ContainerDied","Data":"16b5767f1944f4ee90b288665bf0776adbb40c696639d7d3b7ea18c3e6c0ffc4"} Dec 06 04:35:52 crc kubenswrapper[4802]: I1206 04:35:52.349186 4802 generic.go:334] "Generic (PLEG): container finished" podID="dd042f6c-9dcd-4786-8673-b8b932fdcaea" containerID="e7c86f07811085d9ccdc21be783119f2e4bf40b60d3d5a0faa398fefcd43076b" exitCode=0 Dec 06 04:35:52 crc kubenswrapper[4802]: I1206 04:35:52.349218 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bnk58" event={"ID":"dd042f6c-9dcd-4786-8673-b8b932fdcaea","Type":"ContainerDied","Data":"e7c86f07811085d9ccdc21be783119f2e4bf40b60d3d5a0faa398fefcd43076b"} Dec 06 04:35:52 crc kubenswrapper[4802]: I1206 04:35:52.349251 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bnk58" event={"ID":"dd042f6c-9dcd-4786-8673-b8b932fdcaea","Type":"ContainerStarted","Data":"45e70adbda1740382d06d7220ee9798787a343c02c9cd966cc5cd681098a0671"} Dec 06 04:35:54 crc kubenswrapper[4802]: I1206 04:35:54.368792 4802 generic.go:334] "Generic (PLEG): container finished" podID="dd042f6c-9dcd-4786-8673-b8b932fdcaea" containerID="4526e8d9bdabf81afe332b1f7b5d815ae4beb81e63b4d7a8928779003a5aac91" exitCode=0 Dec 06 04:35:54 crc kubenswrapper[4802]: I1206 04:35:54.368897 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bnk58" event={"ID":"dd042f6c-9dcd-4786-8673-b8b932fdcaea","Type":"ContainerDied","Data":"4526e8d9bdabf81afe332b1f7b5d815ae4beb81e63b4d7a8928779003a5aac91"} Dec 06 04:35:54 crc kubenswrapper[4802]: I1206 04:35:54.372561 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ldgt4" event={"ID":"29b4b708-0ccc-4a02-b18d-ea90aef664b3","Type":"ContainerStarted","Data":"1da0aa045f0e4bcc3ee8b9e2c215f4571e503f545cbb7ff7d67ab5d7dc975b15"} Dec 06 04:35:58 crc kubenswrapper[4802]: I1206 04:35:58.414450 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bnk58" event={"ID":"dd042f6c-9dcd-4786-8673-b8b932fdcaea","Type":"ContainerStarted","Data":"34d8c80e4a4c1c7fa4f1c83ad0ad31f30b10955fd2790c8c61ca004804ab5920"} Dec 06 04:35:58 crc kubenswrapper[4802]: I1206 04:35:58.439686 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bnk58" podStartSLOduration=3.707295135 podStartE2EDuration="8.439668162s" podCreationTimestamp="2025-12-06 04:35:50 +0000 UTC" firstStartedPulling="2025-12-06 04:35:52.350832815 +0000 UTC m=+3345.222741967" lastFinishedPulling="2025-12-06 04:35:57.083205842 +0000 UTC m=+3349.955114994" observedRunningTime="2025-12-06 04:35:58.435116589 +0000 UTC m=+3351.307025761" watchObservedRunningTime="2025-12-06 04:35:58.439668162 +0000 UTC m=+3351.311577314" Dec 06 04:35:59 crc kubenswrapper[4802]: I1206 04:35:59.422777 4802 generic.go:334] "Generic (PLEG): container finished" podID="29b4b708-0ccc-4a02-b18d-ea90aef664b3" containerID="1da0aa045f0e4bcc3ee8b9e2c215f4571e503f545cbb7ff7d67ab5d7dc975b15" exitCode=0 Dec 06 04:35:59 crc kubenswrapper[4802]: I1206 04:35:59.422783 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ldgt4" event={"ID":"29b4b708-0ccc-4a02-b18d-ea90aef664b3","Type":"ContainerDied","Data":"1da0aa045f0e4bcc3ee8b9e2c215f4571e503f545cbb7ff7d67ab5d7dc975b15"} Dec 06 04:36:00 crc kubenswrapper[4802]: I1206 04:36:00.434759 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ldgt4" event={"ID":"29b4b708-0ccc-4a02-b18d-ea90aef664b3","Type":"ContainerStarted","Data":"834f48dc64512c3e83af7401f9854baed1ac90aad1585293e21e5fc364fccc92"} Dec 06 04:36:00 crc kubenswrapper[4802]: I1206 04:36:00.459185 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-ldgt4" podStartSLOduration=2.988306907 podStartE2EDuration="10.45916438s" podCreationTimestamp="2025-12-06 04:35:50 +0000 UTC" firstStartedPulling="2025-12-06 04:35:52.348921272 +0000 UTC m=+3345.220830424" lastFinishedPulling="2025-12-06 04:35:59.819778745 +0000 UTC m=+3352.691687897" observedRunningTime="2025-12-06 04:36:00.456032155 +0000 UTC m=+3353.327941317" watchObservedRunningTime="2025-12-06 04:36:00.45916438 +0000 UTC m=+3353.331073532" Dec 06 04:36:00 crc kubenswrapper[4802]: I1206 04:36:00.815228 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-ldgt4" Dec 06 04:36:00 crc kubenswrapper[4802]: I1206 04:36:00.815598 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-ldgt4" Dec 06 04:36:01 crc kubenswrapper[4802]: I1206 04:36:01.040310 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bnk58" Dec 06 04:36:01 crc kubenswrapper[4802]: I1206 04:36:01.040374 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bnk58" Dec 06 04:36:01 crc kubenswrapper[4802]: I1206 04:36:01.886214 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-ldgt4" podUID="29b4b708-0ccc-4a02-b18d-ea90aef664b3" containerName="registry-server" probeResult="failure" output=< Dec 06 04:36:01 crc kubenswrapper[4802]: timeout: failed to connect service ":50051" within 1s Dec 06 04:36:01 crc kubenswrapper[4802]: > Dec 06 04:36:02 crc kubenswrapper[4802]: I1206 04:36:02.107623 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-bnk58" podUID="dd042f6c-9dcd-4786-8673-b8b932fdcaea" containerName="registry-server" probeResult="failure" output=< Dec 06 04:36:02 crc kubenswrapper[4802]: timeout: failed to connect service ":50051" within 1s Dec 06 04:36:02 crc kubenswrapper[4802]: > Dec 06 04:36:10 crc kubenswrapper[4802]: I1206 04:36:10.867787 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-ldgt4" Dec 06 04:36:10 crc kubenswrapper[4802]: I1206 04:36:10.942009 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-ldgt4" Dec 06 04:36:11 crc kubenswrapper[4802]: I1206 04:36:11.095306 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bnk58" Dec 06 04:36:11 crc kubenswrapper[4802]: I1206 04:36:11.106898 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ldgt4"] Dec 06 04:36:11 crc kubenswrapper[4802]: I1206 04:36:11.158647 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bnk58" Dec 06 04:36:12 crc kubenswrapper[4802]: I1206 04:36:12.546120 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-ldgt4" podUID="29b4b708-0ccc-4a02-b18d-ea90aef664b3" containerName="registry-server" containerID="cri-o://834f48dc64512c3e83af7401f9854baed1ac90aad1585293e21e5fc364fccc92" gracePeriod=2 Dec 06 04:36:13 crc kubenswrapper[4802]: I1206 04:36:13.122462 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ldgt4" Dec 06 04:36:13 crc kubenswrapper[4802]: I1206 04:36:13.279368 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29b4b708-0ccc-4a02-b18d-ea90aef664b3-utilities\") pod \"29b4b708-0ccc-4a02-b18d-ea90aef664b3\" (UID: \"29b4b708-0ccc-4a02-b18d-ea90aef664b3\") " Dec 06 04:36:13 crc kubenswrapper[4802]: I1206 04:36:13.279785 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g9d2p\" (UniqueName: \"kubernetes.io/projected/29b4b708-0ccc-4a02-b18d-ea90aef664b3-kube-api-access-g9d2p\") pod \"29b4b708-0ccc-4a02-b18d-ea90aef664b3\" (UID: \"29b4b708-0ccc-4a02-b18d-ea90aef664b3\") " Dec 06 04:36:13 crc kubenswrapper[4802]: I1206 04:36:13.279869 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29b4b708-0ccc-4a02-b18d-ea90aef664b3-catalog-content\") pod \"29b4b708-0ccc-4a02-b18d-ea90aef664b3\" (UID: \"29b4b708-0ccc-4a02-b18d-ea90aef664b3\") " Dec 06 04:36:13 crc kubenswrapper[4802]: I1206 04:36:13.280437 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/29b4b708-0ccc-4a02-b18d-ea90aef664b3-utilities" (OuterVolumeSpecName: "utilities") pod "29b4b708-0ccc-4a02-b18d-ea90aef664b3" (UID: "29b4b708-0ccc-4a02-b18d-ea90aef664b3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:36:13 crc kubenswrapper[4802]: I1206 04:36:13.280993 4802 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29b4b708-0ccc-4a02-b18d-ea90aef664b3-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 04:36:13 crc kubenswrapper[4802]: I1206 04:36:13.283304 4802 patch_prober.go:28] interesting pod/machine-config-daemon-zpxxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:36:13 crc kubenswrapper[4802]: I1206 04:36:13.283362 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:36:13 crc kubenswrapper[4802]: I1206 04:36:13.286024 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29b4b708-0ccc-4a02-b18d-ea90aef664b3-kube-api-access-g9d2p" (OuterVolumeSpecName: "kube-api-access-g9d2p") pod "29b4b708-0ccc-4a02-b18d-ea90aef664b3" (UID: "29b4b708-0ccc-4a02-b18d-ea90aef664b3"). InnerVolumeSpecName "kube-api-access-g9d2p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:36:13 crc kubenswrapper[4802]: I1206 04:36:13.383694 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g9d2p\" (UniqueName: \"kubernetes.io/projected/29b4b708-0ccc-4a02-b18d-ea90aef664b3-kube-api-access-g9d2p\") on node \"crc\" DevicePath \"\"" Dec 06 04:36:13 crc kubenswrapper[4802]: I1206 04:36:13.387482 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/29b4b708-0ccc-4a02-b18d-ea90aef664b3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "29b4b708-0ccc-4a02-b18d-ea90aef664b3" (UID: "29b4b708-0ccc-4a02-b18d-ea90aef664b3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:36:13 crc kubenswrapper[4802]: I1206 04:36:13.485836 4802 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29b4b708-0ccc-4a02-b18d-ea90aef664b3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 04:36:13 crc kubenswrapper[4802]: I1206 04:36:13.506554 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bnk58"] Dec 06 04:36:13 crc kubenswrapper[4802]: I1206 04:36:13.506798 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bnk58" podUID="dd042f6c-9dcd-4786-8673-b8b932fdcaea" containerName="registry-server" containerID="cri-o://34d8c80e4a4c1c7fa4f1c83ad0ad31f30b10955fd2790c8c61ca004804ab5920" gracePeriod=2 Dec 06 04:36:13 crc kubenswrapper[4802]: I1206 04:36:13.562452 4802 generic.go:334] "Generic (PLEG): container finished" podID="29b4b708-0ccc-4a02-b18d-ea90aef664b3" containerID="834f48dc64512c3e83af7401f9854baed1ac90aad1585293e21e5fc364fccc92" exitCode=0 Dec 06 04:36:13 crc kubenswrapper[4802]: I1206 04:36:13.562539 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ldgt4" event={"ID":"29b4b708-0ccc-4a02-b18d-ea90aef664b3","Type":"ContainerDied","Data":"834f48dc64512c3e83af7401f9854baed1ac90aad1585293e21e5fc364fccc92"} Dec 06 04:36:13 crc kubenswrapper[4802]: I1206 04:36:13.562589 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ldgt4" event={"ID":"29b4b708-0ccc-4a02-b18d-ea90aef664b3","Type":"ContainerDied","Data":"728f1fbbaedaa0a95e56ce916f78a2140b00055d49cf0f366393453c35c76545"} Dec 06 04:36:13 crc kubenswrapper[4802]: I1206 04:36:13.562614 4802 scope.go:117] "RemoveContainer" containerID="834f48dc64512c3e83af7401f9854baed1ac90aad1585293e21e5fc364fccc92" Dec 06 04:36:13 crc kubenswrapper[4802]: I1206 04:36:13.562539 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ldgt4" Dec 06 04:36:13 crc kubenswrapper[4802]: I1206 04:36:13.641313 4802 scope.go:117] "RemoveContainer" containerID="1da0aa045f0e4bcc3ee8b9e2c215f4571e503f545cbb7ff7d67ab5d7dc975b15" Dec 06 04:36:13 crc kubenswrapper[4802]: I1206 04:36:13.724469 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ldgt4"] Dec 06 04:36:13 crc kubenswrapper[4802]: I1206 04:36:13.734952 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-ldgt4"] Dec 06 04:36:13 crc kubenswrapper[4802]: I1206 04:36:13.740065 4802 scope.go:117] "RemoveContainer" containerID="16b5767f1944f4ee90b288665bf0776adbb40c696639d7d3b7ea18c3e6c0ffc4" Dec 06 04:36:13 crc kubenswrapper[4802]: I1206 04:36:13.769097 4802 scope.go:117] "RemoveContainer" containerID="834f48dc64512c3e83af7401f9854baed1ac90aad1585293e21e5fc364fccc92" Dec 06 04:36:13 crc kubenswrapper[4802]: E1206 04:36:13.770463 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"834f48dc64512c3e83af7401f9854baed1ac90aad1585293e21e5fc364fccc92\": container with ID starting with 834f48dc64512c3e83af7401f9854baed1ac90aad1585293e21e5fc364fccc92 not found: ID does not exist" containerID="834f48dc64512c3e83af7401f9854baed1ac90aad1585293e21e5fc364fccc92" Dec 06 04:36:13 crc kubenswrapper[4802]: I1206 04:36:13.770485 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"834f48dc64512c3e83af7401f9854baed1ac90aad1585293e21e5fc364fccc92"} err="failed to get container status \"834f48dc64512c3e83af7401f9854baed1ac90aad1585293e21e5fc364fccc92\": rpc error: code = NotFound desc = could not find container \"834f48dc64512c3e83af7401f9854baed1ac90aad1585293e21e5fc364fccc92\": container with ID starting with 834f48dc64512c3e83af7401f9854baed1ac90aad1585293e21e5fc364fccc92 not found: ID does not exist" Dec 06 04:36:13 crc kubenswrapper[4802]: I1206 04:36:13.770504 4802 scope.go:117] "RemoveContainer" containerID="1da0aa045f0e4bcc3ee8b9e2c215f4571e503f545cbb7ff7d67ab5d7dc975b15" Dec 06 04:36:13 crc kubenswrapper[4802]: E1206 04:36:13.770765 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1da0aa045f0e4bcc3ee8b9e2c215f4571e503f545cbb7ff7d67ab5d7dc975b15\": container with ID starting with 1da0aa045f0e4bcc3ee8b9e2c215f4571e503f545cbb7ff7d67ab5d7dc975b15 not found: ID does not exist" containerID="1da0aa045f0e4bcc3ee8b9e2c215f4571e503f545cbb7ff7d67ab5d7dc975b15" Dec 06 04:36:13 crc kubenswrapper[4802]: I1206 04:36:13.770782 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1da0aa045f0e4bcc3ee8b9e2c215f4571e503f545cbb7ff7d67ab5d7dc975b15"} err="failed to get container status \"1da0aa045f0e4bcc3ee8b9e2c215f4571e503f545cbb7ff7d67ab5d7dc975b15\": rpc error: code = NotFound desc = could not find container \"1da0aa045f0e4bcc3ee8b9e2c215f4571e503f545cbb7ff7d67ab5d7dc975b15\": container with ID starting with 1da0aa045f0e4bcc3ee8b9e2c215f4571e503f545cbb7ff7d67ab5d7dc975b15 not found: ID does not exist" Dec 06 04:36:13 crc kubenswrapper[4802]: I1206 04:36:13.770799 4802 scope.go:117] "RemoveContainer" containerID="16b5767f1944f4ee90b288665bf0776adbb40c696639d7d3b7ea18c3e6c0ffc4" Dec 06 04:36:13 crc kubenswrapper[4802]: E1206 04:36:13.771029 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16b5767f1944f4ee90b288665bf0776adbb40c696639d7d3b7ea18c3e6c0ffc4\": container with ID starting with 16b5767f1944f4ee90b288665bf0776adbb40c696639d7d3b7ea18c3e6c0ffc4 not found: ID does not exist" containerID="16b5767f1944f4ee90b288665bf0776adbb40c696639d7d3b7ea18c3e6c0ffc4" Dec 06 04:36:13 crc kubenswrapper[4802]: I1206 04:36:13.771050 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16b5767f1944f4ee90b288665bf0776adbb40c696639d7d3b7ea18c3e6c0ffc4"} err="failed to get container status \"16b5767f1944f4ee90b288665bf0776adbb40c696639d7d3b7ea18c3e6c0ffc4\": rpc error: code = NotFound desc = could not find container \"16b5767f1944f4ee90b288665bf0776adbb40c696639d7d3b7ea18c3e6c0ffc4\": container with ID starting with 16b5767f1944f4ee90b288665bf0776adbb40c696639d7d3b7ea18c3e6c0ffc4 not found: ID does not exist" Dec 06 04:36:14 crc kubenswrapper[4802]: I1206 04:36:14.188101 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bnk58" Dec 06 04:36:14 crc kubenswrapper[4802]: I1206 04:36:14.299008 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd042f6c-9dcd-4786-8673-b8b932fdcaea-utilities\") pod \"dd042f6c-9dcd-4786-8673-b8b932fdcaea\" (UID: \"dd042f6c-9dcd-4786-8673-b8b932fdcaea\") " Dec 06 04:36:14 crc kubenswrapper[4802]: I1206 04:36:14.299134 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd042f6c-9dcd-4786-8673-b8b932fdcaea-catalog-content\") pod \"dd042f6c-9dcd-4786-8673-b8b932fdcaea\" (UID: \"dd042f6c-9dcd-4786-8673-b8b932fdcaea\") " Dec 06 04:36:14 crc kubenswrapper[4802]: I1206 04:36:14.299161 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-74r2r\" (UniqueName: \"kubernetes.io/projected/dd042f6c-9dcd-4786-8673-b8b932fdcaea-kube-api-access-74r2r\") pod \"dd042f6c-9dcd-4786-8673-b8b932fdcaea\" (UID: \"dd042f6c-9dcd-4786-8673-b8b932fdcaea\") " Dec 06 04:36:14 crc kubenswrapper[4802]: I1206 04:36:14.299844 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd042f6c-9dcd-4786-8673-b8b932fdcaea-utilities" (OuterVolumeSpecName: "utilities") pod "dd042f6c-9dcd-4786-8673-b8b932fdcaea" (UID: "dd042f6c-9dcd-4786-8673-b8b932fdcaea"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:36:14 crc kubenswrapper[4802]: I1206 04:36:14.305094 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd042f6c-9dcd-4786-8673-b8b932fdcaea-kube-api-access-74r2r" (OuterVolumeSpecName: "kube-api-access-74r2r") pod "dd042f6c-9dcd-4786-8673-b8b932fdcaea" (UID: "dd042f6c-9dcd-4786-8673-b8b932fdcaea"). InnerVolumeSpecName "kube-api-access-74r2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:36:14 crc kubenswrapper[4802]: I1206 04:36:14.318895 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd042f6c-9dcd-4786-8673-b8b932fdcaea-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dd042f6c-9dcd-4786-8673-b8b932fdcaea" (UID: "dd042f6c-9dcd-4786-8673-b8b932fdcaea"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:36:14 crc kubenswrapper[4802]: I1206 04:36:14.401779 4802 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd042f6c-9dcd-4786-8673-b8b932fdcaea-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 04:36:14 crc kubenswrapper[4802]: I1206 04:36:14.401997 4802 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd042f6c-9dcd-4786-8673-b8b932fdcaea-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 04:36:14 crc kubenswrapper[4802]: I1206 04:36:14.402082 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-74r2r\" (UniqueName: \"kubernetes.io/projected/dd042f6c-9dcd-4786-8673-b8b932fdcaea-kube-api-access-74r2r\") on node \"crc\" DevicePath \"\"" Dec 06 04:36:14 crc kubenswrapper[4802]: I1206 04:36:14.580961 4802 generic.go:334] "Generic (PLEG): container finished" podID="dd042f6c-9dcd-4786-8673-b8b932fdcaea" containerID="34d8c80e4a4c1c7fa4f1c83ad0ad31f30b10955fd2790c8c61ca004804ab5920" exitCode=0 Dec 06 04:36:14 crc kubenswrapper[4802]: I1206 04:36:14.581021 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bnk58" event={"ID":"dd042f6c-9dcd-4786-8673-b8b932fdcaea","Type":"ContainerDied","Data":"34d8c80e4a4c1c7fa4f1c83ad0ad31f30b10955fd2790c8c61ca004804ab5920"} Dec 06 04:36:14 crc kubenswrapper[4802]: I1206 04:36:14.581058 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bnk58" event={"ID":"dd042f6c-9dcd-4786-8673-b8b932fdcaea","Type":"ContainerDied","Data":"45e70adbda1740382d06d7220ee9798787a343c02c9cd966cc5cd681098a0671"} Dec 06 04:36:14 crc kubenswrapper[4802]: I1206 04:36:14.581067 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bnk58" Dec 06 04:36:14 crc kubenswrapper[4802]: I1206 04:36:14.581085 4802 scope.go:117] "RemoveContainer" containerID="34d8c80e4a4c1c7fa4f1c83ad0ad31f30b10955fd2790c8c61ca004804ab5920" Dec 06 04:36:14 crc kubenswrapper[4802]: I1206 04:36:14.615064 4802 scope.go:117] "RemoveContainer" containerID="4526e8d9bdabf81afe332b1f7b5d815ae4beb81e63b4d7a8928779003a5aac91" Dec 06 04:36:14 crc kubenswrapper[4802]: I1206 04:36:14.669179 4802 scope.go:117] "RemoveContainer" containerID="e7c86f07811085d9ccdc21be783119f2e4bf40b60d3d5a0faa398fefcd43076b" Dec 06 04:36:14 crc kubenswrapper[4802]: I1206 04:36:14.670416 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bnk58"] Dec 06 04:36:14 crc kubenswrapper[4802]: I1206 04:36:14.683017 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bnk58"] Dec 06 04:36:14 crc kubenswrapper[4802]: I1206 04:36:14.704269 4802 scope.go:117] "RemoveContainer" containerID="34d8c80e4a4c1c7fa4f1c83ad0ad31f30b10955fd2790c8c61ca004804ab5920" Dec 06 04:36:14 crc kubenswrapper[4802]: E1206 04:36:14.704833 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34d8c80e4a4c1c7fa4f1c83ad0ad31f30b10955fd2790c8c61ca004804ab5920\": container with ID starting with 34d8c80e4a4c1c7fa4f1c83ad0ad31f30b10955fd2790c8c61ca004804ab5920 not found: ID does not exist" containerID="34d8c80e4a4c1c7fa4f1c83ad0ad31f30b10955fd2790c8c61ca004804ab5920" Dec 06 04:36:14 crc kubenswrapper[4802]: I1206 04:36:14.704871 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34d8c80e4a4c1c7fa4f1c83ad0ad31f30b10955fd2790c8c61ca004804ab5920"} err="failed to get container status \"34d8c80e4a4c1c7fa4f1c83ad0ad31f30b10955fd2790c8c61ca004804ab5920\": rpc error: code = NotFound desc = could not find container \"34d8c80e4a4c1c7fa4f1c83ad0ad31f30b10955fd2790c8c61ca004804ab5920\": container with ID starting with 34d8c80e4a4c1c7fa4f1c83ad0ad31f30b10955fd2790c8c61ca004804ab5920 not found: ID does not exist" Dec 06 04:36:14 crc kubenswrapper[4802]: I1206 04:36:14.704899 4802 scope.go:117] "RemoveContainer" containerID="4526e8d9bdabf81afe332b1f7b5d815ae4beb81e63b4d7a8928779003a5aac91" Dec 06 04:36:14 crc kubenswrapper[4802]: E1206 04:36:14.705478 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4526e8d9bdabf81afe332b1f7b5d815ae4beb81e63b4d7a8928779003a5aac91\": container with ID starting with 4526e8d9bdabf81afe332b1f7b5d815ae4beb81e63b4d7a8928779003a5aac91 not found: ID does not exist" containerID="4526e8d9bdabf81afe332b1f7b5d815ae4beb81e63b4d7a8928779003a5aac91" Dec 06 04:36:14 crc kubenswrapper[4802]: I1206 04:36:14.705575 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4526e8d9bdabf81afe332b1f7b5d815ae4beb81e63b4d7a8928779003a5aac91"} err="failed to get container status \"4526e8d9bdabf81afe332b1f7b5d815ae4beb81e63b4d7a8928779003a5aac91\": rpc error: code = NotFound desc = could not find container \"4526e8d9bdabf81afe332b1f7b5d815ae4beb81e63b4d7a8928779003a5aac91\": container with ID starting with 4526e8d9bdabf81afe332b1f7b5d815ae4beb81e63b4d7a8928779003a5aac91 not found: ID does not exist" Dec 06 04:36:14 crc kubenswrapper[4802]: I1206 04:36:14.705637 4802 scope.go:117] "RemoveContainer" containerID="e7c86f07811085d9ccdc21be783119f2e4bf40b60d3d5a0faa398fefcd43076b" Dec 06 04:36:14 crc kubenswrapper[4802]: E1206 04:36:14.706167 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7c86f07811085d9ccdc21be783119f2e4bf40b60d3d5a0faa398fefcd43076b\": container with ID starting with e7c86f07811085d9ccdc21be783119f2e4bf40b60d3d5a0faa398fefcd43076b not found: ID does not exist" containerID="e7c86f07811085d9ccdc21be783119f2e4bf40b60d3d5a0faa398fefcd43076b" Dec 06 04:36:14 crc kubenswrapper[4802]: I1206 04:36:14.706197 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7c86f07811085d9ccdc21be783119f2e4bf40b60d3d5a0faa398fefcd43076b"} err="failed to get container status \"e7c86f07811085d9ccdc21be783119f2e4bf40b60d3d5a0faa398fefcd43076b\": rpc error: code = NotFound desc = could not find container \"e7c86f07811085d9ccdc21be783119f2e4bf40b60d3d5a0faa398fefcd43076b\": container with ID starting with e7c86f07811085d9ccdc21be783119f2e4bf40b60d3d5a0faa398fefcd43076b not found: ID does not exist" Dec 06 04:36:15 crc kubenswrapper[4802]: I1206 04:36:15.464307 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="29b4b708-0ccc-4a02-b18d-ea90aef664b3" path="/var/lib/kubelet/pods/29b4b708-0ccc-4a02-b18d-ea90aef664b3/volumes" Dec 06 04:36:15 crc kubenswrapper[4802]: I1206 04:36:15.465796 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd042f6c-9dcd-4786-8673-b8b932fdcaea" path="/var/lib/kubelet/pods/dd042f6c-9dcd-4786-8673-b8b932fdcaea/volumes" Dec 06 04:36:43 crc kubenswrapper[4802]: I1206 04:36:43.283200 4802 patch_prober.go:28] interesting pod/machine-config-daemon-zpxxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:36:43 crc kubenswrapper[4802]: I1206 04:36:43.284691 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:37:13 crc kubenswrapper[4802]: I1206 04:37:13.283107 4802 patch_prober.go:28] interesting pod/machine-config-daemon-zpxxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:37:13 crc kubenswrapper[4802]: I1206 04:37:13.283674 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:37:13 crc kubenswrapper[4802]: I1206 04:37:13.283721 4802 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" Dec 06 04:37:13 crc kubenswrapper[4802]: I1206 04:37:13.284492 4802 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"549c98317cf5beda6d72e30c3f31e49c51734f812384e19e7ca9ae6871d665af"} pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 04:37:13 crc kubenswrapper[4802]: I1206 04:37:13.284558 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" containerID="cri-o://549c98317cf5beda6d72e30c3f31e49c51734f812384e19e7ca9ae6871d665af" gracePeriod=600 Dec 06 04:37:14 crc kubenswrapper[4802]: I1206 04:37:14.138308 4802 generic.go:334] "Generic (PLEG): container finished" podID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerID="549c98317cf5beda6d72e30c3f31e49c51734f812384e19e7ca9ae6871d665af" exitCode=0 Dec 06 04:37:14 crc kubenswrapper[4802]: I1206 04:37:14.138449 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" event={"ID":"6df38316-e0d3-4018-8d27-3620eba3a68d","Type":"ContainerDied","Data":"549c98317cf5beda6d72e30c3f31e49c51734f812384e19e7ca9ae6871d665af"} Dec 06 04:37:14 crc kubenswrapper[4802]: I1206 04:37:14.138932 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" event={"ID":"6df38316-e0d3-4018-8d27-3620eba3a68d","Type":"ContainerStarted","Data":"3037e38e5a7009f353fa7ac5ea266cd2366825eeb99656721b87c6c25f49afdb"} Dec 06 04:37:14 crc kubenswrapper[4802]: I1206 04:37:14.138953 4802 scope.go:117] "RemoveContainer" containerID="ce5c9338b4ba535e43ce4e2eb6bfc97dd2ebe59199fb4ec623134cb1b6016b9c" Dec 06 04:38:57 crc kubenswrapper[4802]: I1206 04:38:57.719180 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-j7dgj"] Dec 06 04:38:57 crc kubenswrapper[4802]: E1206 04:38:57.720219 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd042f6c-9dcd-4786-8673-b8b932fdcaea" containerName="registry-server" Dec 06 04:38:57 crc kubenswrapper[4802]: I1206 04:38:57.720237 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd042f6c-9dcd-4786-8673-b8b932fdcaea" containerName="registry-server" Dec 06 04:38:57 crc kubenswrapper[4802]: E1206 04:38:57.720253 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29b4b708-0ccc-4a02-b18d-ea90aef664b3" containerName="registry-server" Dec 06 04:38:57 crc kubenswrapper[4802]: I1206 04:38:57.720261 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="29b4b708-0ccc-4a02-b18d-ea90aef664b3" containerName="registry-server" Dec 06 04:38:57 crc kubenswrapper[4802]: E1206 04:38:57.720287 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29b4b708-0ccc-4a02-b18d-ea90aef664b3" containerName="extract-content" Dec 06 04:38:57 crc kubenswrapper[4802]: I1206 04:38:57.720295 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="29b4b708-0ccc-4a02-b18d-ea90aef664b3" containerName="extract-content" Dec 06 04:38:57 crc kubenswrapper[4802]: E1206 04:38:57.720309 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd042f6c-9dcd-4786-8673-b8b932fdcaea" containerName="extract-content" Dec 06 04:38:57 crc kubenswrapper[4802]: I1206 04:38:57.720318 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd042f6c-9dcd-4786-8673-b8b932fdcaea" containerName="extract-content" Dec 06 04:38:57 crc kubenswrapper[4802]: E1206 04:38:57.720329 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29b4b708-0ccc-4a02-b18d-ea90aef664b3" containerName="extract-utilities" Dec 06 04:38:57 crc kubenswrapper[4802]: I1206 04:38:57.720338 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="29b4b708-0ccc-4a02-b18d-ea90aef664b3" containerName="extract-utilities" Dec 06 04:38:57 crc kubenswrapper[4802]: E1206 04:38:57.720355 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd042f6c-9dcd-4786-8673-b8b932fdcaea" containerName="extract-utilities" Dec 06 04:38:57 crc kubenswrapper[4802]: I1206 04:38:57.720363 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd042f6c-9dcd-4786-8673-b8b932fdcaea" containerName="extract-utilities" Dec 06 04:38:57 crc kubenswrapper[4802]: I1206 04:38:57.720598 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd042f6c-9dcd-4786-8673-b8b932fdcaea" containerName="registry-server" Dec 06 04:38:57 crc kubenswrapper[4802]: I1206 04:38:57.720637 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="29b4b708-0ccc-4a02-b18d-ea90aef664b3" containerName="registry-server" Dec 06 04:38:57 crc kubenswrapper[4802]: I1206 04:38:57.722450 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j7dgj" Dec 06 04:38:57 crc kubenswrapper[4802]: I1206 04:38:57.742476 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-j7dgj"] Dec 06 04:38:57 crc kubenswrapper[4802]: I1206 04:38:57.802868 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90937dee-5d01-4b56-9b2e-5a9c3cd78232-catalog-content\") pod \"community-operators-j7dgj\" (UID: \"90937dee-5d01-4b56-9b2e-5a9c3cd78232\") " pod="openshift-marketplace/community-operators-j7dgj" Dec 06 04:38:57 crc kubenswrapper[4802]: I1206 04:38:57.802930 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90937dee-5d01-4b56-9b2e-5a9c3cd78232-utilities\") pod \"community-operators-j7dgj\" (UID: \"90937dee-5d01-4b56-9b2e-5a9c3cd78232\") " pod="openshift-marketplace/community-operators-j7dgj" Dec 06 04:38:57 crc kubenswrapper[4802]: I1206 04:38:57.803013 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tk4qp\" (UniqueName: \"kubernetes.io/projected/90937dee-5d01-4b56-9b2e-5a9c3cd78232-kube-api-access-tk4qp\") pod \"community-operators-j7dgj\" (UID: \"90937dee-5d01-4b56-9b2e-5a9c3cd78232\") " pod="openshift-marketplace/community-operators-j7dgj" Dec 06 04:38:57 crc kubenswrapper[4802]: I1206 04:38:57.904575 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90937dee-5d01-4b56-9b2e-5a9c3cd78232-catalog-content\") pod \"community-operators-j7dgj\" (UID: \"90937dee-5d01-4b56-9b2e-5a9c3cd78232\") " pod="openshift-marketplace/community-operators-j7dgj" Dec 06 04:38:57 crc kubenswrapper[4802]: I1206 04:38:57.904879 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90937dee-5d01-4b56-9b2e-5a9c3cd78232-utilities\") pod \"community-operators-j7dgj\" (UID: \"90937dee-5d01-4b56-9b2e-5a9c3cd78232\") " pod="openshift-marketplace/community-operators-j7dgj" Dec 06 04:38:57 crc kubenswrapper[4802]: I1206 04:38:57.905016 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tk4qp\" (UniqueName: \"kubernetes.io/projected/90937dee-5d01-4b56-9b2e-5a9c3cd78232-kube-api-access-tk4qp\") pod \"community-operators-j7dgj\" (UID: \"90937dee-5d01-4b56-9b2e-5a9c3cd78232\") " pod="openshift-marketplace/community-operators-j7dgj" Dec 06 04:38:57 crc kubenswrapper[4802]: I1206 04:38:57.905143 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90937dee-5d01-4b56-9b2e-5a9c3cd78232-catalog-content\") pod \"community-operators-j7dgj\" (UID: \"90937dee-5d01-4b56-9b2e-5a9c3cd78232\") " pod="openshift-marketplace/community-operators-j7dgj" Dec 06 04:38:57 crc kubenswrapper[4802]: I1206 04:38:57.905296 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90937dee-5d01-4b56-9b2e-5a9c3cd78232-utilities\") pod \"community-operators-j7dgj\" (UID: \"90937dee-5d01-4b56-9b2e-5a9c3cd78232\") " pod="openshift-marketplace/community-operators-j7dgj" Dec 06 04:38:57 crc kubenswrapper[4802]: I1206 04:38:57.930451 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tk4qp\" (UniqueName: \"kubernetes.io/projected/90937dee-5d01-4b56-9b2e-5a9c3cd78232-kube-api-access-tk4qp\") pod \"community-operators-j7dgj\" (UID: \"90937dee-5d01-4b56-9b2e-5a9c3cd78232\") " pod="openshift-marketplace/community-operators-j7dgj" Dec 06 04:38:58 crc kubenswrapper[4802]: I1206 04:38:58.047547 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j7dgj" Dec 06 04:38:58 crc kubenswrapper[4802]: I1206 04:38:58.566239 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-j7dgj"] Dec 06 04:38:59 crc kubenswrapper[4802]: I1206 04:38:59.320007 4802 generic.go:334] "Generic (PLEG): container finished" podID="90937dee-5d01-4b56-9b2e-5a9c3cd78232" containerID="ed4a34fa3c169cbeea727adaa8c0b809f57c3489d4fd920b16dfd99c43e7c597" exitCode=0 Dec 06 04:38:59 crc kubenswrapper[4802]: I1206 04:38:59.320058 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j7dgj" event={"ID":"90937dee-5d01-4b56-9b2e-5a9c3cd78232","Type":"ContainerDied","Data":"ed4a34fa3c169cbeea727adaa8c0b809f57c3489d4fd920b16dfd99c43e7c597"} Dec 06 04:38:59 crc kubenswrapper[4802]: I1206 04:38:59.320392 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j7dgj" event={"ID":"90937dee-5d01-4b56-9b2e-5a9c3cd78232","Type":"ContainerStarted","Data":"56e08eeade658f3c9857d30e23561b1594b3be53165251a3311a611c8978b45d"} Dec 06 04:39:00 crc kubenswrapper[4802]: I1206 04:39:00.329773 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j7dgj" event={"ID":"90937dee-5d01-4b56-9b2e-5a9c3cd78232","Type":"ContainerStarted","Data":"c023ed625bbf537855ff7a4e3b9a59a7762b231c434dca6ace754431fc2dc07d"} Dec 06 04:39:01 crc kubenswrapper[4802]: I1206 04:39:01.342384 4802 generic.go:334] "Generic (PLEG): container finished" podID="90937dee-5d01-4b56-9b2e-5a9c3cd78232" containerID="c023ed625bbf537855ff7a4e3b9a59a7762b231c434dca6ace754431fc2dc07d" exitCode=0 Dec 06 04:39:01 crc kubenswrapper[4802]: I1206 04:39:01.342442 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j7dgj" event={"ID":"90937dee-5d01-4b56-9b2e-5a9c3cd78232","Type":"ContainerDied","Data":"c023ed625bbf537855ff7a4e3b9a59a7762b231c434dca6ace754431fc2dc07d"} Dec 06 04:39:02 crc kubenswrapper[4802]: I1206 04:39:02.351345 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j7dgj" event={"ID":"90937dee-5d01-4b56-9b2e-5a9c3cd78232","Type":"ContainerStarted","Data":"49b5cc49463125ed9872c056e35fde22b09e2cb3a98a49c6d388448c1e0ddc2a"} Dec 06 04:39:02 crc kubenswrapper[4802]: I1206 04:39:02.369455 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-j7dgj" podStartSLOduration=2.9065332440000002 podStartE2EDuration="5.369434127s" podCreationTimestamp="2025-12-06 04:38:57 +0000 UTC" firstStartedPulling="2025-12-06 04:38:59.322208203 +0000 UTC m=+3532.194117355" lastFinishedPulling="2025-12-06 04:39:01.785109086 +0000 UTC m=+3534.657018238" observedRunningTime="2025-12-06 04:39:02.366111657 +0000 UTC m=+3535.238020809" watchObservedRunningTime="2025-12-06 04:39:02.369434127 +0000 UTC m=+3535.241343279" Dec 06 04:39:08 crc kubenswrapper[4802]: I1206 04:39:08.048789 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-j7dgj" Dec 06 04:39:08 crc kubenswrapper[4802]: I1206 04:39:08.049335 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-j7dgj" Dec 06 04:39:08 crc kubenswrapper[4802]: I1206 04:39:08.117202 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-j7dgj" Dec 06 04:39:08 crc kubenswrapper[4802]: I1206 04:39:08.507348 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-j7dgj" Dec 06 04:39:08 crc kubenswrapper[4802]: I1206 04:39:08.622847 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-j7dgj"] Dec 06 04:39:10 crc kubenswrapper[4802]: I1206 04:39:10.428765 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-j7dgj" podUID="90937dee-5d01-4b56-9b2e-5a9c3cd78232" containerName="registry-server" containerID="cri-o://49b5cc49463125ed9872c056e35fde22b09e2cb3a98a49c6d388448c1e0ddc2a" gracePeriod=2 Dec 06 04:39:10 crc kubenswrapper[4802]: I1206 04:39:10.932536 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j7dgj" Dec 06 04:39:10 crc kubenswrapper[4802]: I1206 04:39:10.964281 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk4qp\" (UniqueName: \"kubernetes.io/projected/90937dee-5d01-4b56-9b2e-5a9c3cd78232-kube-api-access-tk4qp\") pod \"90937dee-5d01-4b56-9b2e-5a9c3cd78232\" (UID: \"90937dee-5d01-4b56-9b2e-5a9c3cd78232\") " Dec 06 04:39:10 crc kubenswrapper[4802]: I1206 04:39:10.964691 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90937dee-5d01-4b56-9b2e-5a9c3cd78232-catalog-content\") pod \"90937dee-5d01-4b56-9b2e-5a9c3cd78232\" (UID: \"90937dee-5d01-4b56-9b2e-5a9c3cd78232\") " Dec 06 04:39:10 crc kubenswrapper[4802]: I1206 04:39:10.964870 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90937dee-5d01-4b56-9b2e-5a9c3cd78232-utilities\") pod \"90937dee-5d01-4b56-9b2e-5a9c3cd78232\" (UID: \"90937dee-5d01-4b56-9b2e-5a9c3cd78232\") " Dec 06 04:39:10 crc kubenswrapper[4802]: I1206 04:39:10.967345 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90937dee-5d01-4b56-9b2e-5a9c3cd78232-utilities" (OuterVolumeSpecName: "utilities") pod "90937dee-5d01-4b56-9b2e-5a9c3cd78232" (UID: "90937dee-5d01-4b56-9b2e-5a9c3cd78232"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:39:10 crc kubenswrapper[4802]: I1206 04:39:10.972626 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90937dee-5d01-4b56-9b2e-5a9c3cd78232-kube-api-access-tk4qp" (OuterVolumeSpecName: "kube-api-access-tk4qp") pod "90937dee-5d01-4b56-9b2e-5a9c3cd78232" (UID: "90937dee-5d01-4b56-9b2e-5a9c3cd78232"). InnerVolumeSpecName "kube-api-access-tk4qp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:39:11 crc kubenswrapper[4802]: I1206 04:39:11.034515 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90937dee-5d01-4b56-9b2e-5a9c3cd78232-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "90937dee-5d01-4b56-9b2e-5a9c3cd78232" (UID: "90937dee-5d01-4b56-9b2e-5a9c3cd78232"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:39:11 crc kubenswrapper[4802]: I1206 04:39:11.066861 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk4qp\" (UniqueName: \"kubernetes.io/projected/90937dee-5d01-4b56-9b2e-5a9c3cd78232-kube-api-access-tk4qp\") on node \"crc\" DevicePath \"\"" Dec 06 04:39:11 crc kubenswrapper[4802]: I1206 04:39:11.066894 4802 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90937dee-5d01-4b56-9b2e-5a9c3cd78232-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 04:39:11 crc kubenswrapper[4802]: I1206 04:39:11.066904 4802 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90937dee-5d01-4b56-9b2e-5a9c3cd78232-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 04:39:11 crc kubenswrapper[4802]: I1206 04:39:11.439842 4802 generic.go:334] "Generic (PLEG): container finished" podID="90937dee-5d01-4b56-9b2e-5a9c3cd78232" containerID="49b5cc49463125ed9872c056e35fde22b09e2cb3a98a49c6d388448c1e0ddc2a" exitCode=0 Dec 06 04:39:11 crc kubenswrapper[4802]: I1206 04:39:11.439919 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j7dgj" event={"ID":"90937dee-5d01-4b56-9b2e-5a9c3cd78232","Type":"ContainerDied","Data":"49b5cc49463125ed9872c056e35fde22b09e2cb3a98a49c6d388448c1e0ddc2a"} Dec 06 04:39:11 crc kubenswrapper[4802]: I1206 04:39:11.439939 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j7dgj" Dec 06 04:39:11 crc kubenswrapper[4802]: I1206 04:39:11.440786 4802 scope.go:117] "RemoveContainer" containerID="49b5cc49463125ed9872c056e35fde22b09e2cb3a98a49c6d388448c1e0ddc2a" Dec 06 04:39:11 crc kubenswrapper[4802]: I1206 04:39:11.440728 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j7dgj" event={"ID":"90937dee-5d01-4b56-9b2e-5a9c3cd78232","Type":"ContainerDied","Data":"56e08eeade658f3c9857d30e23561b1594b3be53165251a3311a611c8978b45d"} Dec 06 04:39:11 crc kubenswrapper[4802]: I1206 04:39:11.464666 4802 scope.go:117] "RemoveContainer" containerID="c023ed625bbf537855ff7a4e3b9a59a7762b231c434dca6ace754431fc2dc07d" Dec 06 04:39:11 crc kubenswrapper[4802]: I1206 04:39:11.499223 4802 scope.go:117] "RemoveContainer" containerID="ed4a34fa3c169cbeea727adaa8c0b809f57c3489d4fd920b16dfd99c43e7c597" Dec 06 04:39:11 crc kubenswrapper[4802]: I1206 04:39:11.506309 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-j7dgj"] Dec 06 04:39:11 crc kubenswrapper[4802]: I1206 04:39:11.519652 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-j7dgj"] Dec 06 04:39:11 crc kubenswrapper[4802]: I1206 04:39:11.551049 4802 scope.go:117] "RemoveContainer" containerID="49b5cc49463125ed9872c056e35fde22b09e2cb3a98a49c6d388448c1e0ddc2a" Dec 06 04:39:11 crc kubenswrapper[4802]: E1206 04:39:11.551783 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49b5cc49463125ed9872c056e35fde22b09e2cb3a98a49c6d388448c1e0ddc2a\": container with ID starting with 49b5cc49463125ed9872c056e35fde22b09e2cb3a98a49c6d388448c1e0ddc2a not found: ID does not exist" containerID="49b5cc49463125ed9872c056e35fde22b09e2cb3a98a49c6d388448c1e0ddc2a" Dec 06 04:39:11 crc kubenswrapper[4802]: I1206 04:39:11.551827 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49b5cc49463125ed9872c056e35fde22b09e2cb3a98a49c6d388448c1e0ddc2a"} err="failed to get container status \"49b5cc49463125ed9872c056e35fde22b09e2cb3a98a49c6d388448c1e0ddc2a\": rpc error: code = NotFound desc = could not find container \"49b5cc49463125ed9872c056e35fde22b09e2cb3a98a49c6d388448c1e0ddc2a\": container with ID starting with 49b5cc49463125ed9872c056e35fde22b09e2cb3a98a49c6d388448c1e0ddc2a not found: ID does not exist" Dec 06 04:39:11 crc kubenswrapper[4802]: I1206 04:39:11.551862 4802 scope.go:117] "RemoveContainer" containerID="c023ed625bbf537855ff7a4e3b9a59a7762b231c434dca6ace754431fc2dc07d" Dec 06 04:39:11 crc kubenswrapper[4802]: E1206 04:39:11.552432 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c023ed625bbf537855ff7a4e3b9a59a7762b231c434dca6ace754431fc2dc07d\": container with ID starting with c023ed625bbf537855ff7a4e3b9a59a7762b231c434dca6ace754431fc2dc07d not found: ID does not exist" containerID="c023ed625bbf537855ff7a4e3b9a59a7762b231c434dca6ace754431fc2dc07d" Dec 06 04:39:11 crc kubenswrapper[4802]: I1206 04:39:11.552454 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c023ed625bbf537855ff7a4e3b9a59a7762b231c434dca6ace754431fc2dc07d"} err="failed to get container status \"c023ed625bbf537855ff7a4e3b9a59a7762b231c434dca6ace754431fc2dc07d\": rpc error: code = NotFound desc = could not find container \"c023ed625bbf537855ff7a4e3b9a59a7762b231c434dca6ace754431fc2dc07d\": container with ID starting with c023ed625bbf537855ff7a4e3b9a59a7762b231c434dca6ace754431fc2dc07d not found: ID does not exist" Dec 06 04:39:11 crc kubenswrapper[4802]: I1206 04:39:11.552469 4802 scope.go:117] "RemoveContainer" containerID="ed4a34fa3c169cbeea727adaa8c0b809f57c3489d4fd920b16dfd99c43e7c597" Dec 06 04:39:11 crc kubenswrapper[4802]: E1206 04:39:11.552872 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed4a34fa3c169cbeea727adaa8c0b809f57c3489d4fd920b16dfd99c43e7c597\": container with ID starting with ed4a34fa3c169cbeea727adaa8c0b809f57c3489d4fd920b16dfd99c43e7c597 not found: ID does not exist" containerID="ed4a34fa3c169cbeea727adaa8c0b809f57c3489d4fd920b16dfd99c43e7c597" Dec 06 04:39:11 crc kubenswrapper[4802]: I1206 04:39:11.552920 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed4a34fa3c169cbeea727adaa8c0b809f57c3489d4fd920b16dfd99c43e7c597"} err="failed to get container status \"ed4a34fa3c169cbeea727adaa8c0b809f57c3489d4fd920b16dfd99c43e7c597\": rpc error: code = NotFound desc = could not find container \"ed4a34fa3c169cbeea727adaa8c0b809f57c3489d4fd920b16dfd99c43e7c597\": container with ID starting with ed4a34fa3c169cbeea727adaa8c0b809f57c3489d4fd920b16dfd99c43e7c597 not found: ID does not exist" Dec 06 04:39:13 crc kubenswrapper[4802]: I1206 04:39:13.283993 4802 patch_prober.go:28] interesting pod/machine-config-daemon-zpxxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:39:13 crc kubenswrapper[4802]: I1206 04:39:13.284984 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:39:13 crc kubenswrapper[4802]: I1206 04:39:13.467440 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90937dee-5d01-4b56-9b2e-5a9c3cd78232" path="/var/lib/kubelet/pods/90937dee-5d01-4b56-9b2e-5a9c3cd78232/volumes" Dec 06 04:39:43 crc kubenswrapper[4802]: I1206 04:39:43.283027 4802 patch_prober.go:28] interesting pod/machine-config-daemon-zpxxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:39:43 crc kubenswrapper[4802]: I1206 04:39:43.283807 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:40:13 crc kubenswrapper[4802]: I1206 04:40:13.283675 4802 patch_prober.go:28] interesting pod/machine-config-daemon-zpxxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:40:13 crc kubenswrapper[4802]: I1206 04:40:13.284696 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:40:13 crc kubenswrapper[4802]: I1206 04:40:13.284804 4802 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" Dec 06 04:40:13 crc kubenswrapper[4802]: I1206 04:40:13.286009 4802 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3037e38e5a7009f353fa7ac5ea266cd2366825eeb99656721b87c6c25f49afdb"} pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 04:40:13 crc kubenswrapper[4802]: I1206 04:40:13.286080 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" containerID="cri-o://3037e38e5a7009f353fa7ac5ea266cd2366825eeb99656721b87c6c25f49afdb" gracePeriod=600 Dec 06 04:40:13 crc kubenswrapper[4802]: E1206 04:40:13.409984 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:40:14 crc kubenswrapper[4802]: I1206 04:40:14.099769 4802 generic.go:334] "Generic (PLEG): container finished" podID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerID="3037e38e5a7009f353fa7ac5ea266cd2366825eeb99656721b87c6c25f49afdb" exitCode=0 Dec 06 04:40:14 crc kubenswrapper[4802]: I1206 04:40:14.099818 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" event={"ID":"6df38316-e0d3-4018-8d27-3620eba3a68d","Type":"ContainerDied","Data":"3037e38e5a7009f353fa7ac5ea266cd2366825eeb99656721b87c6c25f49afdb"} Dec 06 04:40:14 crc kubenswrapper[4802]: I1206 04:40:14.099890 4802 scope.go:117] "RemoveContainer" containerID="549c98317cf5beda6d72e30c3f31e49c51734f812384e19e7ca9ae6871d665af" Dec 06 04:40:14 crc kubenswrapper[4802]: I1206 04:40:14.100735 4802 scope.go:117] "RemoveContainer" containerID="3037e38e5a7009f353fa7ac5ea266cd2366825eeb99656721b87c6c25f49afdb" Dec 06 04:40:14 crc kubenswrapper[4802]: E1206 04:40:14.101077 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:40:29 crc kubenswrapper[4802]: I1206 04:40:29.449956 4802 scope.go:117] "RemoveContainer" containerID="3037e38e5a7009f353fa7ac5ea266cd2366825eeb99656721b87c6c25f49afdb" Dec 06 04:40:29 crc kubenswrapper[4802]: E1206 04:40:29.451131 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:40:40 crc kubenswrapper[4802]: I1206 04:40:40.450705 4802 scope.go:117] "RemoveContainer" containerID="3037e38e5a7009f353fa7ac5ea266cd2366825eeb99656721b87c6c25f49afdb" Dec 06 04:40:40 crc kubenswrapper[4802]: E1206 04:40:40.452009 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:40:53 crc kubenswrapper[4802]: I1206 04:40:53.450545 4802 scope.go:117] "RemoveContainer" containerID="3037e38e5a7009f353fa7ac5ea266cd2366825eeb99656721b87c6c25f49afdb" Dec 06 04:40:53 crc kubenswrapper[4802]: E1206 04:40:53.451206 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:41:05 crc kubenswrapper[4802]: I1206 04:41:05.450871 4802 scope.go:117] "RemoveContainer" containerID="3037e38e5a7009f353fa7ac5ea266cd2366825eeb99656721b87c6c25f49afdb" Dec 06 04:41:05 crc kubenswrapper[4802]: E1206 04:41:05.451696 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:41:16 crc kubenswrapper[4802]: I1206 04:41:16.450164 4802 scope.go:117] "RemoveContainer" containerID="3037e38e5a7009f353fa7ac5ea266cd2366825eeb99656721b87c6c25f49afdb" Dec 06 04:41:16 crc kubenswrapper[4802]: E1206 04:41:16.450989 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:41:29 crc kubenswrapper[4802]: I1206 04:41:29.450360 4802 scope.go:117] "RemoveContainer" containerID="3037e38e5a7009f353fa7ac5ea266cd2366825eeb99656721b87c6c25f49afdb" Dec 06 04:41:29 crc kubenswrapper[4802]: E1206 04:41:29.451084 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:41:42 crc kubenswrapper[4802]: I1206 04:41:42.451131 4802 scope.go:117] "RemoveContainer" containerID="3037e38e5a7009f353fa7ac5ea266cd2366825eeb99656721b87c6c25f49afdb" Dec 06 04:41:42 crc kubenswrapper[4802]: E1206 04:41:42.453353 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:41:48 crc kubenswrapper[4802]: I1206 04:41:48.010012 4802 generic.go:334] "Generic (PLEG): container finished" podID="63c9a1aa-a443-4071-b60e-317a7b81395e" containerID="6d5db2b1a0c0fa5c3abd518f7e7502c7095b137577f7b3ca876dd132cbfe3ad6" exitCode=0 Dec 06 04:41:48 crc kubenswrapper[4802]: I1206 04:41:48.010142 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"63c9a1aa-a443-4071-b60e-317a7b81395e","Type":"ContainerDied","Data":"6d5db2b1a0c0fa5c3abd518f7e7502c7095b137577f7b3ca876dd132cbfe3ad6"} Dec 06 04:41:49 crc kubenswrapper[4802]: I1206 04:41:49.805324 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 06 04:41:49 crc kubenswrapper[4802]: I1206 04:41:49.925107 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"63c9a1aa-a443-4071-b60e-317a7b81395e\" (UID: \"63c9a1aa-a443-4071-b60e-317a7b81395e\") " Dec 06 04:41:49 crc kubenswrapper[4802]: I1206 04:41:49.925175 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/63c9a1aa-a443-4071-b60e-317a7b81395e-test-operator-ephemeral-temporary\") pod \"63c9a1aa-a443-4071-b60e-317a7b81395e\" (UID: \"63c9a1aa-a443-4071-b60e-317a7b81395e\") " Dec 06 04:41:49 crc kubenswrapper[4802]: I1206 04:41:49.925205 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-srnml\" (UniqueName: \"kubernetes.io/projected/63c9a1aa-a443-4071-b60e-317a7b81395e-kube-api-access-srnml\") pod \"63c9a1aa-a443-4071-b60e-317a7b81395e\" (UID: \"63c9a1aa-a443-4071-b60e-317a7b81395e\") " Dec 06 04:41:49 crc kubenswrapper[4802]: I1206 04:41:49.925296 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/63c9a1aa-a443-4071-b60e-317a7b81395e-ca-certs\") pod \"63c9a1aa-a443-4071-b60e-317a7b81395e\" (UID: \"63c9a1aa-a443-4071-b60e-317a7b81395e\") " Dec 06 04:41:49 crc kubenswrapper[4802]: I1206 04:41:49.925311 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/63c9a1aa-a443-4071-b60e-317a7b81395e-test-operator-ephemeral-workdir\") pod \"63c9a1aa-a443-4071-b60e-317a7b81395e\" (UID: \"63c9a1aa-a443-4071-b60e-317a7b81395e\") " Dec 06 04:41:49 crc kubenswrapper[4802]: I1206 04:41:49.925444 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/63c9a1aa-a443-4071-b60e-317a7b81395e-openstack-config\") pod \"63c9a1aa-a443-4071-b60e-317a7b81395e\" (UID: \"63c9a1aa-a443-4071-b60e-317a7b81395e\") " Dec 06 04:41:49 crc kubenswrapper[4802]: I1206 04:41:49.925465 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/63c9a1aa-a443-4071-b60e-317a7b81395e-openstack-config-secret\") pod \"63c9a1aa-a443-4071-b60e-317a7b81395e\" (UID: \"63c9a1aa-a443-4071-b60e-317a7b81395e\") " Dec 06 04:41:49 crc kubenswrapper[4802]: I1206 04:41:49.925489 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/63c9a1aa-a443-4071-b60e-317a7b81395e-ssh-key\") pod \"63c9a1aa-a443-4071-b60e-317a7b81395e\" (UID: \"63c9a1aa-a443-4071-b60e-317a7b81395e\") " Dec 06 04:41:49 crc kubenswrapper[4802]: I1206 04:41:49.925518 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/63c9a1aa-a443-4071-b60e-317a7b81395e-config-data\") pod \"63c9a1aa-a443-4071-b60e-317a7b81395e\" (UID: \"63c9a1aa-a443-4071-b60e-317a7b81395e\") " Dec 06 04:41:49 crc kubenswrapper[4802]: I1206 04:41:49.926577 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63c9a1aa-a443-4071-b60e-317a7b81395e-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "63c9a1aa-a443-4071-b60e-317a7b81395e" (UID: "63c9a1aa-a443-4071-b60e-317a7b81395e"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:41:49 crc kubenswrapper[4802]: I1206 04:41:49.927183 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63c9a1aa-a443-4071-b60e-317a7b81395e-config-data" (OuterVolumeSpecName: "config-data") pod "63c9a1aa-a443-4071-b60e-317a7b81395e" (UID: "63c9a1aa-a443-4071-b60e-317a7b81395e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:41:49 crc kubenswrapper[4802]: I1206 04:41:49.929066 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63c9a1aa-a443-4071-b60e-317a7b81395e-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "63c9a1aa-a443-4071-b60e-317a7b81395e" (UID: "63c9a1aa-a443-4071-b60e-317a7b81395e"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:41:49 crc kubenswrapper[4802]: I1206 04:41:49.931961 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63c9a1aa-a443-4071-b60e-317a7b81395e-kube-api-access-srnml" (OuterVolumeSpecName: "kube-api-access-srnml") pod "63c9a1aa-a443-4071-b60e-317a7b81395e" (UID: "63c9a1aa-a443-4071-b60e-317a7b81395e"). InnerVolumeSpecName "kube-api-access-srnml". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:41:49 crc kubenswrapper[4802]: I1206 04:41:49.933964 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "test-operator-logs") pod "63c9a1aa-a443-4071-b60e-317a7b81395e" (UID: "63c9a1aa-a443-4071-b60e-317a7b81395e"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 04:41:49 crc kubenswrapper[4802]: I1206 04:41:49.956436 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63c9a1aa-a443-4071-b60e-317a7b81395e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "63c9a1aa-a443-4071-b60e-317a7b81395e" (UID: "63c9a1aa-a443-4071-b60e-317a7b81395e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:41:49 crc kubenswrapper[4802]: I1206 04:41:49.960201 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63c9a1aa-a443-4071-b60e-317a7b81395e-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "63c9a1aa-a443-4071-b60e-317a7b81395e" (UID: "63c9a1aa-a443-4071-b60e-317a7b81395e"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:41:49 crc kubenswrapper[4802]: I1206 04:41:49.962390 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63c9a1aa-a443-4071-b60e-317a7b81395e-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "63c9a1aa-a443-4071-b60e-317a7b81395e" (UID: "63c9a1aa-a443-4071-b60e-317a7b81395e"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:41:49 crc kubenswrapper[4802]: I1206 04:41:49.995672 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63c9a1aa-a443-4071-b60e-317a7b81395e-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "63c9a1aa-a443-4071-b60e-317a7b81395e" (UID: "63c9a1aa-a443-4071-b60e-317a7b81395e"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:41:50 crc kubenswrapper[4802]: I1206 04:41:50.027773 4802 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/63c9a1aa-a443-4071-b60e-317a7b81395e-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Dec 06 04:41:50 crc kubenswrapper[4802]: I1206 04:41:50.027885 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-srnml\" (UniqueName: \"kubernetes.io/projected/63c9a1aa-a443-4071-b60e-317a7b81395e-kube-api-access-srnml\") on node \"crc\" DevicePath \"\"" Dec 06 04:41:50 crc kubenswrapper[4802]: I1206 04:41:50.027902 4802 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/63c9a1aa-a443-4071-b60e-317a7b81395e-ca-certs\") on node \"crc\" DevicePath \"\"" Dec 06 04:41:50 crc kubenswrapper[4802]: I1206 04:41:50.027914 4802 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/63c9a1aa-a443-4071-b60e-317a7b81395e-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Dec 06 04:41:50 crc kubenswrapper[4802]: I1206 04:41:50.027928 4802 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/63c9a1aa-a443-4071-b60e-317a7b81395e-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 06 04:41:50 crc kubenswrapper[4802]: I1206 04:41:50.027940 4802 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/63c9a1aa-a443-4071-b60e-317a7b81395e-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 06 04:41:50 crc kubenswrapper[4802]: I1206 04:41:50.027952 4802 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/63c9a1aa-a443-4071-b60e-317a7b81395e-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 04:41:50 crc kubenswrapper[4802]: I1206 04:41:50.027963 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/63c9a1aa-a443-4071-b60e-317a7b81395e-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 04:41:50 crc kubenswrapper[4802]: I1206 04:41:50.027994 4802 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Dec 06 04:41:50 crc kubenswrapper[4802]: I1206 04:41:50.030103 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"63c9a1aa-a443-4071-b60e-317a7b81395e","Type":"ContainerDied","Data":"07b4a4f8606f4f4fb91971ba1cbdfe0e279ec2deaa25ec341c8ea507ecc7f035"} Dec 06 04:41:50 crc kubenswrapper[4802]: I1206 04:41:50.030148 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="07b4a4f8606f4f4fb91971ba1cbdfe0e279ec2deaa25ec341c8ea507ecc7f035" Dec 06 04:41:50 crc kubenswrapper[4802]: I1206 04:41:50.030400 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 06 04:41:50 crc kubenswrapper[4802]: I1206 04:41:50.051136 4802 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Dec 06 04:41:50 crc kubenswrapper[4802]: I1206 04:41:50.129355 4802 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Dec 06 04:41:53 crc kubenswrapper[4802]: I1206 04:41:53.450720 4802 scope.go:117] "RemoveContainer" containerID="3037e38e5a7009f353fa7ac5ea266cd2366825eeb99656721b87c6c25f49afdb" Dec 06 04:41:53 crc kubenswrapper[4802]: E1206 04:41:53.452657 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:41:55 crc kubenswrapper[4802]: I1206 04:41:55.185581 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 06 04:41:55 crc kubenswrapper[4802]: E1206 04:41:55.186496 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90937dee-5d01-4b56-9b2e-5a9c3cd78232" containerName="registry-server" Dec 06 04:41:55 crc kubenswrapper[4802]: I1206 04:41:55.186515 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="90937dee-5d01-4b56-9b2e-5a9c3cd78232" containerName="registry-server" Dec 06 04:41:55 crc kubenswrapper[4802]: E1206 04:41:55.186535 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90937dee-5d01-4b56-9b2e-5a9c3cd78232" containerName="extract-utilities" Dec 06 04:41:55 crc kubenswrapper[4802]: I1206 04:41:55.186542 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="90937dee-5d01-4b56-9b2e-5a9c3cd78232" containerName="extract-utilities" Dec 06 04:41:55 crc kubenswrapper[4802]: E1206 04:41:55.186556 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90937dee-5d01-4b56-9b2e-5a9c3cd78232" containerName="extract-content" Dec 06 04:41:55 crc kubenswrapper[4802]: I1206 04:41:55.186563 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="90937dee-5d01-4b56-9b2e-5a9c3cd78232" containerName="extract-content" Dec 06 04:41:55 crc kubenswrapper[4802]: E1206 04:41:55.186589 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63c9a1aa-a443-4071-b60e-317a7b81395e" containerName="tempest-tests-tempest-tests-runner" Dec 06 04:41:55 crc kubenswrapper[4802]: I1206 04:41:55.186599 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="63c9a1aa-a443-4071-b60e-317a7b81395e" containerName="tempest-tests-tempest-tests-runner" Dec 06 04:41:55 crc kubenswrapper[4802]: I1206 04:41:55.186867 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="90937dee-5d01-4b56-9b2e-5a9c3cd78232" containerName="registry-server" Dec 06 04:41:55 crc kubenswrapper[4802]: I1206 04:41:55.186884 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="63c9a1aa-a443-4071-b60e-317a7b81395e" containerName="tempest-tests-tempest-tests-runner" Dec 06 04:41:55 crc kubenswrapper[4802]: I1206 04:41:55.187851 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 06 04:41:55 crc kubenswrapper[4802]: I1206 04:41:55.190222 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-grnp2" Dec 06 04:41:55 crc kubenswrapper[4802]: I1206 04:41:55.200388 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 06 04:41:55 crc kubenswrapper[4802]: I1206 04:41:55.328228 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"35caa18c-c7c4-4e74-9c6d-053c38327afc\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 06 04:41:55 crc kubenswrapper[4802]: I1206 04:41:55.328329 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnh54\" (UniqueName: \"kubernetes.io/projected/35caa18c-c7c4-4e74-9c6d-053c38327afc-kube-api-access-mnh54\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"35caa18c-c7c4-4e74-9c6d-053c38327afc\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 06 04:41:55 crc kubenswrapper[4802]: I1206 04:41:55.429742 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"35caa18c-c7c4-4e74-9c6d-053c38327afc\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 06 04:41:55 crc kubenswrapper[4802]: I1206 04:41:55.429811 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnh54\" (UniqueName: \"kubernetes.io/projected/35caa18c-c7c4-4e74-9c6d-053c38327afc-kube-api-access-mnh54\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"35caa18c-c7c4-4e74-9c6d-053c38327afc\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 06 04:41:55 crc kubenswrapper[4802]: I1206 04:41:55.430157 4802 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"35caa18c-c7c4-4e74-9c6d-053c38327afc\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 06 04:41:55 crc kubenswrapper[4802]: I1206 04:41:55.458950 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnh54\" (UniqueName: \"kubernetes.io/projected/35caa18c-c7c4-4e74-9c6d-053c38327afc-kube-api-access-mnh54\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"35caa18c-c7c4-4e74-9c6d-053c38327afc\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 06 04:41:55 crc kubenswrapper[4802]: I1206 04:41:55.460383 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"35caa18c-c7c4-4e74-9c6d-053c38327afc\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 06 04:41:55 crc kubenswrapper[4802]: I1206 04:41:55.510224 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 06 04:41:56 crc kubenswrapper[4802]: I1206 04:41:56.003206 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 06 04:41:56 crc kubenswrapper[4802]: I1206 04:41:56.006509 4802 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 04:41:56 crc kubenswrapper[4802]: I1206 04:41:56.087765 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"35caa18c-c7c4-4e74-9c6d-053c38327afc","Type":"ContainerStarted","Data":"61359687f2eb5a447efbecb9d8bea7e2bbc8e890930ce6fade3628f9e68af439"} Dec 06 04:41:57 crc kubenswrapper[4802]: I1206 04:41:57.097386 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"35caa18c-c7c4-4e74-9c6d-053c38327afc","Type":"ContainerStarted","Data":"0c54cea2f8bda65acbc13c5d83b2626d418b4d3300516cf3e62b72b52e032d8d"} Dec 06 04:41:57 crc kubenswrapper[4802]: I1206 04:41:57.119150 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=1.352688657 podStartE2EDuration="2.119132384s" podCreationTimestamp="2025-12-06 04:41:55 +0000 UTC" firstStartedPulling="2025-12-06 04:41:56.006049053 +0000 UTC m=+3708.877958205" lastFinishedPulling="2025-12-06 04:41:56.77249278 +0000 UTC m=+3709.644401932" observedRunningTime="2025-12-06 04:41:57.113980773 +0000 UTC m=+3709.985889935" watchObservedRunningTime="2025-12-06 04:41:57.119132384 +0000 UTC m=+3709.991041556" Dec 06 04:42:04 crc kubenswrapper[4802]: I1206 04:42:04.456191 4802 scope.go:117] "RemoveContainer" containerID="3037e38e5a7009f353fa7ac5ea266cd2366825eeb99656721b87c6c25f49afdb" Dec 06 04:42:04 crc kubenswrapper[4802]: E1206 04:42:04.457200 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:42:15 crc kubenswrapper[4802]: I1206 04:42:15.450323 4802 scope.go:117] "RemoveContainer" containerID="3037e38e5a7009f353fa7ac5ea266cd2366825eeb99656721b87c6c25f49afdb" Dec 06 04:42:15 crc kubenswrapper[4802]: E1206 04:42:15.450971 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:42:20 crc kubenswrapper[4802]: I1206 04:42:20.483607 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mjbfz/must-gather-drcf4"] Dec 06 04:42:20 crc kubenswrapper[4802]: I1206 04:42:20.486072 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mjbfz/must-gather-drcf4" Dec 06 04:42:20 crc kubenswrapper[4802]: I1206 04:42:20.492047 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-mjbfz"/"openshift-service-ca.crt" Dec 06 04:42:20 crc kubenswrapper[4802]: I1206 04:42:20.493306 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-mjbfz"/"kube-root-ca.crt" Dec 06 04:42:20 crc kubenswrapper[4802]: I1206 04:42:20.507145 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-mjbfz/must-gather-drcf4"] Dec 06 04:42:20 crc kubenswrapper[4802]: I1206 04:42:20.586602 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tppg\" (UniqueName: \"kubernetes.io/projected/db83105e-99af-4dde-8ed4-e69a9c915d98-kube-api-access-2tppg\") pod \"must-gather-drcf4\" (UID: \"db83105e-99af-4dde-8ed4-e69a9c915d98\") " pod="openshift-must-gather-mjbfz/must-gather-drcf4" Dec 06 04:42:20 crc kubenswrapper[4802]: I1206 04:42:20.586658 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/db83105e-99af-4dde-8ed4-e69a9c915d98-must-gather-output\") pod \"must-gather-drcf4\" (UID: \"db83105e-99af-4dde-8ed4-e69a9c915d98\") " pod="openshift-must-gather-mjbfz/must-gather-drcf4" Dec 06 04:42:20 crc kubenswrapper[4802]: I1206 04:42:20.689026 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tppg\" (UniqueName: \"kubernetes.io/projected/db83105e-99af-4dde-8ed4-e69a9c915d98-kube-api-access-2tppg\") pod \"must-gather-drcf4\" (UID: \"db83105e-99af-4dde-8ed4-e69a9c915d98\") " pod="openshift-must-gather-mjbfz/must-gather-drcf4" Dec 06 04:42:20 crc kubenswrapper[4802]: I1206 04:42:20.689072 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/db83105e-99af-4dde-8ed4-e69a9c915d98-must-gather-output\") pod \"must-gather-drcf4\" (UID: \"db83105e-99af-4dde-8ed4-e69a9c915d98\") " pod="openshift-must-gather-mjbfz/must-gather-drcf4" Dec 06 04:42:20 crc kubenswrapper[4802]: I1206 04:42:20.689459 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/db83105e-99af-4dde-8ed4-e69a9c915d98-must-gather-output\") pod \"must-gather-drcf4\" (UID: \"db83105e-99af-4dde-8ed4-e69a9c915d98\") " pod="openshift-must-gather-mjbfz/must-gather-drcf4" Dec 06 04:42:20 crc kubenswrapper[4802]: I1206 04:42:20.706915 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tppg\" (UniqueName: \"kubernetes.io/projected/db83105e-99af-4dde-8ed4-e69a9c915d98-kube-api-access-2tppg\") pod \"must-gather-drcf4\" (UID: \"db83105e-99af-4dde-8ed4-e69a9c915d98\") " pod="openshift-must-gather-mjbfz/must-gather-drcf4" Dec 06 04:42:20 crc kubenswrapper[4802]: I1206 04:42:20.817568 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mjbfz/must-gather-drcf4" Dec 06 04:42:21 crc kubenswrapper[4802]: I1206 04:42:21.368150 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-mjbfz/must-gather-drcf4"] Dec 06 04:42:22 crc kubenswrapper[4802]: I1206 04:42:22.372172 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mjbfz/must-gather-drcf4" event={"ID":"db83105e-99af-4dde-8ed4-e69a9c915d98","Type":"ContainerStarted","Data":"f7ac7b3d67cee9e70f82dcc6f3d76c085b0857a7753daf0aff3eb8903c7d63fc"} Dec 06 04:42:26 crc kubenswrapper[4802]: I1206 04:42:26.413817 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mjbfz/must-gather-drcf4" event={"ID":"db83105e-99af-4dde-8ed4-e69a9c915d98","Type":"ContainerStarted","Data":"baa9c7d7aec1f6ae925ee81bd954c72d99ce577c438ba999e98173249997734a"} Dec 06 04:42:26 crc kubenswrapper[4802]: I1206 04:42:26.414378 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mjbfz/must-gather-drcf4" event={"ID":"db83105e-99af-4dde-8ed4-e69a9c915d98","Type":"ContainerStarted","Data":"d429ca90df1982428ddc9e9dd734b968cf1d9575e289c938b6db28f48f74b400"} Dec 06 04:42:26 crc kubenswrapper[4802]: I1206 04:42:26.427306 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-mjbfz/must-gather-drcf4" podStartSLOduration=2.550181412 podStartE2EDuration="6.427285449s" podCreationTimestamp="2025-12-06 04:42:20 +0000 UTC" firstStartedPulling="2025-12-06 04:42:21.369907093 +0000 UTC m=+3734.241816285" lastFinishedPulling="2025-12-06 04:42:25.24701116 +0000 UTC m=+3738.118920322" observedRunningTime="2025-12-06 04:42:26.426724563 +0000 UTC m=+3739.298633715" watchObservedRunningTime="2025-12-06 04:42:26.427285449 +0000 UTC m=+3739.299194611" Dec 06 04:42:29 crc kubenswrapper[4802]: I1206 04:42:29.100414 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mjbfz/crc-debug-sxdq4"] Dec 06 04:42:29 crc kubenswrapper[4802]: I1206 04:42:29.102094 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mjbfz/crc-debug-sxdq4" Dec 06 04:42:29 crc kubenswrapper[4802]: I1206 04:42:29.104622 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-mjbfz"/"default-dockercfg-th975" Dec 06 04:42:29 crc kubenswrapper[4802]: I1206 04:42:29.154699 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qm5jz\" (UniqueName: \"kubernetes.io/projected/be657733-b363-4c3e-be34-77ce40ad8e4f-kube-api-access-qm5jz\") pod \"crc-debug-sxdq4\" (UID: \"be657733-b363-4c3e-be34-77ce40ad8e4f\") " pod="openshift-must-gather-mjbfz/crc-debug-sxdq4" Dec 06 04:42:29 crc kubenswrapper[4802]: I1206 04:42:29.154902 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/be657733-b363-4c3e-be34-77ce40ad8e4f-host\") pod \"crc-debug-sxdq4\" (UID: \"be657733-b363-4c3e-be34-77ce40ad8e4f\") " pod="openshift-must-gather-mjbfz/crc-debug-sxdq4" Dec 06 04:42:29 crc kubenswrapper[4802]: I1206 04:42:29.256955 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qm5jz\" (UniqueName: \"kubernetes.io/projected/be657733-b363-4c3e-be34-77ce40ad8e4f-kube-api-access-qm5jz\") pod \"crc-debug-sxdq4\" (UID: \"be657733-b363-4c3e-be34-77ce40ad8e4f\") " pod="openshift-must-gather-mjbfz/crc-debug-sxdq4" Dec 06 04:42:29 crc kubenswrapper[4802]: I1206 04:42:29.257021 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/be657733-b363-4c3e-be34-77ce40ad8e4f-host\") pod \"crc-debug-sxdq4\" (UID: \"be657733-b363-4c3e-be34-77ce40ad8e4f\") " pod="openshift-must-gather-mjbfz/crc-debug-sxdq4" Dec 06 04:42:29 crc kubenswrapper[4802]: I1206 04:42:29.257139 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/be657733-b363-4c3e-be34-77ce40ad8e4f-host\") pod \"crc-debug-sxdq4\" (UID: \"be657733-b363-4c3e-be34-77ce40ad8e4f\") " pod="openshift-must-gather-mjbfz/crc-debug-sxdq4" Dec 06 04:42:29 crc kubenswrapper[4802]: I1206 04:42:29.276512 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qm5jz\" (UniqueName: \"kubernetes.io/projected/be657733-b363-4c3e-be34-77ce40ad8e4f-kube-api-access-qm5jz\") pod \"crc-debug-sxdq4\" (UID: \"be657733-b363-4c3e-be34-77ce40ad8e4f\") " pod="openshift-must-gather-mjbfz/crc-debug-sxdq4" Dec 06 04:42:29 crc kubenswrapper[4802]: I1206 04:42:29.423967 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mjbfz/crc-debug-sxdq4" Dec 06 04:42:29 crc kubenswrapper[4802]: W1206 04:42:29.471661 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbe657733_b363_4c3e_be34_77ce40ad8e4f.slice/crio-f1133bcc97c4d2b4902ac3196e4f3e1fc7b987bcad2784592c7cc41e066880d3 WatchSource:0}: Error finding container f1133bcc97c4d2b4902ac3196e4f3e1fc7b987bcad2784592c7cc41e066880d3: Status 404 returned error can't find the container with id f1133bcc97c4d2b4902ac3196e4f3e1fc7b987bcad2784592c7cc41e066880d3 Dec 06 04:42:30 crc kubenswrapper[4802]: I1206 04:42:30.449734 4802 scope.go:117] "RemoveContainer" containerID="3037e38e5a7009f353fa7ac5ea266cd2366825eeb99656721b87c6c25f49afdb" Dec 06 04:42:30 crc kubenswrapper[4802]: E1206 04:42:30.450093 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:42:30 crc kubenswrapper[4802]: I1206 04:42:30.454986 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mjbfz/crc-debug-sxdq4" event={"ID":"be657733-b363-4c3e-be34-77ce40ad8e4f","Type":"ContainerStarted","Data":"f1133bcc97c4d2b4902ac3196e4f3e1fc7b987bcad2784592c7cc41e066880d3"} Dec 06 04:42:41 crc kubenswrapper[4802]: I1206 04:42:41.571548 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mjbfz/crc-debug-sxdq4" event={"ID":"be657733-b363-4c3e-be34-77ce40ad8e4f","Type":"ContainerStarted","Data":"00e0274365781e48d8eb08dbc354376abef272f2869caefe62196347884e7132"} Dec 06 04:42:41 crc kubenswrapper[4802]: I1206 04:42:41.594782 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-mjbfz/crc-debug-sxdq4" podStartSLOduration=1.225458467 podStartE2EDuration="12.594739099s" podCreationTimestamp="2025-12-06 04:42:29 +0000 UTC" firstStartedPulling="2025-12-06 04:42:29.473558536 +0000 UTC m=+3742.345467688" lastFinishedPulling="2025-12-06 04:42:40.842839168 +0000 UTC m=+3753.714748320" observedRunningTime="2025-12-06 04:42:41.587472282 +0000 UTC m=+3754.459381434" watchObservedRunningTime="2025-12-06 04:42:41.594739099 +0000 UTC m=+3754.466648251" Dec 06 04:42:44 crc kubenswrapper[4802]: I1206 04:42:44.450275 4802 scope.go:117] "RemoveContainer" containerID="3037e38e5a7009f353fa7ac5ea266cd2366825eeb99656721b87c6c25f49afdb" Dec 06 04:42:44 crc kubenswrapper[4802]: E1206 04:42:44.450935 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:42:56 crc kubenswrapper[4802]: I1206 04:42:56.451357 4802 scope.go:117] "RemoveContainer" containerID="3037e38e5a7009f353fa7ac5ea266cd2366825eeb99656721b87c6c25f49afdb" Dec 06 04:42:56 crc kubenswrapper[4802]: E1206 04:42:56.452187 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:43:11 crc kubenswrapper[4802]: I1206 04:43:11.450913 4802 scope.go:117] "RemoveContainer" containerID="3037e38e5a7009f353fa7ac5ea266cd2366825eeb99656721b87c6c25f49afdb" Dec 06 04:43:11 crc kubenswrapper[4802]: E1206 04:43:11.451553 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:43:25 crc kubenswrapper[4802]: I1206 04:43:25.450342 4802 scope.go:117] "RemoveContainer" containerID="3037e38e5a7009f353fa7ac5ea266cd2366825eeb99656721b87c6c25f49afdb" Dec 06 04:43:25 crc kubenswrapper[4802]: E1206 04:43:25.451570 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:43:35 crc kubenswrapper[4802]: I1206 04:43:35.116140 4802 generic.go:334] "Generic (PLEG): container finished" podID="be657733-b363-4c3e-be34-77ce40ad8e4f" containerID="00e0274365781e48d8eb08dbc354376abef272f2869caefe62196347884e7132" exitCode=0 Dec 06 04:43:35 crc kubenswrapper[4802]: I1206 04:43:35.116248 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mjbfz/crc-debug-sxdq4" event={"ID":"be657733-b363-4c3e-be34-77ce40ad8e4f","Type":"ContainerDied","Data":"00e0274365781e48d8eb08dbc354376abef272f2869caefe62196347884e7132"} Dec 06 04:43:36 crc kubenswrapper[4802]: I1206 04:43:36.252350 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mjbfz/crc-debug-sxdq4" Dec 06 04:43:36 crc kubenswrapper[4802]: I1206 04:43:36.290595 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mjbfz/crc-debug-sxdq4"] Dec 06 04:43:36 crc kubenswrapper[4802]: I1206 04:43:36.299271 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mjbfz/crc-debug-sxdq4"] Dec 06 04:43:36 crc kubenswrapper[4802]: I1206 04:43:36.346266 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qm5jz\" (UniqueName: \"kubernetes.io/projected/be657733-b363-4c3e-be34-77ce40ad8e4f-kube-api-access-qm5jz\") pod \"be657733-b363-4c3e-be34-77ce40ad8e4f\" (UID: \"be657733-b363-4c3e-be34-77ce40ad8e4f\") " Dec 06 04:43:36 crc kubenswrapper[4802]: I1206 04:43:36.346426 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/be657733-b363-4c3e-be34-77ce40ad8e4f-host\") pod \"be657733-b363-4c3e-be34-77ce40ad8e4f\" (UID: \"be657733-b363-4c3e-be34-77ce40ad8e4f\") " Dec 06 04:43:36 crc kubenswrapper[4802]: I1206 04:43:36.346886 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/be657733-b363-4c3e-be34-77ce40ad8e4f-host" (OuterVolumeSpecName: "host") pod "be657733-b363-4c3e-be34-77ce40ad8e4f" (UID: "be657733-b363-4c3e-be34-77ce40ad8e4f"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 04:43:36 crc kubenswrapper[4802]: I1206 04:43:36.354798 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be657733-b363-4c3e-be34-77ce40ad8e4f-kube-api-access-qm5jz" (OuterVolumeSpecName: "kube-api-access-qm5jz") pod "be657733-b363-4c3e-be34-77ce40ad8e4f" (UID: "be657733-b363-4c3e-be34-77ce40ad8e4f"). InnerVolumeSpecName "kube-api-access-qm5jz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:43:36 crc kubenswrapper[4802]: I1206 04:43:36.448444 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qm5jz\" (UniqueName: \"kubernetes.io/projected/be657733-b363-4c3e-be34-77ce40ad8e4f-kube-api-access-qm5jz\") on node \"crc\" DevicePath \"\"" Dec 06 04:43:36 crc kubenswrapper[4802]: I1206 04:43:36.448480 4802 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/be657733-b363-4c3e-be34-77ce40ad8e4f-host\") on node \"crc\" DevicePath \"\"" Dec 06 04:43:37 crc kubenswrapper[4802]: I1206 04:43:37.140590 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f1133bcc97c4d2b4902ac3196e4f3e1fc7b987bcad2784592c7cc41e066880d3" Dec 06 04:43:37 crc kubenswrapper[4802]: I1206 04:43:37.140687 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mjbfz/crc-debug-sxdq4" Dec 06 04:43:37 crc kubenswrapper[4802]: I1206 04:43:37.458222 4802 scope.go:117] "RemoveContainer" containerID="3037e38e5a7009f353fa7ac5ea266cd2366825eeb99656721b87c6c25f49afdb" Dec 06 04:43:37 crc kubenswrapper[4802]: E1206 04:43:37.458497 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:43:37 crc kubenswrapper[4802]: I1206 04:43:37.462774 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be657733-b363-4c3e-be34-77ce40ad8e4f" path="/var/lib/kubelet/pods/be657733-b363-4c3e-be34-77ce40ad8e4f/volumes" Dec 06 04:43:37 crc kubenswrapper[4802]: I1206 04:43:37.481103 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mjbfz/crc-debug-4mfpk"] Dec 06 04:43:37 crc kubenswrapper[4802]: E1206 04:43:37.481488 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be657733-b363-4c3e-be34-77ce40ad8e4f" containerName="container-00" Dec 06 04:43:37 crc kubenswrapper[4802]: I1206 04:43:37.481505 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="be657733-b363-4c3e-be34-77ce40ad8e4f" containerName="container-00" Dec 06 04:43:37 crc kubenswrapper[4802]: I1206 04:43:37.481680 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="be657733-b363-4c3e-be34-77ce40ad8e4f" containerName="container-00" Dec 06 04:43:37 crc kubenswrapper[4802]: I1206 04:43:37.482284 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mjbfz/crc-debug-4mfpk" Dec 06 04:43:37 crc kubenswrapper[4802]: I1206 04:43:37.487060 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-mjbfz"/"default-dockercfg-th975" Dec 06 04:43:37 crc kubenswrapper[4802]: I1206 04:43:37.566841 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3faa0cf0-3bb3-4e7e-96eb-27489147c8e2-host\") pod \"crc-debug-4mfpk\" (UID: \"3faa0cf0-3bb3-4e7e-96eb-27489147c8e2\") " pod="openshift-must-gather-mjbfz/crc-debug-4mfpk" Dec 06 04:43:37 crc kubenswrapper[4802]: I1206 04:43:37.567027 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkz4n\" (UniqueName: \"kubernetes.io/projected/3faa0cf0-3bb3-4e7e-96eb-27489147c8e2-kube-api-access-hkz4n\") pod \"crc-debug-4mfpk\" (UID: \"3faa0cf0-3bb3-4e7e-96eb-27489147c8e2\") " pod="openshift-must-gather-mjbfz/crc-debug-4mfpk" Dec 06 04:43:37 crc kubenswrapper[4802]: I1206 04:43:37.668132 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkz4n\" (UniqueName: \"kubernetes.io/projected/3faa0cf0-3bb3-4e7e-96eb-27489147c8e2-kube-api-access-hkz4n\") pod \"crc-debug-4mfpk\" (UID: \"3faa0cf0-3bb3-4e7e-96eb-27489147c8e2\") " pod="openshift-must-gather-mjbfz/crc-debug-4mfpk" Dec 06 04:43:37 crc kubenswrapper[4802]: I1206 04:43:37.668240 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3faa0cf0-3bb3-4e7e-96eb-27489147c8e2-host\") pod \"crc-debug-4mfpk\" (UID: \"3faa0cf0-3bb3-4e7e-96eb-27489147c8e2\") " pod="openshift-must-gather-mjbfz/crc-debug-4mfpk" Dec 06 04:43:37 crc kubenswrapper[4802]: I1206 04:43:37.668378 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3faa0cf0-3bb3-4e7e-96eb-27489147c8e2-host\") pod \"crc-debug-4mfpk\" (UID: \"3faa0cf0-3bb3-4e7e-96eb-27489147c8e2\") " pod="openshift-must-gather-mjbfz/crc-debug-4mfpk" Dec 06 04:43:37 crc kubenswrapper[4802]: I1206 04:43:37.691522 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkz4n\" (UniqueName: \"kubernetes.io/projected/3faa0cf0-3bb3-4e7e-96eb-27489147c8e2-kube-api-access-hkz4n\") pod \"crc-debug-4mfpk\" (UID: \"3faa0cf0-3bb3-4e7e-96eb-27489147c8e2\") " pod="openshift-must-gather-mjbfz/crc-debug-4mfpk" Dec 06 04:43:37 crc kubenswrapper[4802]: I1206 04:43:37.805790 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mjbfz/crc-debug-4mfpk" Dec 06 04:43:38 crc kubenswrapper[4802]: I1206 04:43:38.152041 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mjbfz/crc-debug-4mfpk" event={"ID":"3faa0cf0-3bb3-4e7e-96eb-27489147c8e2","Type":"ContainerStarted","Data":"a9609204477a2d60e507ab4a5e3dd05464af394cd5e3ec9522e95f2d8714af9d"} Dec 06 04:43:39 crc kubenswrapper[4802]: I1206 04:43:39.164722 4802 generic.go:334] "Generic (PLEG): container finished" podID="3faa0cf0-3bb3-4e7e-96eb-27489147c8e2" containerID="f626bb7c325f4af3e6d07df8598f90df2a9a83ff8777bcae9446f085f5ebc5a0" exitCode=0 Dec 06 04:43:39 crc kubenswrapper[4802]: I1206 04:43:39.164809 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mjbfz/crc-debug-4mfpk" event={"ID":"3faa0cf0-3bb3-4e7e-96eb-27489147c8e2","Type":"ContainerDied","Data":"f626bb7c325f4af3e6d07df8598f90df2a9a83ff8777bcae9446f085f5ebc5a0"} Dec 06 04:43:40 crc kubenswrapper[4802]: I1206 04:43:40.285921 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mjbfz/crc-debug-4mfpk" Dec 06 04:43:40 crc kubenswrapper[4802]: I1206 04:43:40.318415 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3faa0cf0-3bb3-4e7e-96eb-27489147c8e2-host\") pod \"3faa0cf0-3bb3-4e7e-96eb-27489147c8e2\" (UID: \"3faa0cf0-3bb3-4e7e-96eb-27489147c8e2\") " Dec 06 04:43:40 crc kubenswrapper[4802]: I1206 04:43:40.318598 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3faa0cf0-3bb3-4e7e-96eb-27489147c8e2-host" (OuterVolumeSpecName: "host") pod "3faa0cf0-3bb3-4e7e-96eb-27489147c8e2" (UID: "3faa0cf0-3bb3-4e7e-96eb-27489147c8e2"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 04:43:40 crc kubenswrapper[4802]: I1206 04:43:40.318839 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hkz4n\" (UniqueName: \"kubernetes.io/projected/3faa0cf0-3bb3-4e7e-96eb-27489147c8e2-kube-api-access-hkz4n\") pod \"3faa0cf0-3bb3-4e7e-96eb-27489147c8e2\" (UID: \"3faa0cf0-3bb3-4e7e-96eb-27489147c8e2\") " Dec 06 04:43:40 crc kubenswrapper[4802]: I1206 04:43:40.319279 4802 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3faa0cf0-3bb3-4e7e-96eb-27489147c8e2-host\") on node \"crc\" DevicePath \"\"" Dec 06 04:43:40 crc kubenswrapper[4802]: I1206 04:43:40.326906 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3faa0cf0-3bb3-4e7e-96eb-27489147c8e2-kube-api-access-hkz4n" (OuterVolumeSpecName: "kube-api-access-hkz4n") pod "3faa0cf0-3bb3-4e7e-96eb-27489147c8e2" (UID: "3faa0cf0-3bb3-4e7e-96eb-27489147c8e2"). InnerVolumeSpecName "kube-api-access-hkz4n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:43:40 crc kubenswrapper[4802]: I1206 04:43:40.421687 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hkz4n\" (UniqueName: \"kubernetes.io/projected/3faa0cf0-3bb3-4e7e-96eb-27489147c8e2-kube-api-access-hkz4n\") on node \"crc\" DevicePath \"\"" Dec 06 04:43:41 crc kubenswrapper[4802]: I1206 04:43:41.169032 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mjbfz/crc-debug-4mfpk"] Dec 06 04:43:41 crc kubenswrapper[4802]: I1206 04:43:41.178460 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mjbfz/crc-debug-4mfpk"] Dec 06 04:43:41 crc kubenswrapper[4802]: I1206 04:43:41.197438 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a9609204477a2d60e507ab4a5e3dd05464af394cd5e3ec9522e95f2d8714af9d" Dec 06 04:43:41 crc kubenswrapper[4802]: I1206 04:43:41.197506 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mjbfz/crc-debug-4mfpk" Dec 06 04:43:41 crc kubenswrapper[4802]: I1206 04:43:41.468014 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3faa0cf0-3bb3-4e7e-96eb-27489147c8e2" path="/var/lib/kubelet/pods/3faa0cf0-3bb3-4e7e-96eb-27489147c8e2/volumes" Dec 06 04:43:42 crc kubenswrapper[4802]: I1206 04:43:42.342096 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mjbfz/crc-debug-58sr6"] Dec 06 04:43:42 crc kubenswrapper[4802]: E1206 04:43:42.342504 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3faa0cf0-3bb3-4e7e-96eb-27489147c8e2" containerName="container-00" Dec 06 04:43:42 crc kubenswrapper[4802]: I1206 04:43:42.342524 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="3faa0cf0-3bb3-4e7e-96eb-27489147c8e2" containerName="container-00" Dec 06 04:43:42 crc kubenswrapper[4802]: I1206 04:43:42.342788 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="3faa0cf0-3bb3-4e7e-96eb-27489147c8e2" containerName="container-00" Dec 06 04:43:42 crc kubenswrapper[4802]: I1206 04:43:42.343536 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mjbfz/crc-debug-58sr6" Dec 06 04:43:42 crc kubenswrapper[4802]: I1206 04:43:42.346505 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-mjbfz"/"default-dockercfg-th975" Dec 06 04:43:42 crc kubenswrapper[4802]: I1206 04:43:42.366422 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0663503c-e362-49bd-b5e1-a5f3f35bd04b-host\") pod \"crc-debug-58sr6\" (UID: \"0663503c-e362-49bd-b5e1-a5f3f35bd04b\") " pod="openshift-must-gather-mjbfz/crc-debug-58sr6" Dec 06 04:43:42 crc kubenswrapper[4802]: I1206 04:43:42.366670 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94hmr\" (UniqueName: \"kubernetes.io/projected/0663503c-e362-49bd-b5e1-a5f3f35bd04b-kube-api-access-94hmr\") pod \"crc-debug-58sr6\" (UID: \"0663503c-e362-49bd-b5e1-a5f3f35bd04b\") " pod="openshift-must-gather-mjbfz/crc-debug-58sr6" Dec 06 04:43:42 crc kubenswrapper[4802]: I1206 04:43:42.468280 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0663503c-e362-49bd-b5e1-a5f3f35bd04b-host\") pod \"crc-debug-58sr6\" (UID: \"0663503c-e362-49bd-b5e1-a5f3f35bd04b\") " pod="openshift-must-gather-mjbfz/crc-debug-58sr6" Dec 06 04:43:42 crc kubenswrapper[4802]: I1206 04:43:42.468376 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0663503c-e362-49bd-b5e1-a5f3f35bd04b-host\") pod \"crc-debug-58sr6\" (UID: \"0663503c-e362-49bd-b5e1-a5f3f35bd04b\") " pod="openshift-must-gather-mjbfz/crc-debug-58sr6" Dec 06 04:43:42 crc kubenswrapper[4802]: I1206 04:43:42.468404 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94hmr\" (UniqueName: \"kubernetes.io/projected/0663503c-e362-49bd-b5e1-a5f3f35bd04b-kube-api-access-94hmr\") pod \"crc-debug-58sr6\" (UID: \"0663503c-e362-49bd-b5e1-a5f3f35bd04b\") " pod="openshift-must-gather-mjbfz/crc-debug-58sr6" Dec 06 04:43:42 crc kubenswrapper[4802]: I1206 04:43:42.493423 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94hmr\" (UniqueName: \"kubernetes.io/projected/0663503c-e362-49bd-b5e1-a5f3f35bd04b-kube-api-access-94hmr\") pod \"crc-debug-58sr6\" (UID: \"0663503c-e362-49bd-b5e1-a5f3f35bd04b\") " pod="openshift-must-gather-mjbfz/crc-debug-58sr6" Dec 06 04:43:42 crc kubenswrapper[4802]: I1206 04:43:42.676657 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mjbfz/crc-debug-58sr6" Dec 06 04:43:43 crc kubenswrapper[4802]: I1206 04:43:43.219275 4802 generic.go:334] "Generic (PLEG): container finished" podID="0663503c-e362-49bd-b5e1-a5f3f35bd04b" containerID="a66f928daf978674360023cdd5510d6d01d7538e0898742ab6bed03e4c4a8dc6" exitCode=0 Dec 06 04:43:43 crc kubenswrapper[4802]: I1206 04:43:43.219355 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mjbfz/crc-debug-58sr6" event={"ID":"0663503c-e362-49bd-b5e1-a5f3f35bd04b","Type":"ContainerDied","Data":"a66f928daf978674360023cdd5510d6d01d7538e0898742ab6bed03e4c4a8dc6"} Dec 06 04:43:43 crc kubenswrapper[4802]: I1206 04:43:43.219643 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mjbfz/crc-debug-58sr6" event={"ID":"0663503c-e362-49bd-b5e1-a5f3f35bd04b","Type":"ContainerStarted","Data":"b441e71402b6f9ae4518683575e02eadca516350f7d03c7e63ff5ac81f1a2ba5"} Dec 06 04:43:43 crc kubenswrapper[4802]: I1206 04:43:43.257739 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mjbfz/crc-debug-58sr6"] Dec 06 04:43:43 crc kubenswrapper[4802]: I1206 04:43:43.267205 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mjbfz/crc-debug-58sr6"] Dec 06 04:43:44 crc kubenswrapper[4802]: I1206 04:43:44.337484 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mjbfz/crc-debug-58sr6" Dec 06 04:43:44 crc kubenswrapper[4802]: I1206 04:43:44.410862 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0663503c-e362-49bd-b5e1-a5f3f35bd04b-host\") pod \"0663503c-e362-49bd-b5e1-a5f3f35bd04b\" (UID: \"0663503c-e362-49bd-b5e1-a5f3f35bd04b\") " Dec 06 04:43:44 crc kubenswrapper[4802]: I1206 04:43:44.410960 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0663503c-e362-49bd-b5e1-a5f3f35bd04b-host" (OuterVolumeSpecName: "host") pod "0663503c-e362-49bd-b5e1-a5f3f35bd04b" (UID: "0663503c-e362-49bd-b5e1-a5f3f35bd04b"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 04:43:44 crc kubenswrapper[4802]: I1206 04:43:44.411040 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-94hmr\" (UniqueName: \"kubernetes.io/projected/0663503c-e362-49bd-b5e1-a5f3f35bd04b-kube-api-access-94hmr\") pod \"0663503c-e362-49bd-b5e1-a5f3f35bd04b\" (UID: \"0663503c-e362-49bd-b5e1-a5f3f35bd04b\") " Dec 06 04:43:44 crc kubenswrapper[4802]: I1206 04:43:44.411489 4802 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0663503c-e362-49bd-b5e1-a5f3f35bd04b-host\") on node \"crc\" DevicePath \"\"" Dec 06 04:43:44 crc kubenswrapper[4802]: I1206 04:43:44.417460 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0663503c-e362-49bd-b5e1-a5f3f35bd04b-kube-api-access-94hmr" (OuterVolumeSpecName: "kube-api-access-94hmr") pod "0663503c-e362-49bd-b5e1-a5f3f35bd04b" (UID: "0663503c-e362-49bd-b5e1-a5f3f35bd04b"). InnerVolumeSpecName "kube-api-access-94hmr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:43:44 crc kubenswrapper[4802]: I1206 04:43:44.513088 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-94hmr\" (UniqueName: \"kubernetes.io/projected/0663503c-e362-49bd-b5e1-a5f3f35bd04b-kube-api-access-94hmr\") on node \"crc\" DevicePath \"\"" Dec 06 04:43:45 crc kubenswrapper[4802]: I1206 04:43:45.240808 4802 scope.go:117] "RemoveContainer" containerID="a66f928daf978674360023cdd5510d6d01d7538e0898742ab6bed03e4c4a8dc6" Dec 06 04:43:45 crc kubenswrapper[4802]: I1206 04:43:45.240932 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mjbfz/crc-debug-58sr6" Dec 06 04:43:45 crc kubenswrapper[4802]: I1206 04:43:45.459695 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0663503c-e362-49bd-b5e1-a5f3f35bd04b" path="/var/lib/kubelet/pods/0663503c-e362-49bd-b5e1-a5f3f35bd04b/volumes" Dec 06 04:43:49 crc kubenswrapper[4802]: I1206 04:43:49.451432 4802 scope.go:117] "RemoveContainer" containerID="3037e38e5a7009f353fa7ac5ea266cd2366825eeb99656721b87c6c25f49afdb" Dec 06 04:43:49 crc kubenswrapper[4802]: E1206 04:43:49.452726 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:44:00 crc kubenswrapper[4802]: I1206 04:44:00.952062 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7f79b98d58-l9qc5_21e91f8f-b3aa-4a45-85e6-5192a3f85a04/barbican-api/0.log" Dec 06 04:44:01 crc kubenswrapper[4802]: I1206 04:44:01.115212 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7f79b98d58-l9qc5_21e91f8f-b3aa-4a45-85e6-5192a3f85a04/barbican-api-log/0.log" Dec 06 04:44:01 crc kubenswrapper[4802]: I1206 04:44:01.173383 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-554f6fb55d-hdgjt_b443b1b9-ec2c-499d-ae4d-1937c6f156f7/barbican-keystone-listener/0.log" Dec 06 04:44:01 crc kubenswrapper[4802]: I1206 04:44:01.349195 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-554f6fb55d-hdgjt_b443b1b9-ec2c-499d-ae4d-1937c6f156f7/barbican-keystone-listener-log/0.log" Dec 06 04:44:01 crc kubenswrapper[4802]: I1206 04:44:01.425070 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-68998b57f5-6qgkl_57a20a46-fdfb-410f-8a13-0b58fa5751da/barbican-worker/0.log" Dec 06 04:44:01 crc kubenswrapper[4802]: I1206 04:44:01.463830 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-68998b57f5-6qgkl_57a20a46-fdfb-410f-8a13-0b58fa5751da/barbican-worker-log/0.log" Dec 06 04:44:01 crc kubenswrapper[4802]: I1206 04:44:01.570273 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-n5lkw_a07f3a26-501d-49cb-8dbf-648f1cac66da/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 04:44:01 crc kubenswrapper[4802]: I1206 04:44:01.692092 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3ede1365-588f-4048-9002-034d3dbd2ef3/ceilometer-central-agent/0.log" Dec 06 04:44:01 crc kubenswrapper[4802]: I1206 04:44:01.783492 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3ede1365-588f-4048-9002-034d3dbd2ef3/proxy-httpd/0.log" Dec 06 04:44:01 crc kubenswrapper[4802]: I1206 04:44:01.819222 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3ede1365-588f-4048-9002-034d3dbd2ef3/ceilometer-notification-agent/0.log" Dec 06 04:44:01 crc kubenswrapper[4802]: I1206 04:44:01.846948 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3ede1365-588f-4048-9002-034d3dbd2ef3/sg-core/0.log" Dec 06 04:44:02 crc kubenswrapper[4802]: I1206 04:44:02.011532 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_45876a53-aa92-4ca3-bbc2-170e9855b034/cinder-api-log/0.log" Dec 06 04:44:02 crc kubenswrapper[4802]: I1206 04:44:02.015390 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_45876a53-aa92-4ca3-bbc2-170e9855b034/cinder-api/0.log" Dec 06 04:44:02 crc kubenswrapper[4802]: I1206 04:44:02.198834 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_92c245ae-d893-44d7-99e3-69a9e58735cf/cinder-scheduler/0.log" Dec 06 04:44:02 crc kubenswrapper[4802]: I1206 04:44:02.228999 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_92c245ae-d893-44d7-99e3-69a9e58735cf/probe/0.log" Dec 06 04:44:02 crc kubenswrapper[4802]: I1206 04:44:02.330405 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-22wfn_7d07f677-e982-45d5-b4c0-92e0db096fd1/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 04:44:02 crc kubenswrapper[4802]: I1206 04:44:02.457351 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-zd97g_3d1e4ffc-0f3b-4bc2-9f36-6193b7f33785/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 04:44:02 crc kubenswrapper[4802]: I1206 04:44:02.539159 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5d75f767dc-75bb8_ad17cfaf-5662-41ae-9987-766441edb7a4/init/0.log" Dec 06 04:44:02 crc kubenswrapper[4802]: I1206 04:44:02.683829 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5d75f767dc-75bb8_ad17cfaf-5662-41ae-9987-766441edb7a4/init/0.log" Dec 06 04:44:02 crc kubenswrapper[4802]: I1206 04:44:02.728605 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-sgxfp_db0df1b2-ef63-42ae-96ff-cafa00acc8e1/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 04:44:02 crc kubenswrapper[4802]: I1206 04:44:02.758199 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5d75f767dc-75bb8_ad17cfaf-5662-41ae-9987-766441edb7a4/dnsmasq-dns/0.log" Dec 06 04:44:03 crc kubenswrapper[4802]: I1206 04:44:03.052685 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_02888e6b-6185-45b8-84b7-8b3654270b77/glance-httpd/0.log" Dec 06 04:44:03 crc kubenswrapper[4802]: I1206 04:44:03.089368 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_02888e6b-6185-45b8-84b7-8b3654270b77/glance-log/0.log" Dec 06 04:44:03 crc kubenswrapper[4802]: I1206 04:44:03.227716 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_19c7331e-7c80-4aa6-8a9f-2a7f878c79f2/glance-log/0.log" Dec 06 04:44:03 crc kubenswrapper[4802]: I1206 04:44:03.262471 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_19c7331e-7c80-4aa6-8a9f-2a7f878c79f2/glance-httpd/0.log" Dec 06 04:44:03 crc kubenswrapper[4802]: I1206 04:44:03.450297 4802 scope.go:117] "RemoveContainer" containerID="3037e38e5a7009f353fa7ac5ea266cd2366825eeb99656721b87c6c25f49afdb" Dec 06 04:44:03 crc kubenswrapper[4802]: E1206 04:44:03.450566 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:44:03 crc kubenswrapper[4802]: I1206 04:44:03.968332 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-6ddd7d8889-js75h_71fc6de6-20c6-49fe-9780-3cfdb0002c62/heat-api/0.log" Dec 06 04:44:03 crc kubenswrapper[4802]: I1206 04:44:03.972617 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-55fc64bfdd-779rm_a86ea592-5d80-4415-8c1f-5471274999d4/heat-engine/0.log" Dec 06 04:44:03 crc kubenswrapper[4802]: I1206 04:44:03.973516 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-788b9d8f4c-rk4vn_10210aa2-6d6b-4422-ad35-32b04e1444f2/heat-cfnapi/0.log" Dec 06 04:44:04 crc kubenswrapper[4802]: I1206 04:44:04.090163 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g_79de9688-9efb-44b4-9518-259e1f58b5b4/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 04:44:04 crc kubenswrapper[4802]: I1206 04:44:04.227305 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-g9qlw_b6c3bbd2-7993-420e-91d7-ffc72b1d2439/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 04:44:04 crc kubenswrapper[4802]: I1206 04:44:04.672370 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-75dc574896-4vrbs_bdd2c50b-481a-4ab7-90ce-86cb69efd255/keystone-api/0.log" Dec 06 04:44:04 crc kubenswrapper[4802]: I1206 04:44:04.699866 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29416561-rnfxm_613478db-c10d-476b-b66d-4ae3572067ff/keystone-cron/0.log" Dec 06 04:44:04 crc kubenswrapper[4802]: I1206 04:44:04.887804 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-tckw2_18179f2f-334f-48c2-a6be-6d1c1792e600/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 04:44:04 crc kubenswrapper[4802]: I1206 04:44:04.906301 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_5489ef4c-603a-44ce-9a0e-6ad56188e669/kube-state-metrics/0.log" Dec 06 04:44:05 crc kubenswrapper[4802]: I1206 04:44:05.317449 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-68bd774b7c-9sjk9_8bf882a4-ac26-40da-8bcc-46052a9db457/neutron-httpd/0.log" Dec 06 04:44:05 crc kubenswrapper[4802]: I1206 04:44:05.369305 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-lpc95_64272e53-eccf-4cb7-9176-0bc95c5b7e10/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 04:44:05 crc kubenswrapper[4802]: I1206 04:44:05.385169 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-68bd774b7c-9sjk9_8bf882a4-ac26-40da-8bcc-46052a9db457/neutron-api/0.log" Dec 06 04:44:05 crc kubenswrapper[4802]: I1206 04:44:05.923103 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_2c1b64f2-04f9-4fa9-ba9b-852f1383f6ed/nova-cell0-conductor-conductor/0.log" Dec 06 04:44:05 crc kubenswrapper[4802]: I1206 04:44:05.933229 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_add9c7ee-92a7-4cd8-adbf-02ec103cb6dc/nova-api-log/0.log" Dec 06 04:44:05 crc kubenswrapper[4802]: I1206 04:44:05.967518 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_add9c7ee-92a7-4cd8-adbf-02ec103cb6dc/nova-api-api/0.log" Dec 06 04:44:06 crc kubenswrapper[4802]: I1206 04:44:06.210229 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_eaeadc44-2583-4e10-b0d3-15b754927734/nova-cell1-conductor-conductor/0.log" Dec 06 04:44:06 crc kubenswrapper[4802]: I1206 04:44:06.289124 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_ede13563-3589-4889-a568-55af534782c5/nova-cell1-novncproxy-novncproxy/0.log" Dec 06 04:44:06 crc kubenswrapper[4802]: I1206 04:44:06.488275 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-5plkw_b8e8984b-b60b-48e5-9586-1a1a8d883143/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 04:44:06 crc kubenswrapper[4802]: I1206 04:44:06.609492 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_f1350889-6def-43ce-9827-761796bea156/nova-metadata-log/0.log" Dec 06 04:44:06 crc kubenswrapper[4802]: I1206 04:44:06.924141 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_723826fa-70b3-4be3-9380-42561ca2041b/nova-scheduler-scheduler/0.log" Dec 06 04:44:06 crc kubenswrapper[4802]: I1206 04:44:06.944610 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_db5b3080-8b4d-4232-a7da-66e544c09c60/mysql-bootstrap/0.log" Dec 06 04:44:07 crc kubenswrapper[4802]: I1206 04:44:07.141791 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_db5b3080-8b4d-4232-a7da-66e544c09c60/mysql-bootstrap/0.log" Dec 06 04:44:07 crc kubenswrapper[4802]: I1206 04:44:07.185170 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_db5b3080-8b4d-4232-a7da-66e544c09c60/galera/0.log" Dec 06 04:44:07 crc kubenswrapper[4802]: I1206 04:44:07.344663 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_04b6f630-c38a-411e-b48c-99690653972a/mysql-bootstrap/0.log" Dec 06 04:44:07 crc kubenswrapper[4802]: I1206 04:44:07.558695 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_04b6f630-c38a-411e-b48c-99690653972a/galera/0.log" Dec 06 04:44:07 crc kubenswrapper[4802]: I1206 04:44:07.580668 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_04b6f630-c38a-411e-b48c-99690653972a/mysql-bootstrap/0.log" Dec 06 04:44:07 crc kubenswrapper[4802]: I1206 04:44:07.799056 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_20c275f9-f37c-427a-b8f0-77c35fc880b5/openstackclient/0.log" Dec 06 04:44:07 crc kubenswrapper[4802]: I1206 04:44:07.838041 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_f1350889-6def-43ce-9827-761796bea156/nova-metadata-metadata/0.log" Dec 06 04:44:07 crc kubenswrapper[4802]: I1206 04:44:07.846018 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-7sljg_11e2072e-0127-48c0-b417-ced9cfa9bf64/openstack-network-exporter/0.log" Dec 06 04:44:08 crc kubenswrapper[4802]: I1206 04:44:08.052444 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-49cxn_edb79ab5-ba10-4379-96e9-b2de42f6cbc8/ovsdb-server-init/0.log" Dec 06 04:44:08 crc kubenswrapper[4802]: I1206 04:44:08.241659 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-49cxn_edb79ab5-ba10-4379-96e9-b2de42f6cbc8/ovsdb-server-init/0.log" Dec 06 04:44:08 crc kubenswrapper[4802]: I1206 04:44:08.250453 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-49cxn_edb79ab5-ba10-4379-96e9-b2de42f6cbc8/ovsdb-server/0.log" Dec 06 04:44:08 crc kubenswrapper[4802]: I1206 04:44:08.306011 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-49cxn_edb79ab5-ba10-4379-96e9-b2de42f6cbc8/ovs-vswitchd/0.log" Dec 06 04:44:08 crc kubenswrapper[4802]: I1206 04:44:08.508230 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-9hxgw_e2001db0-90f1-40a7-ba6e-982eabebc117/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 04:44:08 crc kubenswrapper[4802]: I1206 04:44:08.512733 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-sbt8p_6e785e69-07cd-4c5c-8446-3d2a0d7e7b0b/ovn-controller/0.log" Dec 06 04:44:08 crc kubenswrapper[4802]: I1206 04:44:08.716546 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_dca5fd9f-3aaf-4fd7-96d0-7f78c38c12b6/ovn-northd/0.log" Dec 06 04:44:08 crc kubenswrapper[4802]: I1206 04:44:08.721483 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_dca5fd9f-3aaf-4fd7-96d0-7f78c38c12b6/openstack-network-exporter/0.log" Dec 06 04:44:08 crc kubenswrapper[4802]: I1206 04:44:08.780216 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_56f9d065-29e4-424d-bc84-2657eba9929e/openstack-network-exporter/0.log" Dec 06 04:44:08 crc kubenswrapper[4802]: I1206 04:44:08.997575 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_56f9d065-29e4-424d-bc84-2657eba9929e/ovsdbserver-nb/0.log" Dec 06 04:44:09 crc kubenswrapper[4802]: I1206 04:44:09.030155 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_c45b4e0d-c52c-421b-b748-896ebc517784/openstack-network-exporter/0.log" Dec 06 04:44:09 crc kubenswrapper[4802]: I1206 04:44:09.084723 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_c45b4e0d-c52c-421b-b748-896ebc517784/ovsdbserver-sb/0.log" Dec 06 04:44:09 crc kubenswrapper[4802]: I1206 04:44:09.243096 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-788d784b-hjc7w_5800f6d2-346a-4e9a-8585-e839ddfd035f/placement-api/0.log" Dec 06 04:44:09 crc kubenswrapper[4802]: I1206 04:44:09.351434 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-788d784b-hjc7w_5800f6d2-346a-4e9a-8585-e839ddfd035f/placement-log/0.log" Dec 06 04:44:09 crc kubenswrapper[4802]: I1206 04:44:09.408903 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_cdcf8137-9bbe-467f-bbb3-e30f350760a6/setup-container/0.log" Dec 06 04:44:09 crc kubenswrapper[4802]: I1206 04:44:09.583679 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_cdcf8137-9bbe-467f-bbb3-e30f350760a6/rabbitmq/0.log" Dec 06 04:44:09 crc kubenswrapper[4802]: I1206 04:44:09.646973 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_cdcf8137-9bbe-467f-bbb3-e30f350760a6/setup-container/0.log" Dec 06 04:44:09 crc kubenswrapper[4802]: I1206 04:44:09.657153 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_cdc37fb9-aeae-4fe0-b86e-e9d41e48314b/setup-container/0.log" Dec 06 04:44:09 crc kubenswrapper[4802]: I1206 04:44:09.860996 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_cdc37fb9-aeae-4fe0-b86e-e9d41e48314b/setup-container/0.log" Dec 06 04:44:09 crc kubenswrapper[4802]: I1206 04:44:09.906347 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-7hgqr_0de67972-73b4-4f71-aca7-d77ae62d73b3/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 04:44:09 crc kubenswrapper[4802]: I1206 04:44:09.930311 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_cdc37fb9-aeae-4fe0-b86e-e9d41e48314b/rabbitmq/0.log" Dec 06 04:44:10 crc kubenswrapper[4802]: I1206 04:44:10.126199 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-tgsk9_63d8761a-c5bb-41d5-a830-f58cc99b0837/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 04:44:10 crc kubenswrapper[4802]: I1206 04:44:10.200798 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-jrmv7_44aaebb1-a5a3-42e6-aacd-04292e80f33d/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 04:44:10 crc kubenswrapper[4802]: I1206 04:44:10.340043 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-4clcr_bcbbe04a-d1ab-4a9b-a4ff-11806f4928c8/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 04:44:10 crc kubenswrapper[4802]: I1206 04:44:10.443026 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-hgx7d_815e5d24-83cc-4bbe-b602-6cd1da82039b/ssh-known-hosts-edpm-deployment/0.log" Dec 06 04:44:10 crc kubenswrapper[4802]: I1206 04:44:10.677170 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-686fbdfd7f-sfrn5_4985762a-1d53-40ad-85da-c6728bdbaa3d/proxy-server/0.log" Dec 06 04:44:10 crc kubenswrapper[4802]: I1206 04:44:10.720885 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-686fbdfd7f-sfrn5_4985762a-1d53-40ad-85da-c6728bdbaa3d/proxy-httpd/0.log" Dec 06 04:44:10 crc kubenswrapper[4802]: I1206 04:44:10.802683 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-flhmv_de4cab50-e92b-4282-bd9d-1b9439373aad/swift-ring-rebalance/0.log" Dec 06 04:44:10 crc kubenswrapper[4802]: I1206 04:44:10.956605 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_508d9d64-dd16-4d21-b492-052d123f1a6a/account-auditor/0.log" Dec 06 04:44:10 crc kubenswrapper[4802]: I1206 04:44:10.987640 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_508d9d64-dd16-4d21-b492-052d123f1a6a/account-reaper/0.log" Dec 06 04:44:11 crc kubenswrapper[4802]: I1206 04:44:11.090065 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_508d9d64-dd16-4d21-b492-052d123f1a6a/account-replicator/0.log" Dec 06 04:44:11 crc kubenswrapper[4802]: I1206 04:44:11.160795 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_508d9d64-dd16-4d21-b492-052d123f1a6a/account-server/0.log" Dec 06 04:44:11 crc kubenswrapper[4802]: I1206 04:44:11.169336 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_508d9d64-dd16-4d21-b492-052d123f1a6a/container-auditor/0.log" Dec 06 04:44:11 crc kubenswrapper[4802]: I1206 04:44:11.308418 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_508d9d64-dd16-4d21-b492-052d123f1a6a/container-replicator/0.log" Dec 06 04:44:11 crc kubenswrapper[4802]: I1206 04:44:11.323173 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_508d9d64-dd16-4d21-b492-052d123f1a6a/container-server/0.log" Dec 06 04:44:11 crc kubenswrapper[4802]: I1206 04:44:11.400580 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_508d9d64-dd16-4d21-b492-052d123f1a6a/container-updater/0.log" Dec 06 04:44:11 crc kubenswrapper[4802]: I1206 04:44:11.414007 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_508d9d64-dd16-4d21-b492-052d123f1a6a/object-auditor/0.log" Dec 06 04:44:11 crc kubenswrapper[4802]: I1206 04:44:11.540594 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_508d9d64-dd16-4d21-b492-052d123f1a6a/object-expirer/0.log" Dec 06 04:44:11 crc kubenswrapper[4802]: I1206 04:44:11.560704 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_508d9d64-dd16-4d21-b492-052d123f1a6a/object-replicator/0.log" Dec 06 04:44:11 crc kubenswrapper[4802]: I1206 04:44:11.602199 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_508d9d64-dd16-4d21-b492-052d123f1a6a/object-server/0.log" Dec 06 04:44:11 crc kubenswrapper[4802]: I1206 04:44:11.641239 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_508d9d64-dd16-4d21-b492-052d123f1a6a/object-updater/0.log" Dec 06 04:44:11 crc kubenswrapper[4802]: I1206 04:44:11.782513 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_508d9d64-dd16-4d21-b492-052d123f1a6a/rsync/0.log" Dec 06 04:44:11 crc kubenswrapper[4802]: I1206 04:44:11.848103 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_508d9d64-dd16-4d21-b492-052d123f1a6a/swift-recon-cron/0.log" Dec 06 04:44:11 crc kubenswrapper[4802]: I1206 04:44:11.941471 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-ns78h_565b62ff-4b9a-4e53-b61c-d6c492d7b253/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 04:44:12 crc kubenswrapper[4802]: I1206 04:44:12.097512 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_35caa18c-c7c4-4e74-9c6d-053c38327afc/test-operator-logs-container/0.log" Dec 06 04:44:12 crc kubenswrapper[4802]: I1206 04:44:12.143556 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_63c9a1aa-a443-4071-b60e-317a7b81395e/tempest-tests-tempest-tests-runner/0.log" Dec 06 04:44:12 crc kubenswrapper[4802]: I1206 04:44:12.325927 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-7nbhk_5fd4f8df-59a8-47cb-b9e7-4162a5811ed7/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 04:44:16 crc kubenswrapper[4802]: I1206 04:44:16.449625 4802 scope.go:117] "RemoveContainer" containerID="3037e38e5a7009f353fa7ac5ea266cd2366825eeb99656721b87c6c25f49afdb" Dec 06 04:44:16 crc kubenswrapper[4802]: E1206 04:44:16.450287 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:44:19 crc kubenswrapper[4802]: I1206 04:44:19.450821 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_654f214b-8467-41d2-93f4-3c18377653b9/memcached/0.log" Dec 06 04:44:27 crc kubenswrapper[4802]: I1206 04:44:27.457293 4802 scope.go:117] "RemoveContainer" containerID="3037e38e5a7009f353fa7ac5ea266cd2366825eeb99656721b87c6c25f49afdb" Dec 06 04:44:27 crc kubenswrapper[4802]: E1206 04:44:27.457910 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:44:37 crc kubenswrapper[4802]: I1206 04:44:37.795665 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_11a37daaca2b5d044ad5e897ef78a9b78d5fd0aec8475f6e59b41a4af1nhddl_e9a7628c-dba2-482e-b66b-e4046bb06ddc/util/0.log" Dec 06 04:44:37 crc kubenswrapper[4802]: I1206 04:44:37.992872 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_11a37daaca2b5d044ad5e897ef78a9b78d5fd0aec8475f6e59b41a4af1nhddl_e9a7628c-dba2-482e-b66b-e4046bb06ddc/util/0.log" Dec 06 04:44:38 crc kubenswrapper[4802]: I1206 04:44:38.038324 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_11a37daaca2b5d044ad5e897ef78a9b78d5fd0aec8475f6e59b41a4af1nhddl_e9a7628c-dba2-482e-b66b-e4046bb06ddc/pull/0.log" Dec 06 04:44:38 crc kubenswrapper[4802]: I1206 04:44:38.040593 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_11a37daaca2b5d044ad5e897ef78a9b78d5fd0aec8475f6e59b41a4af1nhddl_e9a7628c-dba2-482e-b66b-e4046bb06ddc/pull/0.log" Dec 06 04:44:38 crc kubenswrapper[4802]: I1206 04:44:38.172518 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_11a37daaca2b5d044ad5e897ef78a9b78d5fd0aec8475f6e59b41a4af1nhddl_e9a7628c-dba2-482e-b66b-e4046bb06ddc/pull/0.log" Dec 06 04:44:38 crc kubenswrapper[4802]: I1206 04:44:38.201401 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_11a37daaca2b5d044ad5e897ef78a9b78d5fd0aec8475f6e59b41a4af1nhddl_e9a7628c-dba2-482e-b66b-e4046bb06ddc/util/0.log" Dec 06 04:44:38 crc kubenswrapper[4802]: I1206 04:44:38.225465 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_11a37daaca2b5d044ad5e897ef78a9b78d5fd0aec8475f6e59b41a4af1nhddl_e9a7628c-dba2-482e-b66b-e4046bb06ddc/extract/0.log" Dec 06 04:44:38 crc kubenswrapper[4802]: I1206 04:44:38.352837 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-p8kv6_1c3d06b1-b999-4c82-85b5-11bdd996770c/kube-rbac-proxy/0.log" Dec 06 04:44:38 crc kubenswrapper[4802]: I1206 04:44:38.412400 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-xk9cb_309f12a2-5b02-4d5f-9e28-63116ff8bb47/kube-rbac-proxy/0.log" Dec 06 04:44:38 crc kubenswrapper[4802]: I1206 04:44:38.425232 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-p8kv6_1c3d06b1-b999-4c82-85b5-11bdd996770c/manager/0.log" Dec 06 04:44:38 crc kubenswrapper[4802]: I1206 04:44:38.582245 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-xk9cb_309f12a2-5b02-4d5f-9e28-63116ff8bb47/manager/0.log" Dec 06 04:44:38 crc kubenswrapper[4802]: I1206 04:44:38.619388 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-wwgt6_e2bc662f-efe3-4e89-9b0e-e40eba87e41b/kube-rbac-proxy/0.log" Dec 06 04:44:38 crc kubenswrapper[4802]: I1206 04:44:38.660577 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-wwgt6_e2bc662f-efe3-4e89-9b0e-e40eba87e41b/manager/0.log" Dec 06 04:44:38 crc kubenswrapper[4802]: I1206 04:44:38.786883 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-zkp5c_08b2817d-5061-4375-82d4-726108586852/kube-rbac-proxy/0.log" Dec 06 04:44:38 crc kubenswrapper[4802]: I1206 04:44:38.866423 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-zkp5c_08b2817d-5061-4375-82d4-726108586852/manager/0.log" Dec 06 04:44:38 crc kubenswrapper[4802]: I1206 04:44:38.984435 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-55cb9d4b9c-srntd_faf6c2c3-b028-431f-a302-3ad181dd93ae/kube-rbac-proxy/0.log" Dec 06 04:44:39 crc kubenswrapper[4802]: I1206 04:44:39.044729 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-55cb9d4b9c-srntd_faf6c2c3-b028-431f-a302-3ad181dd93ae/manager/0.log" Dec 06 04:44:39 crc kubenswrapper[4802]: I1206 04:44:39.081378 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-gf27t_5f8e1071-aa69-4e21-94c9-f533d55f2cac/kube-rbac-proxy/0.log" Dec 06 04:44:39 crc kubenswrapper[4802]: I1206 04:44:39.186944 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-gf27t_5f8e1071-aa69-4e21-94c9-f533d55f2cac/manager/0.log" Dec 06 04:44:39 crc kubenswrapper[4802]: I1206 04:44:39.274808 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-hpsfv_19f0c68c-32a3-41e2-a8ca-5ab0dc28d748/kube-rbac-proxy/0.log" Dec 06 04:44:39 crc kubenswrapper[4802]: I1206 04:44:39.499599 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-hpsfv_19f0c68c-32a3-41e2-a8ca-5ab0dc28d748/manager/0.log" Dec 06 04:44:39 crc kubenswrapper[4802]: I1206 04:44:39.507136 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-c4sr8_d71b7639-ee30-4323-8be0-a16d844d259f/manager/0.log" Dec 06 04:44:39 crc kubenswrapper[4802]: I1206 04:44:39.542991 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-c4sr8_d71b7639-ee30-4323-8be0-a16d844d259f/kube-rbac-proxy/0.log" Dec 06 04:44:39 crc kubenswrapper[4802]: I1206 04:44:39.718785 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-h7629_c2e52d5b-b8cf-41f2-933e-af31359dc69d/kube-rbac-proxy/0.log" Dec 06 04:44:39 crc kubenswrapper[4802]: I1206 04:44:39.753430 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-h7629_c2e52d5b-b8cf-41f2-933e-af31359dc69d/manager/0.log" Dec 06 04:44:39 crc kubenswrapper[4802]: I1206 04:44:39.934891 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-74c6z_1d3c0479-80bf-46af-a261-42b3834807f1/kube-rbac-proxy/0.log" Dec 06 04:44:39 crc kubenswrapper[4802]: I1206 04:44:39.951131 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-74c6z_1d3c0479-80bf-46af-a261-42b3834807f1/manager/0.log" Dec 06 04:44:39 crc kubenswrapper[4802]: I1206 04:44:39.963955 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-bkfc6_4460e15d-1d9a-4814-a7d2-042c6a84057f/kube-rbac-proxy/0.log" Dec 06 04:44:40 crc kubenswrapper[4802]: I1206 04:44:40.144634 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-hlx24_a076799b-6631-474a-b552-9cf2261c377a/kube-rbac-proxy/0.log" Dec 06 04:44:40 crc kubenswrapper[4802]: I1206 04:44:40.147340 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-bkfc6_4460e15d-1d9a-4814-a7d2-042c6a84057f/manager/0.log" Dec 06 04:44:40 crc kubenswrapper[4802]: I1206 04:44:40.185743 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-hlx24_a076799b-6631-474a-b552-9cf2261c377a/manager/0.log" Dec 06 04:44:40 crc kubenswrapper[4802]: I1206 04:44:40.315263 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-4m2vm_ef258927-21c4-4147-8fb5-3db5d78658a3/kube-rbac-proxy/0.log" Dec 06 04:44:40 crc kubenswrapper[4802]: I1206 04:44:40.427233 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-4m2vm_ef258927-21c4-4147-8fb5-3db5d78658a3/manager/0.log" Dec 06 04:44:40 crc kubenswrapper[4802]: I1206 04:44:40.450739 4802 scope.go:117] "RemoveContainer" containerID="3037e38e5a7009f353fa7ac5ea266cd2366825eeb99656721b87c6c25f49afdb" Dec 06 04:44:40 crc kubenswrapper[4802]: E1206 04:44:40.450961 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:44:40 crc kubenswrapper[4802]: I1206 04:44:40.501247 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-lk8bf_8555a40e-f3d3-4627-8867-6db5ca5ed06a/manager/0.log" Dec 06 04:44:40 crc kubenswrapper[4802]: I1206 04:44:40.513339 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-lk8bf_8555a40e-f3d3-4627-8867-6db5ca5ed06a/kube-rbac-proxy/0.log" Dec 06 04:44:40 crc kubenswrapper[4802]: I1206 04:44:40.586028 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4pwn82_05736c59-c4db-4cfe-addd-1b29c552596e/kube-rbac-proxy/0.log" Dec 06 04:44:40 crc kubenswrapper[4802]: I1206 04:44:40.710263 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4pwn82_05736c59-c4db-4cfe-addd-1b29c552596e/manager/0.log" Dec 06 04:44:41 crc kubenswrapper[4802]: I1206 04:44:41.004253 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-57d4dc878b-qnwf9_31962197-d27b-41af-9469-27ae39fb5d4f/operator/0.log" Dec 06 04:44:41 crc kubenswrapper[4802]: I1206 04:44:41.048870 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-xd8sc_43e371ec-a36d-4082-827e-21518dc7e9c1/registry-server/0.log" Dec 06 04:44:41 crc kubenswrapper[4802]: I1206 04:44:41.223833 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-8q2cg_119bce90-0fad-424d-89cf-2ca06d7126ae/kube-rbac-proxy/0.log" Dec 06 04:44:41 crc kubenswrapper[4802]: I1206 04:44:41.355545 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-8q2cg_119bce90-0fad-424d-89cf-2ca06d7126ae/manager/0.log" Dec 06 04:44:41 crc kubenswrapper[4802]: I1206 04:44:41.481979 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-4x9rb_79efe004-97e7-4586-b48e-08ce9285bb79/kube-rbac-proxy/0.log" Dec 06 04:44:41 crc kubenswrapper[4802]: I1206 04:44:41.558252 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-4x9rb_79efe004-97e7-4586-b48e-08ce9285bb79/manager/0.log" Dec 06 04:44:41 crc kubenswrapper[4802]: I1206 04:44:41.729787 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-b5wv6_a3bed886-a616-4733-947a-288ada88dc74/operator/0.log" Dec 06 04:44:41 crc kubenswrapper[4802]: I1206 04:44:41.803807 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-mmh24_11885e74-7bc3-4207-b046-9e94ab51e499/kube-rbac-proxy/0.log" Dec 06 04:44:41 crc kubenswrapper[4802]: I1206 04:44:41.916837 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-mmh24_11885e74-7bc3-4207-b046-9e94ab51e499/manager/0.log" Dec 06 04:44:41 crc kubenswrapper[4802]: I1206 04:44:41.994098 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-nd42v_18ad32ab-222c-41f9-8409-90fd51178f2d/kube-rbac-proxy/0.log" Dec 06 04:44:42 crc kubenswrapper[4802]: I1206 04:44:42.041451 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-77f4d6cd-9s876_5f6a5e2f-8db1-4d31-8a2d-5810c5e62083/manager/0.log" Dec 06 04:44:42 crc kubenswrapper[4802]: I1206 04:44:42.119537 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-nd42v_18ad32ab-222c-41f9-8409-90fd51178f2d/manager/0.log" Dec 06 04:44:42 crc kubenswrapper[4802]: I1206 04:44:42.194341 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-q6k7m_e9a14170-cafd-433c-83a4-22793b21d376/kube-rbac-proxy/0.log" Dec 06 04:44:42 crc kubenswrapper[4802]: I1206 04:44:42.228967 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-q6k7m_e9a14170-cafd-433c-83a4-22793b21d376/manager/0.log" Dec 06 04:44:42 crc kubenswrapper[4802]: I1206 04:44:42.305961 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-w64dr_91b87773-2f3d-4e35-9a13-708f07cb2bbf/kube-rbac-proxy/0.log" Dec 06 04:44:42 crc kubenswrapper[4802]: I1206 04:44:42.358452 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-w64dr_91b87773-2f3d-4e35-9a13-708f07cb2bbf/manager/0.log" Dec 06 04:44:53 crc kubenswrapper[4802]: I1206 04:44:53.455744 4802 scope.go:117] "RemoveContainer" containerID="3037e38e5a7009f353fa7ac5ea266cd2366825eeb99656721b87c6c25f49afdb" Dec 06 04:44:53 crc kubenswrapper[4802]: E1206 04:44:53.456469 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:45:00 crc kubenswrapper[4802]: I1206 04:45:00.173651 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416605-k27sx"] Dec 06 04:45:00 crc kubenswrapper[4802]: E1206 04:45:00.174701 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0663503c-e362-49bd-b5e1-a5f3f35bd04b" containerName="container-00" Dec 06 04:45:00 crc kubenswrapper[4802]: I1206 04:45:00.174720 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="0663503c-e362-49bd-b5e1-a5f3f35bd04b" containerName="container-00" Dec 06 04:45:00 crc kubenswrapper[4802]: I1206 04:45:00.175000 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="0663503c-e362-49bd-b5e1-a5f3f35bd04b" containerName="container-00" Dec 06 04:45:00 crc kubenswrapper[4802]: I1206 04:45:00.175848 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416605-k27sx" Dec 06 04:45:00 crc kubenswrapper[4802]: I1206 04:45:00.179081 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 04:45:00 crc kubenswrapper[4802]: I1206 04:45:00.179189 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 04:45:00 crc kubenswrapper[4802]: I1206 04:45:00.185651 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416605-k27sx"] Dec 06 04:45:00 crc kubenswrapper[4802]: I1206 04:45:00.300896 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/605adff1-0732-48d8-bd0a-a499309a6da5-config-volume\") pod \"collect-profiles-29416605-k27sx\" (UID: \"605adff1-0732-48d8-bd0a-a499309a6da5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416605-k27sx" Dec 06 04:45:00 crc kubenswrapper[4802]: I1206 04:45:00.301037 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/605adff1-0732-48d8-bd0a-a499309a6da5-secret-volume\") pod \"collect-profiles-29416605-k27sx\" (UID: \"605adff1-0732-48d8-bd0a-a499309a6da5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416605-k27sx" Dec 06 04:45:00 crc kubenswrapper[4802]: I1206 04:45:00.301088 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7ft9\" (UniqueName: \"kubernetes.io/projected/605adff1-0732-48d8-bd0a-a499309a6da5-kube-api-access-c7ft9\") pod \"collect-profiles-29416605-k27sx\" (UID: \"605adff1-0732-48d8-bd0a-a499309a6da5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416605-k27sx" Dec 06 04:45:00 crc kubenswrapper[4802]: I1206 04:45:00.402575 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/605adff1-0732-48d8-bd0a-a499309a6da5-secret-volume\") pod \"collect-profiles-29416605-k27sx\" (UID: \"605adff1-0732-48d8-bd0a-a499309a6da5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416605-k27sx" Dec 06 04:45:00 crc kubenswrapper[4802]: I1206 04:45:00.402642 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7ft9\" (UniqueName: \"kubernetes.io/projected/605adff1-0732-48d8-bd0a-a499309a6da5-kube-api-access-c7ft9\") pod \"collect-profiles-29416605-k27sx\" (UID: \"605adff1-0732-48d8-bd0a-a499309a6da5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416605-k27sx" Dec 06 04:45:00 crc kubenswrapper[4802]: I1206 04:45:00.402730 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/605adff1-0732-48d8-bd0a-a499309a6da5-config-volume\") pod \"collect-profiles-29416605-k27sx\" (UID: \"605adff1-0732-48d8-bd0a-a499309a6da5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416605-k27sx" Dec 06 04:45:00 crc kubenswrapper[4802]: I1206 04:45:00.404175 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/605adff1-0732-48d8-bd0a-a499309a6da5-config-volume\") pod \"collect-profiles-29416605-k27sx\" (UID: \"605adff1-0732-48d8-bd0a-a499309a6da5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416605-k27sx" Dec 06 04:45:00 crc kubenswrapper[4802]: I1206 04:45:00.411293 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/605adff1-0732-48d8-bd0a-a499309a6da5-secret-volume\") pod \"collect-profiles-29416605-k27sx\" (UID: \"605adff1-0732-48d8-bd0a-a499309a6da5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416605-k27sx" Dec 06 04:45:00 crc kubenswrapper[4802]: I1206 04:45:00.426525 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7ft9\" (UniqueName: \"kubernetes.io/projected/605adff1-0732-48d8-bd0a-a499309a6da5-kube-api-access-c7ft9\") pod \"collect-profiles-29416605-k27sx\" (UID: \"605adff1-0732-48d8-bd0a-a499309a6da5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416605-k27sx" Dec 06 04:45:00 crc kubenswrapper[4802]: I1206 04:45:00.495275 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416605-k27sx" Dec 06 04:45:00 crc kubenswrapper[4802]: I1206 04:45:00.910269 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-6d29n_1a34c64b-194e-41a6-8b55-0a87187c4cfc/control-plane-machine-set-operator/0.log" Dec 06 04:45:01 crc kubenswrapper[4802]: I1206 04:45:00.996713 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416605-k27sx"] Dec 06 04:45:01 crc kubenswrapper[4802]: I1206 04:45:01.007582 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-9s5r8_eb15a464-d0e8-4e49-b018-14353cc9d6cb/kube-rbac-proxy/0.log" Dec 06 04:45:01 crc kubenswrapper[4802]: I1206 04:45:01.115215 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-9s5r8_eb15a464-d0e8-4e49-b018-14353cc9d6cb/machine-api-operator/0.log" Dec 06 04:45:01 crc kubenswrapper[4802]: I1206 04:45:01.926019 4802 generic.go:334] "Generic (PLEG): container finished" podID="605adff1-0732-48d8-bd0a-a499309a6da5" containerID="3309a32b91a9611b166449939d5c4ba1c43d546266ce2aa8f48352144970c7bb" exitCode=0 Dec 06 04:45:01 crc kubenswrapper[4802]: I1206 04:45:01.926118 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416605-k27sx" event={"ID":"605adff1-0732-48d8-bd0a-a499309a6da5","Type":"ContainerDied","Data":"3309a32b91a9611b166449939d5c4ba1c43d546266ce2aa8f48352144970c7bb"} Dec 06 04:45:01 crc kubenswrapper[4802]: I1206 04:45:01.926336 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416605-k27sx" event={"ID":"605adff1-0732-48d8-bd0a-a499309a6da5","Type":"ContainerStarted","Data":"6d303f646f7bc50f29dd6d5a1eb193d4d812f916d56c60431f8538614f3ef082"} Dec 06 04:45:03 crc kubenswrapper[4802]: I1206 04:45:03.360458 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416605-k27sx" Dec 06 04:45:03 crc kubenswrapper[4802]: I1206 04:45:03.468955 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/605adff1-0732-48d8-bd0a-a499309a6da5-secret-volume\") pod \"605adff1-0732-48d8-bd0a-a499309a6da5\" (UID: \"605adff1-0732-48d8-bd0a-a499309a6da5\") " Dec 06 04:45:03 crc kubenswrapper[4802]: I1206 04:45:03.469438 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/605adff1-0732-48d8-bd0a-a499309a6da5-config-volume\") pod \"605adff1-0732-48d8-bd0a-a499309a6da5\" (UID: \"605adff1-0732-48d8-bd0a-a499309a6da5\") " Dec 06 04:45:03 crc kubenswrapper[4802]: I1206 04:45:03.469575 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c7ft9\" (UniqueName: \"kubernetes.io/projected/605adff1-0732-48d8-bd0a-a499309a6da5-kube-api-access-c7ft9\") pod \"605adff1-0732-48d8-bd0a-a499309a6da5\" (UID: \"605adff1-0732-48d8-bd0a-a499309a6da5\") " Dec 06 04:45:03 crc kubenswrapper[4802]: I1206 04:45:03.472101 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/605adff1-0732-48d8-bd0a-a499309a6da5-config-volume" (OuterVolumeSpecName: "config-volume") pod "605adff1-0732-48d8-bd0a-a499309a6da5" (UID: "605adff1-0732-48d8-bd0a-a499309a6da5"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:45:03 crc kubenswrapper[4802]: I1206 04:45:03.475743 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/605adff1-0732-48d8-bd0a-a499309a6da5-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "605adff1-0732-48d8-bd0a-a499309a6da5" (UID: "605adff1-0732-48d8-bd0a-a499309a6da5"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:45:03 crc kubenswrapper[4802]: I1206 04:45:03.476037 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/605adff1-0732-48d8-bd0a-a499309a6da5-kube-api-access-c7ft9" (OuterVolumeSpecName: "kube-api-access-c7ft9") pod "605adff1-0732-48d8-bd0a-a499309a6da5" (UID: "605adff1-0732-48d8-bd0a-a499309a6da5"). InnerVolumeSpecName "kube-api-access-c7ft9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:45:03 crc kubenswrapper[4802]: I1206 04:45:03.571622 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c7ft9\" (UniqueName: \"kubernetes.io/projected/605adff1-0732-48d8-bd0a-a499309a6da5-kube-api-access-c7ft9\") on node \"crc\" DevicePath \"\"" Dec 06 04:45:03 crc kubenswrapper[4802]: I1206 04:45:03.571666 4802 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/605adff1-0732-48d8-bd0a-a499309a6da5-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 04:45:03 crc kubenswrapper[4802]: I1206 04:45:03.571680 4802 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/605adff1-0732-48d8-bd0a-a499309a6da5-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 04:45:03 crc kubenswrapper[4802]: I1206 04:45:03.944335 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416605-k27sx" event={"ID":"605adff1-0732-48d8-bd0a-a499309a6da5","Type":"ContainerDied","Data":"6d303f646f7bc50f29dd6d5a1eb193d4d812f916d56c60431f8538614f3ef082"} Dec 06 04:45:03 crc kubenswrapper[4802]: I1206 04:45:03.944396 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6d303f646f7bc50f29dd6d5a1eb193d4d812f916d56c60431f8538614f3ef082" Dec 06 04:45:03 crc kubenswrapper[4802]: I1206 04:45:03.944399 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416605-k27sx" Dec 06 04:45:04 crc kubenswrapper[4802]: I1206 04:45:04.442896 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416560-zzktw"] Dec 06 04:45:04 crc kubenswrapper[4802]: I1206 04:45:04.450490 4802 scope.go:117] "RemoveContainer" containerID="3037e38e5a7009f353fa7ac5ea266cd2366825eeb99656721b87c6c25f49afdb" Dec 06 04:45:04 crc kubenswrapper[4802]: E1206 04:45:04.450736 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:45:04 crc kubenswrapper[4802]: I1206 04:45:04.451097 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416560-zzktw"] Dec 06 04:45:05 crc kubenswrapper[4802]: I1206 04:45:05.462328 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5898b86d-3906-4c87-8509-8c37fe50f544" path="/var/lib/kubelet/pods/5898b86d-3906-4c87-8509-8c37fe50f544/volumes" Dec 06 04:45:12 crc kubenswrapper[4802]: I1206 04:45:12.928136 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-sdbpz_3f3aa4aa-3196-4dd8-802f-73235899e452/cert-manager-controller/0.log" Dec 06 04:45:13 crc kubenswrapper[4802]: I1206 04:45:13.091033 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-7r6gm_5aaa82f6-dd8a-4ca3-8e3d-b574617c240a/cert-manager-cainjector/0.log" Dec 06 04:45:13 crc kubenswrapper[4802]: I1206 04:45:13.161931 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-54qqk_a50fe25c-7be9-42ea-a81b-e6ba817043e8/cert-manager-webhook/0.log" Dec 06 04:45:18 crc kubenswrapper[4802]: I1206 04:45:18.449616 4802 scope.go:117] "RemoveContainer" containerID="3037e38e5a7009f353fa7ac5ea266cd2366825eeb99656721b87c6c25f49afdb" Dec 06 04:45:19 crc kubenswrapper[4802]: I1206 04:45:19.075692 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" event={"ID":"6df38316-e0d3-4018-8d27-3620eba3a68d","Type":"ContainerStarted","Data":"659eedd07500dc9cedfea13e32706d57d3d8130ed952a340282ebbc31bb990e2"} Dec 06 04:45:23 crc kubenswrapper[4802]: I1206 04:45:23.107687 4802 scope.go:117] "RemoveContainer" containerID="b44f058acf9a6945816306974c5c4f9921892f61e3f4067def14e9ff7e313600" Dec 06 04:45:28 crc kubenswrapper[4802]: I1206 04:45:28.150960 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-8wwln_db89a256-eef7-476e-ab22-755b4d6bb780/nmstate-console-plugin/0.log" Dec 06 04:45:28 crc kubenswrapper[4802]: I1206 04:45:28.277824 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-swx72_485510d8-9971-4648-bb7c-120875a7c00d/nmstate-handler/0.log" Dec 06 04:45:28 crc kubenswrapper[4802]: I1206 04:45:28.327409 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-d8q2s_75129a1d-9a3e-4efb-b2bf-bf353a4f6ea2/kube-rbac-proxy/0.log" Dec 06 04:45:28 crc kubenswrapper[4802]: I1206 04:45:28.342843 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-d8q2s_75129a1d-9a3e-4efb-b2bf-bf353a4f6ea2/nmstate-metrics/0.log" Dec 06 04:45:28 crc kubenswrapper[4802]: I1206 04:45:28.775576 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-l8t5h_a1e67ef9-99a6-4d92-a5d4-6c00837b8993/nmstate-operator/0.log" Dec 06 04:45:28 crc kubenswrapper[4802]: I1206 04:45:28.810898 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-2trjk_92122afa-11ef-42cf-9b97-86d3a41c6e13/nmstate-webhook/0.log" Dec 06 04:45:43 crc kubenswrapper[4802]: I1206 04:45:43.934449 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-46vm8_3a078497-d2a1-4be5-87ee-ba151c992ecd/kube-rbac-proxy/0.log" Dec 06 04:45:44 crc kubenswrapper[4802]: I1206 04:45:44.136440 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nzpzt_9ff1fa79-a414-498e-9f64-659502eb6ae7/cp-frr-files/0.log" Dec 06 04:45:44 crc kubenswrapper[4802]: I1206 04:45:44.141036 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-46vm8_3a078497-d2a1-4be5-87ee-ba151c992ecd/controller/0.log" Dec 06 04:45:44 crc kubenswrapper[4802]: I1206 04:45:44.367828 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nzpzt_9ff1fa79-a414-498e-9f64-659502eb6ae7/cp-reloader/0.log" Dec 06 04:45:44 crc kubenswrapper[4802]: I1206 04:45:44.395070 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nzpzt_9ff1fa79-a414-498e-9f64-659502eb6ae7/cp-metrics/0.log" Dec 06 04:45:44 crc kubenswrapper[4802]: I1206 04:45:44.411181 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nzpzt_9ff1fa79-a414-498e-9f64-659502eb6ae7/cp-frr-files/0.log" Dec 06 04:45:44 crc kubenswrapper[4802]: I1206 04:45:44.417023 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nzpzt_9ff1fa79-a414-498e-9f64-659502eb6ae7/cp-reloader/0.log" Dec 06 04:45:44 crc kubenswrapper[4802]: I1206 04:45:44.550531 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nzpzt_9ff1fa79-a414-498e-9f64-659502eb6ae7/cp-frr-files/0.log" Dec 06 04:45:44 crc kubenswrapper[4802]: I1206 04:45:44.576079 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nzpzt_9ff1fa79-a414-498e-9f64-659502eb6ae7/cp-reloader/0.log" Dec 06 04:45:44 crc kubenswrapper[4802]: I1206 04:45:44.624091 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nzpzt_9ff1fa79-a414-498e-9f64-659502eb6ae7/cp-metrics/0.log" Dec 06 04:45:44 crc kubenswrapper[4802]: I1206 04:45:44.657763 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nzpzt_9ff1fa79-a414-498e-9f64-659502eb6ae7/cp-metrics/0.log" Dec 06 04:45:44 crc kubenswrapper[4802]: I1206 04:45:44.837922 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nzpzt_9ff1fa79-a414-498e-9f64-659502eb6ae7/cp-reloader/0.log" Dec 06 04:45:44 crc kubenswrapper[4802]: I1206 04:45:44.838624 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nzpzt_9ff1fa79-a414-498e-9f64-659502eb6ae7/cp-frr-files/0.log" Dec 06 04:45:44 crc kubenswrapper[4802]: I1206 04:45:44.842257 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nzpzt_9ff1fa79-a414-498e-9f64-659502eb6ae7/controller/0.log" Dec 06 04:45:44 crc kubenswrapper[4802]: I1206 04:45:44.856976 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nzpzt_9ff1fa79-a414-498e-9f64-659502eb6ae7/cp-metrics/0.log" Dec 06 04:45:45 crc kubenswrapper[4802]: I1206 04:45:45.030127 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nzpzt_9ff1fa79-a414-498e-9f64-659502eb6ae7/frr-metrics/0.log" Dec 06 04:45:45 crc kubenswrapper[4802]: I1206 04:45:45.061570 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nzpzt_9ff1fa79-a414-498e-9f64-659502eb6ae7/kube-rbac-proxy/0.log" Dec 06 04:45:45 crc kubenswrapper[4802]: I1206 04:45:45.091950 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nzpzt_9ff1fa79-a414-498e-9f64-659502eb6ae7/kube-rbac-proxy-frr/0.log" Dec 06 04:45:45 crc kubenswrapper[4802]: I1206 04:45:45.286540 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nzpzt_9ff1fa79-a414-498e-9f64-659502eb6ae7/reloader/0.log" Dec 06 04:45:45 crc kubenswrapper[4802]: I1206 04:45:45.323270 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-bjbtx_94e7ed3e-c72c-4130-81e6-ef954e2e62dd/frr-k8s-webhook-server/0.log" Dec 06 04:45:45 crc kubenswrapper[4802]: I1206 04:45:45.603359 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5c78fd4b89-5m46c_392ad0d8-35bb-40df-abf2-7a98117f24f7/manager/0.log" Dec 06 04:45:45 crc kubenswrapper[4802]: I1206 04:45:45.781566 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-868c9d9dfb-lxrtk_9b4b22b1-625a-4bd0-9c82-a142a6f7f806/webhook-server/0.log" Dec 06 04:45:45 crc kubenswrapper[4802]: I1206 04:45:45.840928 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-rlwm5_1e43b1ba-4d5f-47a4-8f9c-359875710a1e/kube-rbac-proxy/0.log" Dec 06 04:45:46 crc kubenswrapper[4802]: I1206 04:45:46.505057 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-rlwm5_1e43b1ba-4d5f-47a4-8f9c-359875710a1e/speaker/0.log" Dec 06 04:45:46 crc kubenswrapper[4802]: I1206 04:45:46.517572 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nzpzt_9ff1fa79-a414-498e-9f64-659502eb6ae7/frr/0.log" Dec 06 04:46:00 crc kubenswrapper[4802]: I1206 04:46:00.030261 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkz86h_7dfbb06c-f867-4d00-aef7-e731fa8579a7/util/0.log" Dec 06 04:46:00 crc kubenswrapper[4802]: I1206 04:46:00.209815 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkz86h_7dfbb06c-f867-4d00-aef7-e731fa8579a7/pull/0.log" Dec 06 04:46:00 crc kubenswrapper[4802]: I1206 04:46:00.214613 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkz86h_7dfbb06c-f867-4d00-aef7-e731fa8579a7/util/0.log" Dec 06 04:46:00 crc kubenswrapper[4802]: I1206 04:46:00.302174 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkz86h_7dfbb06c-f867-4d00-aef7-e731fa8579a7/pull/0.log" Dec 06 04:46:00 crc kubenswrapper[4802]: I1206 04:46:00.469491 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkz86h_7dfbb06c-f867-4d00-aef7-e731fa8579a7/util/0.log" Dec 06 04:46:00 crc kubenswrapper[4802]: I1206 04:46:00.469961 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkz86h_7dfbb06c-f867-4d00-aef7-e731fa8579a7/extract/0.log" Dec 06 04:46:00 crc kubenswrapper[4802]: I1206 04:46:00.483903 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkz86h_7dfbb06c-f867-4d00-aef7-e731fa8579a7/pull/0.log" Dec 06 04:46:00 crc kubenswrapper[4802]: I1206 04:46:00.672687 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838j9dk_207dc360-1c22-429b-b49e-f27bdd02bc03/util/0.log" Dec 06 04:46:00 crc kubenswrapper[4802]: I1206 04:46:00.857786 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838j9dk_207dc360-1c22-429b-b49e-f27bdd02bc03/util/0.log" Dec 06 04:46:00 crc kubenswrapper[4802]: I1206 04:46:00.937568 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838j9dk_207dc360-1c22-429b-b49e-f27bdd02bc03/pull/0.log" Dec 06 04:46:00 crc kubenswrapper[4802]: I1206 04:46:00.942428 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838j9dk_207dc360-1c22-429b-b49e-f27bdd02bc03/pull/0.log" Dec 06 04:46:01 crc kubenswrapper[4802]: I1206 04:46:01.531982 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838j9dk_207dc360-1c22-429b-b49e-f27bdd02bc03/util/0.log" Dec 06 04:46:01 crc kubenswrapper[4802]: I1206 04:46:01.556725 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838j9dk_207dc360-1c22-429b-b49e-f27bdd02bc03/extract/0.log" Dec 06 04:46:01 crc kubenswrapper[4802]: I1206 04:46:01.561430 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838j9dk_207dc360-1c22-429b-b49e-f27bdd02bc03/pull/0.log" Dec 06 04:46:01 crc kubenswrapper[4802]: I1206 04:46:01.713989 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-skt5b_042f382d-836e-4cc2-9065-352a210f6593/extract-utilities/0.log" Dec 06 04:46:01 crc kubenswrapper[4802]: I1206 04:46:01.902249 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-skt5b_042f382d-836e-4cc2-9065-352a210f6593/extract-utilities/0.log" Dec 06 04:46:01 crc kubenswrapper[4802]: I1206 04:46:01.916736 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-skt5b_042f382d-836e-4cc2-9065-352a210f6593/extract-content/0.log" Dec 06 04:46:01 crc kubenswrapper[4802]: I1206 04:46:01.920708 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-skt5b_042f382d-836e-4cc2-9065-352a210f6593/extract-content/0.log" Dec 06 04:46:02 crc kubenswrapper[4802]: I1206 04:46:02.117087 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-skt5b_042f382d-836e-4cc2-9065-352a210f6593/extract-utilities/0.log" Dec 06 04:46:02 crc kubenswrapper[4802]: I1206 04:46:02.135133 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-skt5b_042f382d-836e-4cc2-9065-352a210f6593/extract-content/0.log" Dec 06 04:46:02 crc kubenswrapper[4802]: I1206 04:46:02.324187 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rnrvp_cae8cfbe-5644-45a5-92e6-0ce169cc47c5/extract-utilities/0.log" Dec 06 04:46:02 crc kubenswrapper[4802]: I1206 04:46:02.548741 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-skt5b_042f382d-836e-4cc2-9065-352a210f6593/registry-server/0.log" Dec 06 04:46:02 crc kubenswrapper[4802]: I1206 04:46:02.583725 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rnrvp_cae8cfbe-5644-45a5-92e6-0ce169cc47c5/extract-content/0.log" Dec 06 04:46:02 crc kubenswrapper[4802]: I1206 04:46:02.614036 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rnrvp_cae8cfbe-5644-45a5-92e6-0ce169cc47c5/extract-utilities/0.log" Dec 06 04:46:02 crc kubenswrapper[4802]: I1206 04:46:02.635790 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rnrvp_cae8cfbe-5644-45a5-92e6-0ce169cc47c5/extract-content/0.log" Dec 06 04:46:03 crc kubenswrapper[4802]: I1206 04:46:03.584285 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rnrvp_cae8cfbe-5644-45a5-92e6-0ce169cc47c5/extract-content/0.log" Dec 06 04:46:03 crc kubenswrapper[4802]: I1206 04:46:03.825146 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rnrvp_cae8cfbe-5644-45a5-92e6-0ce169cc47c5/extract-utilities/0.log" Dec 06 04:46:04 crc kubenswrapper[4802]: I1206 04:46:04.075883 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2l4qq_199a835e-3d62-42ef-8df0-6f087ae5586e/extract-utilities/0.log" Dec 06 04:46:04 crc kubenswrapper[4802]: I1206 04:46:04.086081 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-wxqsx_cffba732-050e-470a-9876-5954b2854ba5/marketplace-operator/0.log" Dec 06 04:46:04 crc kubenswrapper[4802]: I1206 04:46:04.265994 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2l4qq_199a835e-3d62-42ef-8df0-6f087ae5586e/extract-content/0.log" Dec 06 04:46:04 crc kubenswrapper[4802]: I1206 04:46:04.307025 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2l4qq_199a835e-3d62-42ef-8df0-6f087ae5586e/extract-content/0.log" Dec 06 04:46:04 crc kubenswrapper[4802]: I1206 04:46:04.402701 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2l4qq_199a835e-3d62-42ef-8df0-6f087ae5586e/extract-utilities/0.log" Dec 06 04:46:04 crc kubenswrapper[4802]: I1206 04:46:04.457713 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rnrvp_cae8cfbe-5644-45a5-92e6-0ce169cc47c5/registry-server/0.log" Dec 06 04:46:04 crc kubenswrapper[4802]: I1206 04:46:04.568996 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2l4qq_199a835e-3d62-42ef-8df0-6f087ae5586e/extract-utilities/0.log" Dec 06 04:46:04 crc kubenswrapper[4802]: I1206 04:46:04.618119 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2l4qq_199a835e-3d62-42ef-8df0-6f087ae5586e/extract-content/0.log" Dec 06 04:46:04 crc kubenswrapper[4802]: I1206 04:46:04.684447 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-ldbdv_a640d4db-76d9-4c21-b394-cbc18a3fe2c4/extract-utilities/0.log" Dec 06 04:46:04 crc kubenswrapper[4802]: I1206 04:46:04.711199 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2l4qq_199a835e-3d62-42ef-8df0-6f087ae5586e/registry-server/0.log" Dec 06 04:46:04 crc kubenswrapper[4802]: I1206 04:46:04.875551 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-ldbdv_a640d4db-76d9-4c21-b394-cbc18a3fe2c4/extract-content/0.log" Dec 06 04:46:04 crc kubenswrapper[4802]: I1206 04:46:04.931908 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-ldbdv_a640d4db-76d9-4c21-b394-cbc18a3fe2c4/extract-utilities/0.log" Dec 06 04:46:04 crc kubenswrapper[4802]: I1206 04:46:04.938569 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-ldbdv_a640d4db-76d9-4c21-b394-cbc18a3fe2c4/extract-content/0.log" Dec 06 04:46:05 crc kubenswrapper[4802]: I1206 04:46:05.136743 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-ldbdv_a640d4db-76d9-4c21-b394-cbc18a3fe2c4/extract-utilities/0.log" Dec 06 04:46:05 crc kubenswrapper[4802]: I1206 04:46:05.184935 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-ldbdv_a640d4db-76d9-4c21-b394-cbc18a3fe2c4/extract-content/0.log" Dec 06 04:46:05 crc kubenswrapper[4802]: I1206 04:46:05.583451 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-ldbdv_a640d4db-76d9-4c21-b394-cbc18a3fe2c4/registry-server/0.log" Dec 06 04:46:23 crc kubenswrapper[4802]: I1206 04:46:23.651129 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8gkpv"] Dec 06 04:46:23 crc kubenswrapper[4802]: E1206 04:46:23.652057 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="605adff1-0732-48d8-bd0a-a499309a6da5" containerName="collect-profiles" Dec 06 04:46:23 crc kubenswrapper[4802]: I1206 04:46:23.652077 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="605adff1-0732-48d8-bd0a-a499309a6da5" containerName="collect-profiles" Dec 06 04:46:23 crc kubenswrapper[4802]: I1206 04:46:23.652276 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="605adff1-0732-48d8-bd0a-a499309a6da5" containerName="collect-profiles" Dec 06 04:46:23 crc kubenswrapper[4802]: I1206 04:46:23.655893 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8gkpv" Dec 06 04:46:23 crc kubenswrapper[4802]: I1206 04:46:23.673353 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8gkpv"] Dec 06 04:46:23 crc kubenswrapper[4802]: I1206 04:46:23.757798 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b46e8527-a03c-4d9f-aff7-85c720167a79-utilities\") pod \"redhat-marketplace-8gkpv\" (UID: \"b46e8527-a03c-4d9f-aff7-85c720167a79\") " pod="openshift-marketplace/redhat-marketplace-8gkpv" Dec 06 04:46:23 crc kubenswrapper[4802]: I1206 04:46:23.757858 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b46e8527-a03c-4d9f-aff7-85c720167a79-catalog-content\") pod \"redhat-marketplace-8gkpv\" (UID: \"b46e8527-a03c-4d9f-aff7-85c720167a79\") " pod="openshift-marketplace/redhat-marketplace-8gkpv" Dec 06 04:46:23 crc kubenswrapper[4802]: I1206 04:46:23.757893 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgll2\" (UniqueName: \"kubernetes.io/projected/b46e8527-a03c-4d9f-aff7-85c720167a79-kube-api-access-zgll2\") pod \"redhat-marketplace-8gkpv\" (UID: \"b46e8527-a03c-4d9f-aff7-85c720167a79\") " pod="openshift-marketplace/redhat-marketplace-8gkpv" Dec 06 04:46:23 crc kubenswrapper[4802]: I1206 04:46:23.859622 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b46e8527-a03c-4d9f-aff7-85c720167a79-utilities\") pod \"redhat-marketplace-8gkpv\" (UID: \"b46e8527-a03c-4d9f-aff7-85c720167a79\") " pod="openshift-marketplace/redhat-marketplace-8gkpv" Dec 06 04:46:23 crc kubenswrapper[4802]: I1206 04:46:23.859717 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b46e8527-a03c-4d9f-aff7-85c720167a79-catalog-content\") pod \"redhat-marketplace-8gkpv\" (UID: \"b46e8527-a03c-4d9f-aff7-85c720167a79\") " pod="openshift-marketplace/redhat-marketplace-8gkpv" Dec 06 04:46:23 crc kubenswrapper[4802]: I1206 04:46:23.859839 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgll2\" (UniqueName: \"kubernetes.io/projected/b46e8527-a03c-4d9f-aff7-85c720167a79-kube-api-access-zgll2\") pod \"redhat-marketplace-8gkpv\" (UID: \"b46e8527-a03c-4d9f-aff7-85c720167a79\") " pod="openshift-marketplace/redhat-marketplace-8gkpv" Dec 06 04:46:23 crc kubenswrapper[4802]: I1206 04:46:23.860204 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b46e8527-a03c-4d9f-aff7-85c720167a79-utilities\") pod \"redhat-marketplace-8gkpv\" (UID: \"b46e8527-a03c-4d9f-aff7-85c720167a79\") " pod="openshift-marketplace/redhat-marketplace-8gkpv" Dec 06 04:46:23 crc kubenswrapper[4802]: I1206 04:46:23.860314 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b46e8527-a03c-4d9f-aff7-85c720167a79-catalog-content\") pod \"redhat-marketplace-8gkpv\" (UID: \"b46e8527-a03c-4d9f-aff7-85c720167a79\") " pod="openshift-marketplace/redhat-marketplace-8gkpv" Dec 06 04:46:23 crc kubenswrapper[4802]: I1206 04:46:23.882039 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgll2\" (UniqueName: \"kubernetes.io/projected/b46e8527-a03c-4d9f-aff7-85c720167a79-kube-api-access-zgll2\") pod \"redhat-marketplace-8gkpv\" (UID: \"b46e8527-a03c-4d9f-aff7-85c720167a79\") " pod="openshift-marketplace/redhat-marketplace-8gkpv" Dec 06 04:46:23 crc kubenswrapper[4802]: I1206 04:46:23.979153 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8gkpv" Dec 06 04:46:24 crc kubenswrapper[4802]: I1206 04:46:24.542339 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8gkpv"] Dec 06 04:46:24 crc kubenswrapper[4802]: W1206 04:46:24.548173 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb46e8527_a03c_4d9f_aff7_85c720167a79.slice/crio-10664038f5eafe3efa43fe38d1c1e73f703f94fbc29428f951f5835e44d5154e WatchSource:0}: Error finding container 10664038f5eafe3efa43fe38d1c1e73f703f94fbc29428f951f5835e44d5154e: Status 404 returned error can't find the container with id 10664038f5eafe3efa43fe38d1c1e73f703f94fbc29428f951f5835e44d5154e Dec 06 04:46:24 crc kubenswrapper[4802]: I1206 04:46:24.641064 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8gkpv" event={"ID":"b46e8527-a03c-4d9f-aff7-85c720167a79","Type":"ContainerStarted","Data":"10664038f5eafe3efa43fe38d1c1e73f703f94fbc29428f951f5835e44d5154e"} Dec 06 04:46:25 crc kubenswrapper[4802]: I1206 04:46:25.662177 4802 generic.go:334] "Generic (PLEG): container finished" podID="b46e8527-a03c-4d9f-aff7-85c720167a79" containerID="c7c023c2a5626840690d799714aaa41a0bb2bf87d466eb234f41d41f1dca378d" exitCode=0 Dec 06 04:46:25 crc kubenswrapper[4802]: I1206 04:46:25.662337 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8gkpv" event={"ID":"b46e8527-a03c-4d9f-aff7-85c720167a79","Type":"ContainerDied","Data":"c7c023c2a5626840690d799714aaa41a0bb2bf87d466eb234f41d41f1dca378d"} Dec 06 04:46:26 crc kubenswrapper[4802]: I1206 04:46:26.677090 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8gkpv" event={"ID":"b46e8527-a03c-4d9f-aff7-85c720167a79","Type":"ContainerStarted","Data":"5dbd7066a3c7377a49163defc08f815be13c2694f799b5308c4e61726ba4668a"} Dec 06 04:46:28 crc kubenswrapper[4802]: I1206 04:46:28.707915 4802 generic.go:334] "Generic (PLEG): container finished" podID="b46e8527-a03c-4d9f-aff7-85c720167a79" containerID="5dbd7066a3c7377a49163defc08f815be13c2694f799b5308c4e61726ba4668a" exitCode=0 Dec 06 04:46:28 crc kubenswrapper[4802]: I1206 04:46:28.708126 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8gkpv" event={"ID":"b46e8527-a03c-4d9f-aff7-85c720167a79","Type":"ContainerDied","Data":"5dbd7066a3c7377a49163defc08f815be13c2694f799b5308c4e61726ba4668a"} Dec 06 04:46:29 crc kubenswrapper[4802]: I1206 04:46:29.748112 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8gkpv" event={"ID":"b46e8527-a03c-4d9f-aff7-85c720167a79","Type":"ContainerStarted","Data":"d6a546ffdcb72e969e211d5e5e1b32f37f563c7a5f8bd6c182b9a554476b4e98"} Dec 06 04:46:29 crc kubenswrapper[4802]: I1206 04:46:29.801729 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8gkpv" podStartSLOduration=3.175483027 podStartE2EDuration="6.801709791s" podCreationTimestamp="2025-12-06 04:46:23 +0000 UTC" firstStartedPulling="2025-12-06 04:46:25.665069081 +0000 UTC m=+3978.536978233" lastFinishedPulling="2025-12-06 04:46:29.291295845 +0000 UTC m=+3982.163204997" observedRunningTime="2025-12-06 04:46:29.784303028 +0000 UTC m=+3982.656212180" watchObservedRunningTime="2025-12-06 04:46:29.801709791 +0000 UTC m=+3982.673618943" Dec 06 04:46:33 crc kubenswrapper[4802]: I1206 04:46:33.979282 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8gkpv" Dec 06 04:46:33 crc kubenswrapper[4802]: I1206 04:46:33.979897 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8gkpv" Dec 06 04:46:34 crc kubenswrapper[4802]: I1206 04:46:34.027332 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8gkpv" Dec 06 04:46:34 crc kubenswrapper[4802]: I1206 04:46:34.881114 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8gkpv" Dec 06 04:46:34 crc kubenswrapper[4802]: I1206 04:46:34.928880 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8gkpv"] Dec 06 04:46:36 crc kubenswrapper[4802]: I1206 04:46:36.814722 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8gkpv" podUID="b46e8527-a03c-4d9f-aff7-85c720167a79" containerName="registry-server" containerID="cri-o://d6a546ffdcb72e969e211d5e5e1b32f37f563c7a5f8bd6c182b9a554476b4e98" gracePeriod=2 Dec 06 04:46:37 crc kubenswrapper[4802]: I1206 04:46:37.416673 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8gkpv" Dec 06 04:46:37 crc kubenswrapper[4802]: I1206 04:46:37.471823 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgll2\" (UniqueName: \"kubernetes.io/projected/b46e8527-a03c-4d9f-aff7-85c720167a79-kube-api-access-zgll2\") pod \"b46e8527-a03c-4d9f-aff7-85c720167a79\" (UID: \"b46e8527-a03c-4d9f-aff7-85c720167a79\") " Dec 06 04:46:37 crc kubenswrapper[4802]: I1206 04:46:37.471904 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b46e8527-a03c-4d9f-aff7-85c720167a79-utilities\") pod \"b46e8527-a03c-4d9f-aff7-85c720167a79\" (UID: \"b46e8527-a03c-4d9f-aff7-85c720167a79\") " Dec 06 04:46:37 crc kubenswrapper[4802]: I1206 04:46:37.471921 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b46e8527-a03c-4d9f-aff7-85c720167a79-catalog-content\") pod \"b46e8527-a03c-4d9f-aff7-85c720167a79\" (UID: \"b46e8527-a03c-4d9f-aff7-85c720167a79\") " Dec 06 04:46:37 crc kubenswrapper[4802]: I1206 04:46:37.473023 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b46e8527-a03c-4d9f-aff7-85c720167a79-utilities" (OuterVolumeSpecName: "utilities") pod "b46e8527-a03c-4d9f-aff7-85c720167a79" (UID: "b46e8527-a03c-4d9f-aff7-85c720167a79"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:46:37 crc kubenswrapper[4802]: I1206 04:46:37.494128 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b46e8527-a03c-4d9f-aff7-85c720167a79-kube-api-access-zgll2" (OuterVolumeSpecName: "kube-api-access-zgll2") pod "b46e8527-a03c-4d9f-aff7-85c720167a79" (UID: "b46e8527-a03c-4d9f-aff7-85c720167a79"). InnerVolumeSpecName "kube-api-access-zgll2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:46:37 crc kubenswrapper[4802]: I1206 04:46:37.532956 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b46e8527-a03c-4d9f-aff7-85c720167a79-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b46e8527-a03c-4d9f-aff7-85c720167a79" (UID: "b46e8527-a03c-4d9f-aff7-85c720167a79"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:46:37 crc kubenswrapper[4802]: I1206 04:46:37.574249 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgll2\" (UniqueName: \"kubernetes.io/projected/b46e8527-a03c-4d9f-aff7-85c720167a79-kube-api-access-zgll2\") on node \"crc\" DevicePath \"\"" Dec 06 04:46:37 crc kubenswrapper[4802]: I1206 04:46:37.574281 4802 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b46e8527-a03c-4d9f-aff7-85c720167a79-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 04:46:37 crc kubenswrapper[4802]: I1206 04:46:37.574290 4802 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b46e8527-a03c-4d9f-aff7-85c720167a79-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 04:46:37 crc kubenswrapper[4802]: I1206 04:46:37.691439 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xctfm"] Dec 06 04:46:37 crc kubenswrapper[4802]: E1206 04:46:37.692213 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b46e8527-a03c-4d9f-aff7-85c720167a79" containerName="registry-server" Dec 06 04:46:37 crc kubenswrapper[4802]: I1206 04:46:37.692234 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="b46e8527-a03c-4d9f-aff7-85c720167a79" containerName="registry-server" Dec 06 04:46:37 crc kubenswrapper[4802]: E1206 04:46:37.692257 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b46e8527-a03c-4d9f-aff7-85c720167a79" containerName="extract-content" Dec 06 04:46:37 crc kubenswrapper[4802]: I1206 04:46:37.692265 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="b46e8527-a03c-4d9f-aff7-85c720167a79" containerName="extract-content" Dec 06 04:46:37 crc kubenswrapper[4802]: E1206 04:46:37.692283 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b46e8527-a03c-4d9f-aff7-85c720167a79" containerName="extract-utilities" Dec 06 04:46:37 crc kubenswrapper[4802]: I1206 04:46:37.692292 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="b46e8527-a03c-4d9f-aff7-85c720167a79" containerName="extract-utilities" Dec 06 04:46:37 crc kubenswrapper[4802]: I1206 04:46:37.692503 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="b46e8527-a03c-4d9f-aff7-85c720167a79" containerName="registry-server" Dec 06 04:46:37 crc kubenswrapper[4802]: I1206 04:46:37.695305 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xctfm" Dec 06 04:46:37 crc kubenswrapper[4802]: I1206 04:46:37.750807 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xctfm"] Dec 06 04:46:37 crc kubenswrapper[4802]: I1206 04:46:37.777771 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9vp9\" (UniqueName: \"kubernetes.io/projected/abf099b1-802d-4d53-b469-1a5f986e17c0-kube-api-access-v9vp9\") pod \"redhat-operators-xctfm\" (UID: \"abf099b1-802d-4d53-b469-1a5f986e17c0\") " pod="openshift-marketplace/redhat-operators-xctfm" Dec 06 04:46:37 crc kubenswrapper[4802]: I1206 04:46:37.777955 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abf099b1-802d-4d53-b469-1a5f986e17c0-catalog-content\") pod \"redhat-operators-xctfm\" (UID: \"abf099b1-802d-4d53-b469-1a5f986e17c0\") " pod="openshift-marketplace/redhat-operators-xctfm" Dec 06 04:46:37 crc kubenswrapper[4802]: I1206 04:46:37.778059 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abf099b1-802d-4d53-b469-1a5f986e17c0-utilities\") pod \"redhat-operators-xctfm\" (UID: \"abf099b1-802d-4d53-b469-1a5f986e17c0\") " pod="openshift-marketplace/redhat-operators-xctfm" Dec 06 04:46:37 crc kubenswrapper[4802]: I1206 04:46:37.859557 4802 generic.go:334] "Generic (PLEG): container finished" podID="b46e8527-a03c-4d9f-aff7-85c720167a79" containerID="d6a546ffdcb72e969e211d5e5e1b32f37f563c7a5f8bd6c182b9a554476b4e98" exitCode=0 Dec 06 04:46:37 crc kubenswrapper[4802]: I1206 04:46:37.859633 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8gkpv" event={"ID":"b46e8527-a03c-4d9f-aff7-85c720167a79","Type":"ContainerDied","Data":"d6a546ffdcb72e969e211d5e5e1b32f37f563c7a5f8bd6c182b9a554476b4e98"} Dec 06 04:46:37 crc kubenswrapper[4802]: I1206 04:46:37.859889 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8gkpv" Dec 06 04:46:37 crc kubenswrapper[4802]: I1206 04:46:37.860479 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8gkpv" event={"ID":"b46e8527-a03c-4d9f-aff7-85c720167a79","Type":"ContainerDied","Data":"10664038f5eafe3efa43fe38d1c1e73f703f94fbc29428f951f5835e44d5154e"} Dec 06 04:46:37 crc kubenswrapper[4802]: I1206 04:46:37.860526 4802 scope.go:117] "RemoveContainer" containerID="d6a546ffdcb72e969e211d5e5e1b32f37f563c7a5f8bd6c182b9a554476b4e98" Dec 06 04:46:37 crc kubenswrapper[4802]: I1206 04:46:37.880852 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abf099b1-802d-4d53-b469-1a5f986e17c0-utilities\") pod \"redhat-operators-xctfm\" (UID: \"abf099b1-802d-4d53-b469-1a5f986e17c0\") " pod="openshift-marketplace/redhat-operators-xctfm" Dec 06 04:46:37 crc kubenswrapper[4802]: I1206 04:46:37.880925 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9vp9\" (UniqueName: \"kubernetes.io/projected/abf099b1-802d-4d53-b469-1a5f986e17c0-kube-api-access-v9vp9\") pod \"redhat-operators-xctfm\" (UID: \"abf099b1-802d-4d53-b469-1a5f986e17c0\") " pod="openshift-marketplace/redhat-operators-xctfm" Dec 06 04:46:37 crc kubenswrapper[4802]: I1206 04:46:37.881054 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abf099b1-802d-4d53-b469-1a5f986e17c0-catalog-content\") pod \"redhat-operators-xctfm\" (UID: \"abf099b1-802d-4d53-b469-1a5f986e17c0\") " pod="openshift-marketplace/redhat-operators-xctfm" Dec 06 04:46:37 crc kubenswrapper[4802]: I1206 04:46:37.881477 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abf099b1-802d-4d53-b469-1a5f986e17c0-utilities\") pod \"redhat-operators-xctfm\" (UID: \"abf099b1-802d-4d53-b469-1a5f986e17c0\") " pod="openshift-marketplace/redhat-operators-xctfm" Dec 06 04:46:37 crc kubenswrapper[4802]: I1206 04:46:37.882219 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abf099b1-802d-4d53-b469-1a5f986e17c0-catalog-content\") pod \"redhat-operators-xctfm\" (UID: \"abf099b1-802d-4d53-b469-1a5f986e17c0\") " pod="openshift-marketplace/redhat-operators-xctfm" Dec 06 04:46:37 crc kubenswrapper[4802]: I1206 04:46:37.899571 4802 scope.go:117] "RemoveContainer" containerID="5dbd7066a3c7377a49163defc08f815be13c2694f799b5308c4e61726ba4668a" Dec 06 04:46:37 crc kubenswrapper[4802]: I1206 04:46:37.911809 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9vp9\" (UniqueName: \"kubernetes.io/projected/abf099b1-802d-4d53-b469-1a5f986e17c0-kube-api-access-v9vp9\") pod \"redhat-operators-xctfm\" (UID: \"abf099b1-802d-4d53-b469-1a5f986e17c0\") " pod="openshift-marketplace/redhat-operators-xctfm" Dec 06 04:46:37 crc kubenswrapper[4802]: I1206 04:46:37.983537 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8gkpv"] Dec 06 04:46:37 crc kubenswrapper[4802]: I1206 04:46:37.991577 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8gkpv"] Dec 06 04:46:38 crc kubenswrapper[4802]: I1206 04:46:38.003342 4802 scope.go:117] "RemoveContainer" containerID="c7c023c2a5626840690d799714aaa41a0bb2bf87d466eb234f41d41f1dca378d" Dec 06 04:46:38 crc kubenswrapper[4802]: I1206 04:46:38.024799 4802 scope.go:117] "RemoveContainer" containerID="d6a546ffdcb72e969e211d5e5e1b32f37f563c7a5f8bd6c182b9a554476b4e98" Dec 06 04:46:38 crc kubenswrapper[4802]: E1206 04:46:38.025504 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6a546ffdcb72e969e211d5e5e1b32f37f563c7a5f8bd6c182b9a554476b4e98\": container with ID starting with d6a546ffdcb72e969e211d5e5e1b32f37f563c7a5f8bd6c182b9a554476b4e98 not found: ID does not exist" containerID="d6a546ffdcb72e969e211d5e5e1b32f37f563c7a5f8bd6c182b9a554476b4e98" Dec 06 04:46:38 crc kubenswrapper[4802]: I1206 04:46:38.025532 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6a546ffdcb72e969e211d5e5e1b32f37f563c7a5f8bd6c182b9a554476b4e98"} err="failed to get container status \"d6a546ffdcb72e969e211d5e5e1b32f37f563c7a5f8bd6c182b9a554476b4e98\": rpc error: code = NotFound desc = could not find container \"d6a546ffdcb72e969e211d5e5e1b32f37f563c7a5f8bd6c182b9a554476b4e98\": container with ID starting with d6a546ffdcb72e969e211d5e5e1b32f37f563c7a5f8bd6c182b9a554476b4e98 not found: ID does not exist" Dec 06 04:46:38 crc kubenswrapper[4802]: I1206 04:46:38.025553 4802 scope.go:117] "RemoveContainer" containerID="5dbd7066a3c7377a49163defc08f815be13c2694f799b5308c4e61726ba4668a" Dec 06 04:46:38 crc kubenswrapper[4802]: E1206 04:46:38.025951 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5dbd7066a3c7377a49163defc08f815be13c2694f799b5308c4e61726ba4668a\": container with ID starting with 5dbd7066a3c7377a49163defc08f815be13c2694f799b5308c4e61726ba4668a not found: ID does not exist" containerID="5dbd7066a3c7377a49163defc08f815be13c2694f799b5308c4e61726ba4668a" Dec 06 04:46:38 crc kubenswrapper[4802]: I1206 04:46:38.025973 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5dbd7066a3c7377a49163defc08f815be13c2694f799b5308c4e61726ba4668a"} err="failed to get container status \"5dbd7066a3c7377a49163defc08f815be13c2694f799b5308c4e61726ba4668a\": rpc error: code = NotFound desc = could not find container \"5dbd7066a3c7377a49163defc08f815be13c2694f799b5308c4e61726ba4668a\": container with ID starting with 5dbd7066a3c7377a49163defc08f815be13c2694f799b5308c4e61726ba4668a not found: ID does not exist" Dec 06 04:46:38 crc kubenswrapper[4802]: I1206 04:46:38.025986 4802 scope.go:117] "RemoveContainer" containerID="c7c023c2a5626840690d799714aaa41a0bb2bf87d466eb234f41d41f1dca378d" Dec 06 04:46:38 crc kubenswrapper[4802]: E1206 04:46:38.026334 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7c023c2a5626840690d799714aaa41a0bb2bf87d466eb234f41d41f1dca378d\": container with ID starting with c7c023c2a5626840690d799714aaa41a0bb2bf87d466eb234f41d41f1dca378d not found: ID does not exist" containerID="c7c023c2a5626840690d799714aaa41a0bb2bf87d466eb234f41d41f1dca378d" Dec 06 04:46:38 crc kubenswrapper[4802]: I1206 04:46:38.026352 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7c023c2a5626840690d799714aaa41a0bb2bf87d466eb234f41d41f1dca378d"} err="failed to get container status \"c7c023c2a5626840690d799714aaa41a0bb2bf87d466eb234f41d41f1dca378d\": rpc error: code = NotFound desc = could not find container \"c7c023c2a5626840690d799714aaa41a0bb2bf87d466eb234f41d41f1dca378d\": container with ID starting with c7c023c2a5626840690d799714aaa41a0bb2bf87d466eb234f41d41f1dca378d not found: ID does not exist" Dec 06 04:46:38 crc kubenswrapper[4802]: I1206 04:46:38.049409 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xctfm" Dec 06 04:46:38 crc kubenswrapper[4802]: I1206 04:46:38.370194 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xctfm"] Dec 06 04:46:38 crc kubenswrapper[4802]: I1206 04:46:38.870810 4802 generic.go:334] "Generic (PLEG): container finished" podID="abf099b1-802d-4d53-b469-1a5f986e17c0" containerID="faf68939e8922c3409e139e9f981860462e6b3d725f848432cba6f526f2b1143" exitCode=0 Dec 06 04:46:38 crc kubenswrapper[4802]: I1206 04:46:38.870913 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xctfm" event={"ID":"abf099b1-802d-4d53-b469-1a5f986e17c0","Type":"ContainerDied","Data":"faf68939e8922c3409e139e9f981860462e6b3d725f848432cba6f526f2b1143"} Dec 06 04:46:38 crc kubenswrapper[4802]: I1206 04:46:38.871215 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xctfm" event={"ID":"abf099b1-802d-4d53-b469-1a5f986e17c0","Type":"ContainerStarted","Data":"1440a7b442bee748f7bcb4aacba4eebbbd702223d2adbc3415fd336952ee45a4"} Dec 06 04:46:39 crc kubenswrapper[4802]: I1206 04:46:39.463984 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b46e8527-a03c-4d9f-aff7-85c720167a79" path="/var/lib/kubelet/pods/b46e8527-a03c-4d9f-aff7-85c720167a79/volumes" Dec 06 04:46:39 crc kubenswrapper[4802]: I1206 04:46:39.883083 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xctfm" event={"ID":"abf099b1-802d-4d53-b469-1a5f986e17c0","Type":"ContainerStarted","Data":"f3ea0133e68bb6412317fe5c00f1398bcbf3c5e706147e6ccacb0b492b92f197"} Dec 06 04:46:41 crc kubenswrapper[4802]: I1206 04:46:41.906543 4802 generic.go:334] "Generic (PLEG): container finished" podID="abf099b1-802d-4d53-b469-1a5f986e17c0" containerID="f3ea0133e68bb6412317fe5c00f1398bcbf3c5e706147e6ccacb0b492b92f197" exitCode=0 Dec 06 04:46:41 crc kubenswrapper[4802]: I1206 04:46:41.906828 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xctfm" event={"ID":"abf099b1-802d-4d53-b469-1a5f986e17c0","Type":"ContainerDied","Data":"f3ea0133e68bb6412317fe5c00f1398bcbf3c5e706147e6ccacb0b492b92f197"} Dec 06 04:46:42 crc kubenswrapper[4802]: I1206 04:46:42.920911 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xctfm" event={"ID":"abf099b1-802d-4d53-b469-1a5f986e17c0","Type":"ContainerStarted","Data":"c5185c58d41dd96228771e85bd6b19fe69ac48d721ea0bcf6cefea717beeb49c"} Dec 06 04:46:42 crc kubenswrapper[4802]: I1206 04:46:42.938999 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xctfm" podStartSLOduration=2.42196502 podStartE2EDuration="5.938974093s" podCreationTimestamp="2025-12-06 04:46:37 +0000 UTC" firstStartedPulling="2025-12-06 04:46:38.873140408 +0000 UTC m=+3991.745049560" lastFinishedPulling="2025-12-06 04:46:42.390149481 +0000 UTC m=+3995.262058633" observedRunningTime="2025-12-06 04:46:42.934362777 +0000 UTC m=+3995.806271929" watchObservedRunningTime="2025-12-06 04:46:42.938974093 +0000 UTC m=+3995.810883245" Dec 06 04:46:48 crc kubenswrapper[4802]: I1206 04:46:48.049677 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xctfm" Dec 06 04:46:48 crc kubenswrapper[4802]: I1206 04:46:48.050316 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xctfm" Dec 06 04:46:49 crc kubenswrapper[4802]: I1206 04:46:49.107782 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-xctfm" podUID="abf099b1-802d-4d53-b469-1a5f986e17c0" containerName="registry-server" probeResult="failure" output=< Dec 06 04:46:49 crc kubenswrapper[4802]: timeout: failed to connect service ":50051" within 1s Dec 06 04:46:49 crc kubenswrapper[4802]: > Dec 06 04:46:58 crc kubenswrapper[4802]: I1206 04:46:58.101320 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xctfm" Dec 06 04:46:58 crc kubenswrapper[4802]: I1206 04:46:58.157425 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xctfm" Dec 06 04:46:58 crc kubenswrapper[4802]: I1206 04:46:58.346453 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xctfm"] Dec 06 04:47:00 crc kubenswrapper[4802]: I1206 04:47:00.140847 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xctfm" podUID="abf099b1-802d-4d53-b469-1a5f986e17c0" containerName="registry-server" containerID="cri-o://c5185c58d41dd96228771e85bd6b19fe69ac48d721ea0bcf6cefea717beeb49c" gracePeriod=2 Dec 06 04:47:00 crc kubenswrapper[4802]: I1206 04:47:00.697204 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xctfm" Dec 06 04:47:00 crc kubenswrapper[4802]: I1206 04:47:00.866725 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abf099b1-802d-4d53-b469-1a5f986e17c0-utilities\") pod \"abf099b1-802d-4d53-b469-1a5f986e17c0\" (UID: \"abf099b1-802d-4d53-b469-1a5f986e17c0\") " Dec 06 04:47:00 crc kubenswrapper[4802]: I1206 04:47:00.866844 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v9vp9\" (UniqueName: \"kubernetes.io/projected/abf099b1-802d-4d53-b469-1a5f986e17c0-kube-api-access-v9vp9\") pod \"abf099b1-802d-4d53-b469-1a5f986e17c0\" (UID: \"abf099b1-802d-4d53-b469-1a5f986e17c0\") " Dec 06 04:47:00 crc kubenswrapper[4802]: I1206 04:47:00.866894 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abf099b1-802d-4d53-b469-1a5f986e17c0-catalog-content\") pod \"abf099b1-802d-4d53-b469-1a5f986e17c0\" (UID: \"abf099b1-802d-4d53-b469-1a5f986e17c0\") " Dec 06 04:47:00 crc kubenswrapper[4802]: I1206 04:47:00.867667 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/abf099b1-802d-4d53-b469-1a5f986e17c0-utilities" (OuterVolumeSpecName: "utilities") pod "abf099b1-802d-4d53-b469-1a5f986e17c0" (UID: "abf099b1-802d-4d53-b469-1a5f986e17c0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:47:00 crc kubenswrapper[4802]: I1206 04:47:00.873015 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abf099b1-802d-4d53-b469-1a5f986e17c0-kube-api-access-v9vp9" (OuterVolumeSpecName: "kube-api-access-v9vp9") pod "abf099b1-802d-4d53-b469-1a5f986e17c0" (UID: "abf099b1-802d-4d53-b469-1a5f986e17c0"). InnerVolumeSpecName "kube-api-access-v9vp9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:47:00 crc kubenswrapper[4802]: I1206 04:47:00.970197 4802 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abf099b1-802d-4d53-b469-1a5f986e17c0-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 04:47:00 crc kubenswrapper[4802]: I1206 04:47:00.970270 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v9vp9\" (UniqueName: \"kubernetes.io/projected/abf099b1-802d-4d53-b469-1a5f986e17c0-kube-api-access-v9vp9\") on node \"crc\" DevicePath \"\"" Dec 06 04:47:00 crc kubenswrapper[4802]: I1206 04:47:00.998617 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/abf099b1-802d-4d53-b469-1a5f986e17c0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "abf099b1-802d-4d53-b469-1a5f986e17c0" (UID: "abf099b1-802d-4d53-b469-1a5f986e17c0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:47:01 crc kubenswrapper[4802]: I1206 04:47:01.072212 4802 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abf099b1-802d-4d53-b469-1a5f986e17c0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 04:47:01 crc kubenswrapper[4802]: I1206 04:47:01.155559 4802 generic.go:334] "Generic (PLEG): container finished" podID="abf099b1-802d-4d53-b469-1a5f986e17c0" containerID="c5185c58d41dd96228771e85bd6b19fe69ac48d721ea0bcf6cefea717beeb49c" exitCode=0 Dec 06 04:47:01 crc kubenswrapper[4802]: I1206 04:47:01.155632 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xctfm" event={"ID":"abf099b1-802d-4d53-b469-1a5f986e17c0","Type":"ContainerDied","Data":"c5185c58d41dd96228771e85bd6b19fe69ac48d721ea0bcf6cefea717beeb49c"} Dec 06 04:47:01 crc kubenswrapper[4802]: I1206 04:47:01.155680 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xctfm" event={"ID":"abf099b1-802d-4d53-b469-1a5f986e17c0","Type":"ContainerDied","Data":"1440a7b442bee748f7bcb4aacba4eebbbd702223d2adbc3415fd336952ee45a4"} Dec 06 04:47:01 crc kubenswrapper[4802]: I1206 04:47:01.155710 4802 scope.go:117] "RemoveContainer" containerID="c5185c58d41dd96228771e85bd6b19fe69ac48d721ea0bcf6cefea717beeb49c" Dec 06 04:47:01 crc kubenswrapper[4802]: I1206 04:47:01.155944 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xctfm" Dec 06 04:47:01 crc kubenswrapper[4802]: I1206 04:47:01.202404 4802 scope.go:117] "RemoveContainer" containerID="f3ea0133e68bb6412317fe5c00f1398bcbf3c5e706147e6ccacb0b492b92f197" Dec 06 04:47:01 crc kubenswrapper[4802]: I1206 04:47:01.207914 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xctfm"] Dec 06 04:47:01 crc kubenswrapper[4802]: I1206 04:47:01.220989 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xctfm"] Dec 06 04:47:01 crc kubenswrapper[4802]: I1206 04:47:01.225983 4802 scope.go:117] "RemoveContainer" containerID="faf68939e8922c3409e139e9f981860462e6b3d725f848432cba6f526f2b1143" Dec 06 04:47:01 crc kubenswrapper[4802]: I1206 04:47:01.270741 4802 scope.go:117] "RemoveContainer" containerID="c5185c58d41dd96228771e85bd6b19fe69ac48d721ea0bcf6cefea717beeb49c" Dec 06 04:47:01 crc kubenswrapper[4802]: E1206 04:47:01.271250 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5185c58d41dd96228771e85bd6b19fe69ac48d721ea0bcf6cefea717beeb49c\": container with ID starting with c5185c58d41dd96228771e85bd6b19fe69ac48d721ea0bcf6cefea717beeb49c not found: ID does not exist" containerID="c5185c58d41dd96228771e85bd6b19fe69ac48d721ea0bcf6cefea717beeb49c" Dec 06 04:47:01 crc kubenswrapper[4802]: I1206 04:47:01.271292 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5185c58d41dd96228771e85bd6b19fe69ac48d721ea0bcf6cefea717beeb49c"} err="failed to get container status \"c5185c58d41dd96228771e85bd6b19fe69ac48d721ea0bcf6cefea717beeb49c\": rpc error: code = NotFound desc = could not find container \"c5185c58d41dd96228771e85bd6b19fe69ac48d721ea0bcf6cefea717beeb49c\": container with ID starting with c5185c58d41dd96228771e85bd6b19fe69ac48d721ea0bcf6cefea717beeb49c not found: ID does not exist" Dec 06 04:47:01 crc kubenswrapper[4802]: I1206 04:47:01.271323 4802 scope.go:117] "RemoveContainer" containerID="f3ea0133e68bb6412317fe5c00f1398bcbf3c5e706147e6ccacb0b492b92f197" Dec 06 04:47:01 crc kubenswrapper[4802]: E1206 04:47:01.271783 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3ea0133e68bb6412317fe5c00f1398bcbf3c5e706147e6ccacb0b492b92f197\": container with ID starting with f3ea0133e68bb6412317fe5c00f1398bcbf3c5e706147e6ccacb0b492b92f197 not found: ID does not exist" containerID="f3ea0133e68bb6412317fe5c00f1398bcbf3c5e706147e6ccacb0b492b92f197" Dec 06 04:47:01 crc kubenswrapper[4802]: I1206 04:47:01.271825 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3ea0133e68bb6412317fe5c00f1398bcbf3c5e706147e6ccacb0b492b92f197"} err="failed to get container status \"f3ea0133e68bb6412317fe5c00f1398bcbf3c5e706147e6ccacb0b492b92f197\": rpc error: code = NotFound desc = could not find container \"f3ea0133e68bb6412317fe5c00f1398bcbf3c5e706147e6ccacb0b492b92f197\": container with ID starting with f3ea0133e68bb6412317fe5c00f1398bcbf3c5e706147e6ccacb0b492b92f197 not found: ID does not exist" Dec 06 04:47:01 crc kubenswrapper[4802]: I1206 04:47:01.271853 4802 scope.go:117] "RemoveContainer" containerID="faf68939e8922c3409e139e9f981860462e6b3d725f848432cba6f526f2b1143" Dec 06 04:47:01 crc kubenswrapper[4802]: E1206 04:47:01.272283 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"faf68939e8922c3409e139e9f981860462e6b3d725f848432cba6f526f2b1143\": container with ID starting with faf68939e8922c3409e139e9f981860462e6b3d725f848432cba6f526f2b1143 not found: ID does not exist" containerID="faf68939e8922c3409e139e9f981860462e6b3d725f848432cba6f526f2b1143" Dec 06 04:47:01 crc kubenswrapper[4802]: I1206 04:47:01.272308 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"faf68939e8922c3409e139e9f981860462e6b3d725f848432cba6f526f2b1143"} err="failed to get container status \"faf68939e8922c3409e139e9f981860462e6b3d725f848432cba6f526f2b1143\": rpc error: code = NotFound desc = could not find container \"faf68939e8922c3409e139e9f981860462e6b3d725f848432cba6f526f2b1143\": container with ID starting with faf68939e8922c3409e139e9f981860462e6b3d725f848432cba6f526f2b1143 not found: ID does not exist" Dec 06 04:47:01 crc kubenswrapper[4802]: I1206 04:47:01.465799 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="abf099b1-802d-4d53-b469-1a5f986e17c0" path="/var/lib/kubelet/pods/abf099b1-802d-4d53-b469-1a5f986e17c0/volumes" Dec 06 04:47:43 crc kubenswrapper[4802]: I1206 04:47:43.283056 4802 patch_prober.go:28] interesting pod/machine-config-daemon-zpxxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:47:43 crc kubenswrapper[4802]: I1206 04:47:43.283633 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:48:02 crc kubenswrapper[4802]: E1206 04:48:02.578790 4802 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddb83105e_99af_4dde_8ed4_e69a9c915d98.slice/crio-d429ca90df1982428ddc9e9dd734b968cf1d9575e289c938b6db28f48f74b400.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddb83105e_99af_4dde_8ed4_e69a9c915d98.slice/crio-conmon-d429ca90df1982428ddc9e9dd734b968cf1d9575e289c938b6db28f48f74b400.scope\": RecentStats: unable to find data in memory cache]" Dec 06 04:48:02 crc kubenswrapper[4802]: I1206 04:48:02.818519 4802 generic.go:334] "Generic (PLEG): container finished" podID="db83105e-99af-4dde-8ed4-e69a9c915d98" containerID="d429ca90df1982428ddc9e9dd734b968cf1d9575e289c938b6db28f48f74b400" exitCode=0 Dec 06 04:48:02 crc kubenswrapper[4802]: I1206 04:48:02.818673 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mjbfz/must-gather-drcf4" event={"ID":"db83105e-99af-4dde-8ed4-e69a9c915d98","Type":"ContainerDied","Data":"d429ca90df1982428ddc9e9dd734b968cf1d9575e289c938b6db28f48f74b400"} Dec 06 04:48:02 crc kubenswrapper[4802]: I1206 04:48:02.819452 4802 scope.go:117] "RemoveContainer" containerID="d429ca90df1982428ddc9e9dd734b968cf1d9575e289c938b6db28f48f74b400" Dec 06 04:48:03 crc kubenswrapper[4802]: I1206 04:48:03.493986 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mjbfz_must-gather-drcf4_db83105e-99af-4dde-8ed4-e69a9c915d98/gather/0.log" Dec 06 04:48:11 crc kubenswrapper[4802]: I1206 04:48:11.813567 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mjbfz/must-gather-drcf4"] Dec 06 04:48:11 crc kubenswrapper[4802]: I1206 04:48:11.814367 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-mjbfz/must-gather-drcf4" podUID="db83105e-99af-4dde-8ed4-e69a9c915d98" containerName="copy" containerID="cri-o://baa9c7d7aec1f6ae925ee81bd954c72d99ce577c438ba999e98173249997734a" gracePeriod=2 Dec 06 04:48:11 crc kubenswrapper[4802]: I1206 04:48:11.824669 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mjbfz/must-gather-drcf4"] Dec 06 04:48:12 crc kubenswrapper[4802]: I1206 04:48:12.356138 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mjbfz_must-gather-drcf4_db83105e-99af-4dde-8ed4-e69a9c915d98/copy/0.log" Dec 06 04:48:12 crc kubenswrapper[4802]: I1206 04:48:12.356722 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mjbfz/must-gather-drcf4" Dec 06 04:48:12 crc kubenswrapper[4802]: I1206 04:48:12.499928 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2tppg\" (UniqueName: \"kubernetes.io/projected/db83105e-99af-4dde-8ed4-e69a9c915d98-kube-api-access-2tppg\") pod \"db83105e-99af-4dde-8ed4-e69a9c915d98\" (UID: \"db83105e-99af-4dde-8ed4-e69a9c915d98\") " Dec 06 04:48:12 crc kubenswrapper[4802]: I1206 04:48:12.500036 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/db83105e-99af-4dde-8ed4-e69a9c915d98-must-gather-output\") pod \"db83105e-99af-4dde-8ed4-e69a9c915d98\" (UID: \"db83105e-99af-4dde-8ed4-e69a9c915d98\") " Dec 06 04:48:12 crc kubenswrapper[4802]: I1206 04:48:12.506404 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db83105e-99af-4dde-8ed4-e69a9c915d98-kube-api-access-2tppg" (OuterVolumeSpecName: "kube-api-access-2tppg") pod "db83105e-99af-4dde-8ed4-e69a9c915d98" (UID: "db83105e-99af-4dde-8ed4-e69a9c915d98"). InnerVolumeSpecName "kube-api-access-2tppg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:48:12 crc kubenswrapper[4802]: I1206 04:48:12.602432 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2tppg\" (UniqueName: \"kubernetes.io/projected/db83105e-99af-4dde-8ed4-e69a9c915d98-kube-api-access-2tppg\") on node \"crc\" DevicePath \"\"" Dec 06 04:48:12 crc kubenswrapper[4802]: I1206 04:48:12.662813 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db83105e-99af-4dde-8ed4-e69a9c915d98-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "db83105e-99af-4dde-8ed4-e69a9c915d98" (UID: "db83105e-99af-4dde-8ed4-e69a9c915d98"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:48:12 crc kubenswrapper[4802]: I1206 04:48:12.704144 4802 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/db83105e-99af-4dde-8ed4-e69a9c915d98-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 06 04:48:12 crc kubenswrapper[4802]: I1206 04:48:12.913742 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mjbfz_must-gather-drcf4_db83105e-99af-4dde-8ed4-e69a9c915d98/copy/0.log" Dec 06 04:48:12 crc kubenswrapper[4802]: I1206 04:48:12.915209 4802 generic.go:334] "Generic (PLEG): container finished" podID="db83105e-99af-4dde-8ed4-e69a9c915d98" containerID="baa9c7d7aec1f6ae925ee81bd954c72d99ce577c438ba999e98173249997734a" exitCode=143 Dec 06 04:48:12 crc kubenswrapper[4802]: I1206 04:48:12.915267 4802 scope.go:117] "RemoveContainer" containerID="baa9c7d7aec1f6ae925ee81bd954c72d99ce577c438ba999e98173249997734a" Dec 06 04:48:12 crc kubenswrapper[4802]: I1206 04:48:12.915377 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mjbfz/must-gather-drcf4" Dec 06 04:48:12 crc kubenswrapper[4802]: I1206 04:48:12.939377 4802 scope.go:117] "RemoveContainer" containerID="d429ca90df1982428ddc9e9dd734b968cf1d9575e289c938b6db28f48f74b400" Dec 06 04:48:12 crc kubenswrapper[4802]: I1206 04:48:12.991787 4802 scope.go:117] "RemoveContainer" containerID="baa9c7d7aec1f6ae925ee81bd954c72d99ce577c438ba999e98173249997734a" Dec 06 04:48:12 crc kubenswrapper[4802]: E1206 04:48:12.992285 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"baa9c7d7aec1f6ae925ee81bd954c72d99ce577c438ba999e98173249997734a\": container with ID starting with baa9c7d7aec1f6ae925ee81bd954c72d99ce577c438ba999e98173249997734a not found: ID does not exist" containerID="baa9c7d7aec1f6ae925ee81bd954c72d99ce577c438ba999e98173249997734a" Dec 06 04:48:12 crc kubenswrapper[4802]: I1206 04:48:12.992332 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"baa9c7d7aec1f6ae925ee81bd954c72d99ce577c438ba999e98173249997734a"} err="failed to get container status \"baa9c7d7aec1f6ae925ee81bd954c72d99ce577c438ba999e98173249997734a\": rpc error: code = NotFound desc = could not find container \"baa9c7d7aec1f6ae925ee81bd954c72d99ce577c438ba999e98173249997734a\": container with ID starting with baa9c7d7aec1f6ae925ee81bd954c72d99ce577c438ba999e98173249997734a not found: ID does not exist" Dec 06 04:48:12 crc kubenswrapper[4802]: I1206 04:48:12.992371 4802 scope.go:117] "RemoveContainer" containerID="d429ca90df1982428ddc9e9dd734b968cf1d9575e289c938b6db28f48f74b400" Dec 06 04:48:12 crc kubenswrapper[4802]: E1206 04:48:12.992829 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d429ca90df1982428ddc9e9dd734b968cf1d9575e289c938b6db28f48f74b400\": container with ID starting with d429ca90df1982428ddc9e9dd734b968cf1d9575e289c938b6db28f48f74b400 not found: ID does not exist" containerID="d429ca90df1982428ddc9e9dd734b968cf1d9575e289c938b6db28f48f74b400" Dec 06 04:48:12 crc kubenswrapper[4802]: I1206 04:48:12.992997 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d429ca90df1982428ddc9e9dd734b968cf1d9575e289c938b6db28f48f74b400"} err="failed to get container status \"d429ca90df1982428ddc9e9dd734b968cf1d9575e289c938b6db28f48f74b400\": rpc error: code = NotFound desc = could not find container \"d429ca90df1982428ddc9e9dd734b968cf1d9575e289c938b6db28f48f74b400\": container with ID starting with d429ca90df1982428ddc9e9dd734b968cf1d9575e289c938b6db28f48f74b400 not found: ID does not exist" Dec 06 04:48:13 crc kubenswrapper[4802]: I1206 04:48:13.283918 4802 patch_prober.go:28] interesting pod/machine-config-daemon-zpxxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:48:13 crc kubenswrapper[4802]: I1206 04:48:13.283986 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:48:13 crc kubenswrapper[4802]: I1206 04:48:13.460504 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db83105e-99af-4dde-8ed4-e69a9c915d98" path="/var/lib/kubelet/pods/db83105e-99af-4dde-8ed4-e69a9c915d98/volumes" Dec 06 04:48:21 crc kubenswrapper[4802]: I1206 04:48:21.276326 4802 patch_prober.go:28] interesting pod/route-controller-manager-67644cbbd7-6vq75 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.69:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 04:48:21 crc kubenswrapper[4802]: I1206 04:48:21.277080 4802 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-67644cbbd7-6vq75" podUID="d291f291-567c-4b2b-9641-cce89ef3636d" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.69:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 04:48:21 crc kubenswrapper[4802]: I1206 04:48:21.276368 4802 patch_prober.go:28] interesting pod/route-controller-manager-67644cbbd7-6vq75 container/route-controller-manager namespace/openshift-route-controller-manager: Liveness probe status=failure output="Get \"https://10.217.0.69:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 04:48:21 crc kubenswrapper[4802]: I1206 04:48:21.277896 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-route-controller-manager/route-controller-manager-67644cbbd7-6vq75" podUID="d291f291-567c-4b2b-9641-cce89ef3636d" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.69:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 04:48:43 crc kubenswrapper[4802]: I1206 04:48:43.282973 4802 patch_prober.go:28] interesting pod/machine-config-daemon-zpxxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:48:43 crc kubenswrapper[4802]: I1206 04:48:43.283451 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:48:43 crc kubenswrapper[4802]: I1206 04:48:43.283495 4802 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" Dec 06 04:48:43 crc kubenswrapper[4802]: I1206 04:48:43.284173 4802 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"659eedd07500dc9cedfea13e32706d57d3d8130ed952a340282ebbc31bb990e2"} pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 04:48:43 crc kubenswrapper[4802]: I1206 04:48:43.284226 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" containerID="cri-o://659eedd07500dc9cedfea13e32706d57d3d8130ed952a340282ebbc31bb990e2" gracePeriod=600 Dec 06 04:48:44 crc kubenswrapper[4802]: I1206 04:48:44.239518 4802 generic.go:334] "Generic (PLEG): container finished" podID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerID="659eedd07500dc9cedfea13e32706d57d3d8130ed952a340282ebbc31bb990e2" exitCode=0 Dec 06 04:48:44 crc kubenswrapper[4802]: I1206 04:48:44.239594 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" event={"ID":"6df38316-e0d3-4018-8d27-3620eba3a68d","Type":"ContainerDied","Data":"659eedd07500dc9cedfea13e32706d57d3d8130ed952a340282ebbc31bb990e2"} Dec 06 04:48:44 crc kubenswrapper[4802]: I1206 04:48:44.240262 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" event={"ID":"6df38316-e0d3-4018-8d27-3620eba3a68d","Type":"ContainerStarted","Data":"4334df6618bec8f6cedfb779435b2c4d4e0f7e149dd426e245ab29892f475208"} Dec 06 04:48:44 crc kubenswrapper[4802]: I1206 04:48:44.240326 4802 scope.go:117] "RemoveContainer" containerID="3037e38e5a7009f353fa7ac5ea266cd2366825eeb99656721b87c6c25f49afdb" Dec 06 04:49:23 crc kubenswrapper[4802]: I1206 04:49:23.303727 4802 scope.go:117] "RemoveContainer" containerID="00e0274365781e48d8eb08dbc354376abef272f2869caefe62196347884e7132" Dec 06 04:50:23 crc kubenswrapper[4802]: I1206 04:50:23.358395 4802 scope.go:117] "RemoveContainer" containerID="f626bb7c325f4af3e6d07df8598f90df2a9a83ff8777bcae9446f085f5ebc5a0" Dec 06 04:51:13 crc kubenswrapper[4802]: I1206 04:51:13.283204 4802 patch_prober.go:28] interesting pod/machine-config-daemon-zpxxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:51:13 crc kubenswrapper[4802]: I1206 04:51:13.283717 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:51:19 crc kubenswrapper[4802]: I1206 04:51:19.271809 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-jw8qc/must-gather-k7q7z"] Dec 06 04:51:19 crc kubenswrapper[4802]: E1206 04:51:19.272838 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abf099b1-802d-4d53-b469-1a5f986e17c0" containerName="registry-server" Dec 06 04:51:19 crc kubenswrapper[4802]: I1206 04:51:19.272855 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="abf099b1-802d-4d53-b469-1a5f986e17c0" containerName="registry-server" Dec 06 04:51:19 crc kubenswrapper[4802]: E1206 04:51:19.272881 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db83105e-99af-4dde-8ed4-e69a9c915d98" containerName="gather" Dec 06 04:51:19 crc kubenswrapper[4802]: I1206 04:51:19.272888 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="db83105e-99af-4dde-8ed4-e69a9c915d98" containerName="gather" Dec 06 04:51:19 crc kubenswrapper[4802]: E1206 04:51:19.272916 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abf099b1-802d-4d53-b469-1a5f986e17c0" containerName="extract-utilities" Dec 06 04:51:19 crc kubenswrapper[4802]: I1206 04:51:19.272923 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="abf099b1-802d-4d53-b469-1a5f986e17c0" containerName="extract-utilities" Dec 06 04:51:19 crc kubenswrapper[4802]: E1206 04:51:19.272931 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db83105e-99af-4dde-8ed4-e69a9c915d98" containerName="copy" Dec 06 04:51:19 crc kubenswrapper[4802]: I1206 04:51:19.272938 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="db83105e-99af-4dde-8ed4-e69a9c915d98" containerName="copy" Dec 06 04:51:19 crc kubenswrapper[4802]: E1206 04:51:19.272956 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abf099b1-802d-4d53-b469-1a5f986e17c0" containerName="extract-content" Dec 06 04:51:19 crc kubenswrapper[4802]: I1206 04:51:19.272962 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="abf099b1-802d-4d53-b469-1a5f986e17c0" containerName="extract-content" Dec 06 04:51:19 crc kubenswrapper[4802]: I1206 04:51:19.273157 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="db83105e-99af-4dde-8ed4-e69a9c915d98" containerName="copy" Dec 06 04:51:19 crc kubenswrapper[4802]: I1206 04:51:19.273176 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="abf099b1-802d-4d53-b469-1a5f986e17c0" containerName="registry-server" Dec 06 04:51:19 crc kubenswrapper[4802]: I1206 04:51:19.273190 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="db83105e-99af-4dde-8ed4-e69a9c915d98" containerName="gather" Dec 06 04:51:19 crc kubenswrapper[4802]: I1206 04:51:19.274465 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jw8qc/must-gather-k7q7z" Dec 06 04:51:19 crc kubenswrapper[4802]: I1206 04:51:19.281146 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-jw8qc"/"kube-root-ca.crt" Dec 06 04:51:19 crc kubenswrapper[4802]: I1206 04:51:19.281374 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-jw8qc"/"default-dockercfg-kbq59" Dec 06 04:51:19 crc kubenswrapper[4802]: I1206 04:51:19.281556 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-jw8qc"/"openshift-service-ca.crt" Dec 06 04:51:19 crc kubenswrapper[4802]: I1206 04:51:19.318036 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-jw8qc/must-gather-k7q7z"] Dec 06 04:51:19 crc kubenswrapper[4802]: I1206 04:51:19.398008 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lp8vm\" (UniqueName: \"kubernetes.io/projected/034bdceb-c945-4869-b588-07cc2af48473-kube-api-access-lp8vm\") pod \"must-gather-k7q7z\" (UID: \"034bdceb-c945-4869-b588-07cc2af48473\") " pod="openshift-must-gather-jw8qc/must-gather-k7q7z" Dec 06 04:51:19 crc kubenswrapper[4802]: I1206 04:51:19.398179 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/034bdceb-c945-4869-b588-07cc2af48473-must-gather-output\") pod \"must-gather-k7q7z\" (UID: \"034bdceb-c945-4869-b588-07cc2af48473\") " pod="openshift-must-gather-jw8qc/must-gather-k7q7z" Dec 06 04:51:19 crc kubenswrapper[4802]: I1206 04:51:19.501002 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/034bdceb-c945-4869-b588-07cc2af48473-must-gather-output\") pod \"must-gather-k7q7z\" (UID: \"034bdceb-c945-4869-b588-07cc2af48473\") " pod="openshift-must-gather-jw8qc/must-gather-k7q7z" Dec 06 04:51:19 crc kubenswrapper[4802]: I1206 04:51:19.502978 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/034bdceb-c945-4869-b588-07cc2af48473-must-gather-output\") pod \"must-gather-k7q7z\" (UID: \"034bdceb-c945-4869-b588-07cc2af48473\") " pod="openshift-must-gather-jw8qc/must-gather-k7q7z" Dec 06 04:51:19 crc kubenswrapper[4802]: I1206 04:51:19.503605 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lp8vm\" (UniqueName: \"kubernetes.io/projected/034bdceb-c945-4869-b588-07cc2af48473-kube-api-access-lp8vm\") pod \"must-gather-k7q7z\" (UID: \"034bdceb-c945-4869-b588-07cc2af48473\") " pod="openshift-must-gather-jw8qc/must-gather-k7q7z" Dec 06 04:51:19 crc kubenswrapper[4802]: I1206 04:51:19.528335 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lp8vm\" (UniqueName: \"kubernetes.io/projected/034bdceb-c945-4869-b588-07cc2af48473-kube-api-access-lp8vm\") pod \"must-gather-k7q7z\" (UID: \"034bdceb-c945-4869-b588-07cc2af48473\") " pod="openshift-must-gather-jw8qc/must-gather-k7q7z" Dec 06 04:51:19 crc kubenswrapper[4802]: I1206 04:51:19.609665 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jw8qc/must-gather-k7q7z" Dec 06 04:51:20 crc kubenswrapper[4802]: I1206 04:51:20.087306 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-jw8qc/must-gather-k7q7z"] Dec 06 04:51:20 crc kubenswrapper[4802]: I1206 04:51:20.798962 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jw8qc/must-gather-k7q7z" event={"ID":"034bdceb-c945-4869-b588-07cc2af48473","Type":"ContainerStarted","Data":"2267a36845a696049e85e413be808c02a8110427e541346563bca457919c365c"} Dec 06 04:51:20 crc kubenswrapper[4802]: I1206 04:51:20.799285 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jw8qc/must-gather-k7q7z" event={"ID":"034bdceb-c945-4869-b588-07cc2af48473","Type":"ContainerStarted","Data":"e8c6bbf7dd78397044bc7b821db5f0e43d3aff5aa973687df30eb73efea3d7a8"} Dec 06 04:51:20 crc kubenswrapper[4802]: I1206 04:51:20.799298 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jw8qc/must-gather-k7q7z" event={"ID":"034bdceb-c945-4869-b588-07cc2af48473","Type":"ContainerStarted","Data":"57e5a37e828f34869013fb36a17d39cb88dc9cfeed2cee9ca57dda2c66205c28"} Dec 06 04:51:20 crc kubenswrapper[4802]: I1206 04:51:20.821705 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-jw8qc/must-gather-k7q7z" podStartSLOduration=1.821685611 podStartE2EDuration="1.821685611s" podCreationTimestamp="2025-12-06 04:51:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:51:20.817274411 +0000 UTC m=+4273.689183623" watchObservedRunningTime="2025-12-06 04:51:20.821685611 +0000 UTC m=+4273.693594763" Dec 06 04:51:23 crc kubenswrapper[4802]: I1206 04:51:23.975238 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-jw8qc/crc-debug-ghv6d"] Dec 06 04:51:23 crc kubenswrapper[4802]: I1206 04:51:23.977311 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jw8qc/crc-debug-ghv6d" Dec 06 04:51:23 crc kubenswrapper[4802]: I1206 04:51:23.999793 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c48ed216-435a-435c-bf7e-8f65aa2e3107-host\") pod \"crc-debug-ghv6d\" (UID: \"c48ed216-435a-435c-bf7e-8f65aa2e3107\") " pod="openshift-must-gather-jw8qc/crc-debug-ghv6d" Dec 06 04:51:23 crc kubenswrapper[4802]: I1206 04:51:23.999867 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2ts7\" (UniqueName: \"kubernetes.io/projected/c48ed216-435a-435c-bf7e-8f65aa2e3107-kube-api-access-p2ts7\") pod \"crc-debug-ghv6d\" (UID: \"c48ed216-435a-435c-bf7e-8f65aa2e3107\") " pod="openshift-must-gather-jw8qc/crc-debug-ghv6d" Dec 06 04:51:24 crc kubenswrapper[4802]: I1206 04:51:24.101520 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c48ed216-435a-435c-bf7e-8f65aa2e3107-host\") pod \"crc-debug-ghv6d\" (UID: \"c48ed216-435a-435c-bf7e-8f65aa2e3107\") " pod="openshift-must-gather-jw8qc/crc-debug-ghv6d" Dec 06 04:51:24 crc kubenswrapper[4802]: I1206 04:51:24.101593 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2ts7\" (UniqueName: \"kubernetes.io/projected/c48ed216-435a-435c-bf7e-8f65aa2e3107-kube-api-access-p2ts7\") pod \"crc-debug-ghv6d\" (UID: \"c48ed216-435a-435c-bf7e-8f65aa2e3107\") " pod="openshift-must-gather-jw8qc/crc-debug-ghv6d" Dec 06 04:51:24 crc kubenswrapper[4802]: I1206 04:51:24.101637 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c48ed216-435a-435c-bf7e-8f65aa2e3107-host\") pod \"crc-debug-ghv6d\" (UID: \"c48ed216-435a-435c-bf7e-8f65aa2e3107\") " pod="openshift-must-gather-jw8qc/crc-debug-ghv6d" Dec 06 04:51:24 crc kubenswrapper[4802]: I1206 04:51:24.128796 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2ts7\" (UniqueName: \"kubernetes.io/projected/c48ed216-435a-435c-bf7e-8f65aa2e3107-kube-api-access-p2ts7\") pod \"crc-debug-ghv6d\" (UID: \"c48ed216-435a-435c-bf7e-8f65aa2e3107\") " pod="openshift-must-gather-jw8qc/crc-debug-ghv6d" Dec 06 04:51:24 crc kubenswrapper[4802]: I1206 04:51:24.303178 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jw8qc/crc-debug-ghv6d" Dec 06 04:51:24 crc kubenswrapper[4802]: W1206 04:51:24.345457 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc48ed216_435a_435c_bf7e_8f65aa2e3107.slice/crio-d6aabc59c423b61282cd2664cbd341728121095a045b24c657bf7465d14d734c WatchSource:0}: Error finding container d6aabc59c423b61282cd2664cbd341728121095a045b24c657bf7465d14d734c: Status 404 returned error can't find the container with id d6aabc59c423b61282cd2664cbd341728121095a045b24c657bf7465d14d734c Dec 06 04:51:24 crc kubenswrapper[4802]: I1206 04:51:24.847244 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jw8qc/crc-debug-ghv6d" event={"ID":"c48ed216-435a-435c-bf7e-8f65aa2e3107","Type":"ContainerStarted","Data":"4546dbf93185c740dcb6d447c4af2e431b1691283328f0b87520f2aa7b441401"} Dec 06 04:51:24 crc kubenswrapper[4802]: I1206 04:51:24.847901 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jw8qc/crc-debug-ghv6d" event={"ID":"c48ed216-435a-435c-bf7e-8f65aa2e3107","Type":"ContainerStarted","Data":"d6aabc59c423b61282cd2664cbd341728121095a045b24c657bf7465d14d734c"} Dec 06 04:51:24 crc kubenswrapper[4802]: I1206 04:51:24.878831 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-jw8qc/crc-debug-ghv6d" podStartSLOduration=1.878807269 podStartE2EDuration="1.878807269s" podCreationTimestamp="2025-12-06 04:51:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:51:24.873542765 +0000 UTC m=+4277.745451917" watchObservedRunningTime="2025-12-06 04:51:24.878807269 +0000 UTC m=+4277.750716421" Dec 06 04:51:43 crc kubenswrapper[4802]: I1206 04:51:43.283616 4802 patch_prober.go:28] interesting pod/machine-config-daemon-zpxxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:51:43 crc kubenswrapper[4802]: I1206 04:51:43.284138 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:52:03 crc kubenswrapper[4802]: I1206 04:52:03.210303 4802 generic.go:334] "Generic (PLEG): container finished" podID="c48ed216-435a-435c-bf7e-8f65aa2e3107" containerID="4546dbf93185c740dcb6d447c4af2e431b1691283328f0b87520f2aa7b441401" exitCode=0 Dec 06 04:52:03 crc kubenswrapper[4802]: I1206 04:52:03.210432 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jw8qc/crc-debug-ghv6d" event={"ID":"c48ed216-435a-435c-bf7e-8f65aa2e3107","Type":"ContainerDied","Data":"4546dbf93185c740dcb6d447c4af2e431b1691283328f0b87520f2aa7b441401"} Dec 06 04:52:04 crc kubenswrapper[4802]: I1206 04:52:04.346543 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jw8qc/crc-debug-ghv6d" Dec 06 04:52:04 crc kubenswrapper[4802]: I1206 04:52:04.374307 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-jw8qc/crc-debug-ghv6d"] Dec 06 04:52:04 crc kubenswrapper[4802]: I1206 04:52:04.382208 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-jw8qc/crc-debug-ghv6d"] Dec 06 04:52:04 crc kubenswrapper[4802]: I1206 04:52:04.432298 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c48ed216-435a-435c-bf7e-8f65aa2e3107-host\") pod \"c48ed216-435a-435c-bf7e-8f65aa2e3107\" (UID: \"c48ed216-435a-435c-bf7e-8f65aa2e3107\") " Dec 06 04:52:04 crc kubenswrapper[4802]: I1206 04:52:04.432490 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p2ts7\" (UniqueName: \"kubernetes.io/projected/c48ed216-435a-435c-bf7e-8f65aa2e3107-kube-api-access-p2ts7\") pod \"c48ed216-435a-435c-bf7e-8f65aa2e3107\" (UID: \"c48ed216-435a-435c-bf7e-8f65aa2e3107\") " Dec 06 04:52:04 crc kubenswrapper[4802]: I1206 04:52:04.433606 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c48ed216-435a-435c-bf7e-8f65aa2e3107-host" (OuterVolumeSpecName: "host") pod "c48ed216-435a-435c-bf7e-8f65aa2e3107" (UID: "c48ed216-435a-435c-bf7e-8f65aa2e3107"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 04:52:04 crc kubenswrapper[4802]: I1206 04:52:04.438060 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c48ed216-435a-435c-bf7e-8f65aa2e3107-kube-api-access-p2ts7" (OuterVolumeSpecName: "kube-api-access-p2ts7") pod "c48ed216-435a-435c-bf7e-8f65aa2e3107" (UID: "c48ed216-435a-435c-bf7e-8f65aa2e3107"). InnerVolumeSpecName "kube-api-access-p2ts7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:52:04 crc kubenswrapper[4802]: I1206 04:52:04.534893 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p2ts7\" (UniqueName: \"kubernetes.io/projected/c48ed216-435a-435c-bf7e-8f65aa2e3107-kube-api-access-p2ts7\") on node \"crc\" DevicePath \"\"" Dec 06 04:52:04 crc kubenswrapper[4802]: I1206 04:52:04.534954 4802 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c48ed216-435a-435c-bf7e-8f65aa2e3107-host\") on node \"crc\" DevicePath \"\"" Dec 06 04:52:05 crc kubenswrapper[4802]: I1206 04:52:05.246339 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d6aabc59c423b61282cd2664cbd341728121095a045b24c657bf7465d14d734c" Dec 06 04:52:05 crc kubenswrapper[4802]: I1206 04:52:05.246411 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jw8qc/crc-debug-ghv6d" Dec 06 04:52:05 crc kubenswrapper[4802]: I1206 04:52:05.460182 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c48ed216-435a-435c-bf7e-8f65aa2e3107" path="/var/lib/kubelet/pods/c48ed216-435a-435c-bf7e-8f65aa2e3107/volumes" Dec 06 04:52:05 crc kubenswrapper[4802]: I1206 04:52:05.569692 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-jw8qc/crc-debug-7hwqt"] Dec 06 04:52:05 crc kubenswrapper[4802]: E1206 04:52:05.570146 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c48ed216-435a-435c-bf7e-8f65aa2e3107" containerName="container-00" Dec 06 04:52:05 crc kubenswrapper[4802]: I1206 04:52:05.570165 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="c48ed216-435a-435c-bf7e-8f65aa2e3107" containerName="container-00" Dec 06 04:52:05 crc kubenswrapper[4802]: I1206 04:52:05.570362 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="c48ed216-435a-435c-bf7e-8f65aa2e3107" containerName="container-00" Dec 06 04:52:05 crc kubenswrapper[4802]: I1206 04:52:05.571024 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jw8qc/crc-debug-7hwqt" Dec 06 04:52:05 crc kubenswrapper[4802]: I1206 04:52:05.654354 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0d9cbd29-e544-49ec-96ef-890a3ce1ea19-host\") pod \"crc-debug-7hwqt\" (UID: \"0d9cbd29-e544-49ec-96ef-890a3ce1ea19\") " pod="openshift-must-gather-jw8qc/crc-debug-7hwqt" Dec 06 04:52:05 crc kubenswrapper[4802]: I1206 04:52:05.654684 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bd8w2\" (UniqueName: \"kubernetes.io/projected/0d9cbd29-e544-49ec-96ef-890a3ce1ea19-kube-api-access-bd8w2\") pod \"crc-debug-7hwqt\" (UID: \"0d9cbd29-e544-49ec-96ef-890a3ce1ea19\") " pod="openshift-must-gather-jw8qc/crc-debug-7hwqt" Dec 06 04:52:05 crc kubenswrapper[4802]: I1206 04:52:05.756299 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0d9cbd29-e544-49ec-96ef-890a3ce1ea19-host\") pod \"crc-debug-7hwqt\" (UID: \"0d9cbd29-e544-49ec-96ef-890a3ce1ea19\") " pod="openshift-must-gather-jw8qc/crc-debug-7hwqt" Dec 06 04:52:05 crc kubenswrapper[4802]: I1206 04:52:05.756374 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bd8w2\" (UniqueName: \"kubernetes.io/projected/0d9cbd29-e544-49ec-96ef-890a3ce1ea19-kube-api-access-bd8w2\") pod \"crc-debug-7hwqt\" (UID: \"0d9cbd29-e544-49ec-96ef-890a3ce1ea19\") " pod="openshift-must-gather-jw8qc/crc-debug-7hwqt" Dec 06 04:52:05 crc kubenswrapper[4802]: I1206 04:52:05.756458 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0d9cbd29-e544-49ec-96ef-890a3ce1ea19-host\") pod \"crc-debug-7hwqt\" (UID: \"0d9cbd29-e544-49ec-96ef-890a3ce1ea19\") " pod="openshift-must-gather-jw8qc/crc-debug-7hwqt" Dec 06 04:52:05 crc kubenswrapper[4802]: I1206 04:52:05.773506 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bd8w2\" (UniqueName: \"kubernetes.io/projected/0d9cbd29-e544-49ec-96ef-890a3ce1ea19-kube-api-access-bd8w2\") pod \"crc-debug-7hwqt\" (UID: \"0d9cbd29-e544-49ec-96ef-890a3ce1ea19\") " pod="openshift-must-gather-jw8qc/crc-debug-7hwqt" Dec 06 04:52:05 crc kubenswrapper[4802]: I1206 04:52:05.891929 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jw8qc/crc-debug-7hwqt" Dec 06 04:52:06 crc kubenswrapper[4802]: I1206 04:52:06.256918 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jw8qc/crc-debug-7hwqt" event={"ID":"0d9cbd29-e544-49ec-96ef-890a3ce1ea19","Type":"ContainerStarted","Data":"56125e4880064f244b97c1dcd394f37393054024640494fdfd4c86d4489d152f"} Dec 06 04:52:06 crc kubenswrapper[4802]: I1206 04:52:06.257238 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jw8qc/crc-debug-7hwqt" event={"ID":"0d9cbd29-e544-49ec-96ef-890a3ce1ea19","Type":"ContainerStarted","Data":"4878b29666b49fb4f4d38593cf573e5b6ff6a7e3d81228bec0f9b3d95c6684ff"} Dec 06 04:52:06 crc kubenswrapper[4802]: I1206 04:52:06.280575 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-jw8qc/crc-debug-7hwqt" podStartSLOduration=1.28055831 podStartE2EDuration="1.28055831s" podCreationTimestamp="2025-12-06 04:52:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:52:06.275024829 +0000 UTC m=+4319.146933981" watchObservedRunningTime="2025-12-06 04:52:06.28055831 +0000 UTC m=+4319.152467462" Dec 06 04:52:07 crc kubenswrapper[4802]: I1206 04:52:07.269192 4802 generic.go:334] "Generic (PLEG): container finished" podID="0d9cbd29-e544-49ec-96ef-890a3ce1ea19" containerID="56125e4880064f244b97c1dcd394f37393054024640494fdfd4c86d4489d152f" exitCode=0 Dec 06 04:52:07 crc kubenswrapper[4802]: I1206 04:52:07.269247 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jw8qc/crc-debug-7hwqt" event={"ID":"0d9cbd29-e544-49ec-96ef-890a3ce1ea19","Type":"ContainerDied","Data":"56125e4880064f244b97c1dcd394f37393054024640494fdfd4c86d4489d152f"} Dec 06 04:52:08 crc kubenswrapper[4802]: I1206 04:52:08.372684 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jw8qc/crc-debug-7hwqt" Dec 06 04:52:08 crc kubenswrapper[4802]: I1206 04:52:08.438541 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-jw8qc/crc-debug-7hwqt"] Dec 06 04:52:08 crc kubenswrapper[4802]: I1206 04:52:08.448123 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-jw8qc/crc-debug-7hwqt"] Dec 06 04:52:08 crc kubenswrapper[4802]: I1206 04:52:08.504213 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0d9cbd29-e544-49ec-96ef-890a3ce1ea19-host\") pod \"0d9cbd29-e544-49ec-96ef-890a3ce1ea19\" (UID: \"0d9cbd29-e544-49ec-96ef-890a3ce1ea19\") " Dec 06 04:52:08 crc kubenswrapper[4802]: I1206 04:52:08.504305 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0d9cbd29-e544-49ec-96ef-890a3ce1ea19-host" (OuterVolumeSpecName: "host") pod "0d9cbd29-e544-49ec-96ef-890a3ce1ea19" (UID: "0d9cbd29-e544-49ec-96ef-890a3ce1ea19"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 04:52:08 crc kubenswrapper[4802]: I1206 04:52:08.504487 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bd8w2\" (UniqueName: \"kubernetes.io/projected/0d9cbd29-e544-49ec-96ef-890a3ce1ea19-kube-api-access-bd8w2\") pod \"0d9cbd29-e544-49ec-96ef-890a3ce1ea19\" (UID: \"0d9cbd29-e544-49ec-96ef-890a3ce1ea19\") " Dec 06 04:52:08 crc kubenswrapper[4802]: I1206 04:52:08.504974 4802 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0d9cbd29-e544-49ec-96ef-890a3ce1ea19-host\") on node \"crc\" DevicePath \"\"" Dec 06 04:52:08 crc kubenswrapper[4802]: I1206 04:52:08.510016 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d9cbd29-e544-49ec-96ef-890a3ce1ea19-kube-api-access-bd8w2" (OuterVolumeSpecName: "kube-api-access-bd8w2") pod "0d9cbd29-e544-49ec-96ef-890a3ce1ea19" (UID: "0d9cbd29-e544-49ec-96ef-890a3ce1ea19"). InnerVolumeSpecName "kube-api-access-bd8w2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:52:08 crc kubenswrapper[4802]: I1206 04:52:08.606503 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bd8w2\" (UniqueName: \"kubernetes.io/projected/0d9cbd29-e544-49ec-96ef-890a3ce1ea19-kube-api-access-bd8w2\") on node \"crc\" DevicePath \"\"" Dec 06 04:52:09 crc kubenswrapper[4802]: I1206 04:52:09.291478 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4878b29666b49fb4f4d38593cf573e5b6ff6a7e3d81228bec0f9b3d95c6684ff" Dec 06 04:52:09 crc kubenswrapper[4802]: I1206 04:52:09.291485 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jw8qc/crc-debug-7hwqt" Dec 06 04:52:09 crc kubenswrapper[4802]: I1206 04:52:09.462659 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d9cbd29-e544-49ec-96ef-890a3ce1ea19" path="/var/lib/kubelet/pods/0d9cbd29-e544-49ec-96ef-890a3ce1ea19/volumes" Dec 06 04:52:09 crc kubenswrapper[4802]: I1206 04:52:09.586434 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-jw8qc/crc-debug-nrvp8"] Dec 06 04:52:09 crc kubenswrapper[4802]: E1206 04:52:09.587077 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d9cbd29-e544-49ec-96ef-890a3ce1ea19" containerName="container-00" Dec 06 04:52:09 crc kubenswrapper[4802]: I1206 04:52:09.587094 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d9cbd29-e544-49ec-96ef-890a3ce1ea19" containerName="container-00" Dec 06 04:52:09 crc kubenswrapper[4802]: I1206 04:52:09.587271 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d9cbd29-e544-49ec-96ef-890a3ce1ea19" containerName="container-00" Dec 06 04:52:09 crc kubenswrapper[4802]: I1206 04:52:09.587932 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jw8qc/crc-debug-nrvp8" Dec 06 04:52:09 crc kubenswrapper[4802]: I1206 04:52:09.731417 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5brh\" (UniqueName: \"kubernetes.io/projected/29e3d173-8a4d-4f6a-afa7-d3d992e9400a-kube-api-access-k5brh\") pod \"crc-debug-nrvp8\" (UID: \"29e3d173-8a4d-4f6a-afa7-d3d992e9400a\") " pod="openshift-must-gather-jw8qc/crc-debug-nrvp8" Dec 06 04:52:09 crc kubenswrapper[4802]: I1206 04:52:09.731654 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/29e3d173-8a4d-4f6a-afa7-d3d992e9400a-host\") pod \"crc-debug-nrvp8\" (UID: \"29e3d173-8a4d-4f6a-afa7-d3d992e9400a\") " pod="openshift-must-gather-jw8qc/crc-debug-nrvp8" Dec 06 04:52:09 crc kubenswrapper[4802]: I1206 04:52:09.833899 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/29e3d173-8a4d-4f6a-afa7-d3d992e9400a-host\") pod \"crc-debug-nrvp8\" (UID: \"29e3d173-8a4d-4f6a-afa7-d3d992e9400a\") " pod="openshift-must-gather-jw8qc/crc-debug-nrvp8" Dec 06 04:52:09 crc kubenswrapper[4802]: I1206 04:52:09.834067 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5brh\" (UniqueName: \"kubernetes.io/projected/29e3d173-8a4d-4f6a-afa7-d3d992e9400a-kube-api-access-k5brh\") pod \"crc-debug-nrvp8\" (UID: \"29e3d173-8a4d-4f6a-afa7-d3d992e9400a\") " pod="openshift-must-gather-jw8qc/crc-debug-nrvp8" Dec 06 04:52:09 crc kubenswrapper[4802]: I1206 04:52:09.834374 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/29e3d173-8a4d-4f6a-afa7-d3d992e9400a-host\") pod \"crc-debug-nrvp8\" (UID: \"29e3d173-8a4d-4f6a-afa7-d3d992e9400a\") " pod="openshift-must-gather-jw8qc/crc-debug-nrvp8" Dec 06 04:52:09 crc kubenswrapper[4802]: I1206 04:52:09.864647 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5brh\" (UniqueName: \"kubernetes.io/projected/29e3d173-8a4d-4f6a-afa7-d3d992e9400a-kube-api-access-k5brh\") pod \"crc-debug-nrvp8\" (UID: \"29e3d173-8a4d-4f6a-afa7-d3d992e9400a\") " pod="openshift-must-gather-jw8qc/crc-debug-nrvp8" Dec 06 04:52:09 crc kubenswrapper[4802]: I1206 04:52:09.904366 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jw8qc/crc-debug-nrvp8" Dec 06 04:52:09 crc kubenswrapper[4802]: W1206 04:52:09.943783 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod29e3d173_8a4d_4f6a_afa7_d3d992e9400a.slice/crio-a26e0b6053f65f34d4fd36be22d3c229e9986ff99c3a7b2887a4a64706f3d1f1 WatchSource:0}: Error finding container a26e0b6053f65f34d4fd36be22d3c229e9986ff99c3a7b2887a4a64706f3d1f1: Status 404 returned error can't find the container with id a26e0b6053f65f34d4fd36be22d3c229e9986ff99c3a7b2887a4a64706f3d1f1 Dec 06 04:52:10 crc kubenswrapper[4802]: I1206 04:52:10.302036 4802 generic.go:334] "Generic (PLEG): container finished" podID="29e3d173-8a4d-4f6a-afa7-d3d992e9400a" containerID="8d96ec8f5e773a8afc375075cea9b9723a33cb461a4a15f16ed42ffc4555867b" exitCode=0 Dec 06 04:52:10 crc kubenswrapper[4802]: I1206 04:52:10.302193 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jw8qc/crc-debug-nrvp8" event={"ID":"29e3d173-8a4d-4f6a-afa7-d3d992e9400a","Type":"ContainerDied","Data":"8d96ec8f5e773a8afc375075cea9b9723a33cb461a4a15f16ed42ffc4555867b"} Dec 06 04:52:10 crc kubenswrapper[4802]: I1206 04:52:10.302769 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jw8qc/crc-debug-nrvp8" event={"ID":"29e3d173-8a4d-4f6a-afa7-d3d992e9400a","Type":"ContainerStarted","Data":"a26e0b6053f65f34d4fd36be22d3c229e9986ff99c3a7b2887a4a64706f3d1f1"} Dec 06 04:52:10 crc kubenswrapper[4802]: I1206 04:52:10.339122 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-jw8qc/crc-debug-nrvp8"] Dec 06 04:52:10 crc kubenswrapper[4802]: I1206 04:52:10.348771 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-jw8qc/crc-debug-nrvp8"] Dec 06 04:52:11 crc kubenswrapper[4802]: I1206 04:52:11.425048 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jw8qc/crc-debug-nrvp8" Dec 06 04:52:11 crc kubenswrapper[4802]: I1206 04:52:11.566266 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k5brh\" (UniqueName: \"kubernetes.io/projected/29e3d173-8a4d-4f6a-afa7-d3d992e9400a-kube-api-access-k5brh\") pod \"29e3d173-8a4d-4f6a-afa7-d3d992e9400a\" (UID: \"29e3d173-8a4d-4f6a-afa7-d3d992e9400a\") " Dec 06 04:52:11 crc kubenswrapper[4802]: I1206 04:52:11.566413 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/29e3d173-8a4d-4f6a-afa7-d3d992e9400a-host\") pod \"29e3d173-8a4d-4f6a-afa7-d3d992e9400a\" (UID: \"29e3d173-8a4d-4f6a-afa7-d3d992e9400a\") " Dec 06 04:52:11 crc kubenswrapper[4802]: I1206 04:52:11.566638 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/29e3d173-8a4d-4f6a-afa7-d3d992e9400a-host" (OuterVolumeSpecName: "host") pod "29e3d173-8a4d-4f6a-afa7-d3d992e9400a" (UID: "29e3d173-8a4d-4f6a-afa7-d3d992e9400a"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 04:52:11 crc kubenswrapper[4802]: I1206 04:52:11.568686 4802 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/29e3d173-8a4d-4f6a-afa7-d3d992e9400a-host\") on node \"crc\" DevicePath \"\"" Dec 06 04:52:11 crc kubenswrapper[4802]: I1206 04:52:11.571946 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29e3d173-8a4d-4f6a-afa7-d3d992e9400a-kube-api-access-k5brh" (OuterVolumeSpecName: "kube-api-access-k5brh") pod "29e3d173-8a4d-4f6a-afa7-d3d992e9400a" (UID: "29e3d173-8a4d-4f6a-afa7-d3d992e9400a"). InnerVolumeSpecName "kube-api-access-k5brh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:52:11 crc kubenswrapper[4802]: I1206 04:52:11.670623 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k5brh\" (UniqueName: \"kubernetes.io/projected/29e3d173-8a4d-4f6a-afa7-d3d992e9400a-kube-api-access-k5brh\") on node \"crc\" DevicePath \"\"" Dec 06 04:52:12 crc kubenswrapper[4802]: I1206 04:52:12.329394 4802 scope.go:117] "RemoveContainer" containerID="8d96ec8f5e773a8afc375075cea9b9723a33cb461a4a15f16ed42ffc4555867b" Dec 06 04:52:12 crc kubenswrapper[4802]: I1206 04:52:12.329449 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jw8qc/crc-debug-nrvp8" Dec 06 04:52:13 crc kubenswrapper[4802]: I1206 04:52:13.283597 4802 patch_prober.go:28] interesting pod/machine-config-daemon-zpxxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:52:13 crc kubenswrapper[4802]: I1206 04:52:13.283670 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:52:13 crc kubenswrapper[4802]: I1206 04:52:13.283721 4802 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" Dec 06 04:52:13 crc kubenswrapper[4802]: I1206 04:52:13.284635 4802 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4334df6618bec8f6cedfb779435b2c4d4e0f7e149dd426e245ab29892f475208"} pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 04:52:13 crc kubenswrapper[4802]: I1206 04:52:13.284716 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" containerID="cri-o://4334df6618bec8f6cedfb779435b2c4d4e0f7e149dd426e245ab29892f475208" gracePeriod=600 Dec 06 04:52:13 crc kubenswrapper[4802]: I1206 04:52:13.458897 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="29e3d173-8a4d-4f6a-afa7-d3d992e9400a" path="/var/lib/kubelet/pods/29e3d173-8a4d-4f6a-afa7-d3d992e9400a/volumes" Dec 06 04:52:13 crc kubenswrapper[4802]: E1206 04:52:13.580041 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:52:14 crc kubenswrapper[4802]: I1206 04:52:14.367965 4802 generic.go:334] "Generic (PLEG): container finished" podID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerID="4334df6618bec8f6cedfb779435b2c4d4e0f7e149dd426e245ab29892f475208" exitCode=0 Dec 06 04:52:14 crc kubenswrapper[4802]: I1206 04:52:14.368017 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" event={"ID":"6df38316-e0d3-4018-8d27-3620eba3a68d","Type":"ContainerDied","Data":"4334df6618bec8f6cedfb779435b2c4d4e0f7e149dd426e245ab29892f475208"} Dec 06 04:52:14 crc kubenswrapper[4802]: I1206 04:52:14.368059 4802 scope.go:117] "RemoveContainer" containerID="659eedd07500dc9cedfea13e32706d57d3d8130ed952a340282ebbc31bb990e2" Dec 06 04:52:14 crc kubenswrapper[4802]: I1206 04:52:14.368885 4802 scope.go:117] "RemoveContainer" containerID="4334df6618bec8f6cedfb779435b2c4d4e0f7e149dd426e245ab29892f475208" Dec 06 04:52:14 crc kubenswrapper[4802]: E1206 04:52:14.369300 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:52:29 crc kubenswrapper[4802]: I1206 04:52:29.295955 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9tf9p"] Dec 06 04:52:29 crc kubenswrapper[4802]: E1206 04:52:29.296983 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29e3d173-8a4d-4f6a-afa7-d3d992e9400a" containerName="container-00" Dec 06 04:52:29 crc kubenswrapper[4802]: I1206 04:52:29.296997 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="29e3d173-8a4d-4f6a-afa7-d3d992e9400a" containerName="container-00" Dec 06 04:52:29 crc kubenswrapper[4802]: I1206 04:52:29.297204 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="29e3d173-8a4d-4f6a-afa7-d3d992e9400a" containerName="container-00" Dec 06 04:52:29 crc kubenswrapper[4802]: I1206 04:52:29.304373 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9tf9p" Dec 06 04:52:29 crc kubenswrapper[4802]: I1206 04:52:29.308986 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9tf9p"] Dec 06 04:52:29 crc kubenswrapper[4802]: I1206 04:52:29.356874 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab924c08-8bf0-4f00-8798-866cbdfb04ba-catalog-content\") pod \"community-operators-9tf9p\" (UID: \"ab924c08-8bf0-4f00-8798-866cbdfb04ba\") " pod="openshift-marketplace/community-operators-9tf9p" Dec 06 04:52:29 crc kubenswrapper[4802]: I1206 04:52:29.356923 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab924c08-8bf0-4f00-8798-866cbdfb04ba-utilities\") pod \"community-operators-9tf9p\" (UID: \"ab924c08-8bf0-4f00-8798-866cbdfb04ba\") " pod="openshift-marketplace/community-operators-9tf9p" Dec 06 04:52:29 crc kubenswrapper[4802]: I1206 04:52:29.357207 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2j7n\" (UniqueName: \"kubernetes.io/projected/ab924c08-8bf0-4f00-8798-866cbdfb04ba-kube-api-access-r2j7n\") pod \"community-operators-9tf9p\" (UID: \"ab924c08-8bf0-4f00-8798-866cbdfb04ba\") " pod="openshift-marketplace/community-operators-9tf9p" Dec 06 04:52:29 crc kubenswrapper[4802]: I1206 04:52:29.450417 4802 scope.go:117] "RemoveContainer" containerID="4334df6618bec8f6cedfb779435b2c4d4e0f7e149dd426e245ab29892f475208" Dec 06 04:52:29 crc kubenswrapper[4802]: E1206 04:52:29.450663 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:52:29 crc kubenswrapper[4802]: I1206 04:52:29.458200 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2j7n\" (UniqueName: \"kubernetes.io/projected/ab924c08-8bf0-4f00-8798-866cbdfb04ba-kube-api-access-r2j7n\") pod \"community-operators-9tf9p\" (UID: \"ab924c08-8bf0-4f00-8798-866cbdfb04ba\") " pod="openshift-marketplace/community-operators-9tf9p" Dec 06 04:52:29 crc kubenswrapper[4802]: I1206 04:52:29.458329 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab924c08-8bf0-4f00-8798-866cbdfb04ba-catalog-content\") pod \"community-operators-9tf9p\" (UID: \"ab924c08-8bf0-4f00-8798-866cbdfb04ba\") " pod="openshift-marketplace/community-operators-9tf9p" Dec 06 04:52:29 crc kubenswrapper[4802]: I1206 04:52:29.458351 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab924c08-8bf0-4f00-8798-866cbdfb04ba-utilities\") pod \"community-operators-9tf9p\" (UID: \"ab924c08-8bf0-4f00-8798-866cbdfb04ba\") " pod="openshift-marketplace/community-operators-9tf9p" Dec 06 04:52:29 crc kubenswrapper[4802]: I1206 04:52:29.458823 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab924c08-8bf0-4f00-8798-866cbdfb04ba-utilities\") pod \"community-operators-9tf9p\" (UID: \"ab924c08-8bf0-4f00-8798-866cbdfb04ba\") " pod="openshift-marketplace/community-operators-9tf9p" Dec 06 04:52:29 crc kubenswrapper[4802]: I1206 04:52:29.459625 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab924c08-8bf0-4f00-8798-866cbdfb04ba-catalog-content\") pod \"community-operators-9tf9p\" (UID: \"ab924c08-8bf0-4f00-8798-866cbdfb04ba\") " pod="openshift-marketplace/community-operators-9tf9p" Dec 06 04:52:29 crc kubenswrapper[4802]: I1206 04:52:29.482592 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2j7n\" (UniqueName: \"kubernetes.io/projected/ab924c08-8bf0-4f00-8798-866cbdfb04ba-kube-api-access-r2j7n\") pod \"community-operators-9tf9p\" (UID: \"ab924c08-8bf0-4f00-8798-866cbdfb04ba\") " pod="openshift-marketplace/community-operators-9tf9p" Dec 06 04:52:29 crc kubenswrapper[4802]: I1206 04:52:29.638516 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9tf9p" Dec 06 04:52:30 crc kubenswrapper[4802]: I1206 04:52:30.226984 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9tf9p"] Dec 06 04:52:30 crc kubenswrapper[4802]: I1206 04:52:30.519882 4802 generic.go:334] "Generic (PLEG): container finished" podID="ab924c08-8bf0-4f00-8798-866cbdfb04ba" containerID="8c07245b7d76db5731e27d3771e781e7c11d1069114ead43bd42735b752e4bc6" exitCode=0 Dec 06 04:52:30 crc kubenswrapper[4802]: I1206 04:52:30.519925 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9tf9p" event={"ID":"ab924c08-8bf0-4f00-8798-866cbdfb04ba","Type":"ContainerDied","Data":"8c07245b7d76db5731e27d3771e781e7c11d1069114ead43bd42735b752e4bc6"} Dec 06 04:52:30 crc kubenswrapper[4802]: I1206 04:52:30.519951 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9tf9p" event={"ID":"ab924c08-8bf0-4f00-8798-866cbdfb04ba","Type":"ContainerStarted","Data":"352b2885f0a0289848adad120fa86048b28f6005e007801c7bbd60605db94f80"} Dec 06 04:52:30 crc kubenswrapper[4802]: I1206 04:52:30.522388 4802 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 04:52:32 crc kubenswrapper[4802]: I1206 04:52:32.537581 4802 generic.go:334] "Generic (PLEG): container finished" podID="ab924c08-8bf0-4f00-8798-866cbdfb04ba" containerID="3078f49aee039420e278eccf61aa2e77b16c33f8f9dabb2dfb61470af93ead1a" exitCode=0 Dec 06 04:52:32 crc kubenswrapper[4802]: I1206 04:52:32.537680 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9tf9p" event={"ID":"ab924c08-8bf0-4f00-8798-866cbdfb04ba","Type":"ContainerDied","Data":"3078f49aee039420e278eccf61aa2e77b16c33f8f9dabb2dfb61470af93ead1a"} Dec 06 04:52:32 crc kubenswrapper[4802]: I1206 04:52:32.764236 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7f79b98d58-l9qc5_21e91f8f-b3aa-4a45-85e6-5192a3f85a04/barbican-api/0.log" Dec 06 04:52:32 crc kubenswrapper[4802]: I1206 04:52:32.919868 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7f79b98d58-l9qc5_21e91f8f-b3aa-4a45-85e6-5192a3f85a04/barbican-api-log/0.log" Dec 06 04:52:32 crc kubenswrapper[4802]: I1206 04:52:32.920614 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-554f6fb55d-hdgjt_b443b1b9-ec2c-499d-ae4d-1937c6f156f7/barbican-keystone-listener/0.log" Dec 06 04:52:33 crc kubenswrapper[4802]: I1206 04:52:33.068577 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-554f6fb55d-hdgjt_b443b1b9-ec2c-499d-ae4d-1937c6f156f7/barbican-keystone-listener-log/0.log" Dec 06 04:52:33 crc kubenswrapper[4802]: I1206 04:52:33.100550 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-vqlpv"] Dec 06 04:52:33 crc kubenswrapper[4802]: I1206 04:52:33.102488 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vqlpv" Dec 06 04:52:33 crc kubenswrapper[4802]: I1206 04:52:33.113708 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vqlpv"] Dec 06 04:52:33 crc kubenswrapper[4802]: I1206 04:52:33.127210 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6c8r\" (UniqueName: \"kubernetes.io/projected/0a7427d8-b298-4e7f-bcd4-b36d8787aa07-kube-api-access-q6c8r\") pod \"certified-operators-vqlpv\" (UID: \"0a7427d8-b298-4e7f-bcd4-b36d8787aa07\") " pod="openshift-marketplace/certified-operators-vqlpv" Dec 06 04:52:33 crc kubenswrapper[4802]: I1206 04:52:33.127317 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a7427d8-b298-4e7f-bcd4-b36d8787aa07-catalog-content\") pod \"certified-operators-vqlpv\" (UID: \"0a7427d8-b298-4e7f-bcd4-b36d8787aa07\") " pod="openshift-marketplace/certified-operators-vqlpv" Dec 06 04:52:33 crc kubenswrapper[4802]: I1206 04:52:33.127380 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a7427d8-b298-4e7f-bcd4-b36d8787aa07-utilities\") pod \"certified-operators-vqlpv\" (UID: \"0a7427d8-b298-4e7f-bcd4-b36d8787aa07\") " pod="openshift-marketplace/certified-operators-vqlpv" Dec 06 04:52:33 crc kubenswrapper[4802]: I1206 04:52:33.214623 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-68998b57f5-6qgkl_57a20a46-fdfb-410f-8a13-0b58fa5751da/barbican-worker/0.log" Dec 06 04:52:33 crc kubenswrapper[4802]: I1206 04:52:33.228660 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a7427d8-b298-4e7f-bcd4-b36d8787aa07-catalog-content\") pod \"certified-operators-vqlpv\" (UID: \"0a7427d8-b298-4e7f-bcd4-b36d8787aa07\") " pod="openshift-marketplace/certified-operators-vqlpv" Dec 06 04:52:33 crc kubenswrapper[4802]: I1206 04:52:33.228739 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a7427d8-b298-4e7f-bcd4-b36d8787aa07-utilities\") pod \"certified-operators-vqlpv\" (UID: \"0a7427d8-b298-4e7f-bcd4-b36d8787aa07\") " pod="openshift-marketplace/certified-operators-vqlpv" Dec 06 04:52:33 crc kubenswrapper[4802]: I1206 04:52:33.228794 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6c8r\" (UniqueName: \"kubernetes.io/projected/0a7427d8-b298-4e7f-bcd4-b36d8787aa07-kube-api-access-q6c8r\") pod \"certified-operators-vqlpv\" (UID: \"0a7427d8-b298-4e7f-bcd4-b36d8787aa07\") " pod="openshift-marketplace/certified-operators-vqlpv" Dec 06 04:52:33 crc kubenswrapper[4802]: I1206 04:52:33.229237 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a7427d8-b298-4e7f-bcd4-b36d8787aa07-utilities\") pod \"certified-operators-vqlpv\" (UID: \"0a7427d8-b298-4e7f-bcd4-b36d8787aa07\") " pod="openshift-marketplace/certified-operators-vqlpv" Dec 06 04:52:33 crc kubenswrapper[4802]: I1206 04:52:33.229270 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a7427d8-b298-4e7f-bcd4-b36d8787aa07-catalog-content\") pod \"certified-operators-vqlpv\" (UID: \"0a7427d8-b298-4e7f-bcd4-b36d8787aa07\") " pod="openshift-marketplace/certified-operators-vqlpv" Dec 06 04:52:33 crc kubenswrapper[4802]: I1206 04:52:33.248500 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6c8r\" (UniqueName: \"kubernetes.io/projected/0a7427d8-b298-4e7f-bcd4-b36d8787aa07-kube-api-access-q6c8r\") pod \"certified-operators-vqlpv\" (UID: \"0a7427d8-b298-4e7f-bcd4-b36d8787aa07\") " pod="openshift-marketplace/certified-operators-vqlpv" Dec 06 04:52:33 crc kubenswrapper[4802]: I1206 04:52:33.258175 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-68998b57f5-6qgkl_57a20a46-fdfb-410f-8a13-0b58fa5751da/barbican-worker-log/0.log" Dec 06 04:52:33 crc kubenswrapper[4802]: I1206 04:52:33.429647 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vqlpv" Dec 06 04:52:33 crc kubenswrapper[4802]: I1206 04:52:33.580280 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9tf9p" event={"ID":"ab924c08-8bf0-4f00-8798-866cbdfb04ba","Type":"ContainerStarted","Data":"0e80eb7b72b812173c9701d09e1124a0aa65342071e5d59155e84220b9fa8fae"} Dec 06 04:52:33 crc kubenswrapper[4802]: I1206 04:52:33.606081 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3ede1365-588f-4048-9002-034d3dbd2ef3/ceilometer-central-agent/0.log" Dec 06 04:52:33 crc kubenswrapper[4802]: I1206 04:52:33.642647 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9tf9p" podStartSLOduration=2.225931651 podStartE2EDuration="4.64262419s" podCreationTimestamp="2025-12-06 04:52:29 +0000 UTC" firstStartedPulling="2025-12-06 04:52:30.522179225 +0000 UTC m=+4343.394088377" lastFinishedPulling="2025-12-06 04:52:32.938871764 +0000 UTC m=+4345.810780916" observedRunningTime="2025-12-06 04:52:33.623633093 +0000 UTC m=+4346.495542235" watchObservedRunningTime="2025-12-06 04:52:33.64262419 +0000 UTC m=+4346.514533342" Dec 06 04:52:33 crc kubenswrapper[4802]: I1206 04:52:33.686683 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-n5lkw_a07f3a26-501d-49cb-8dbf-648f1cac66da/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 04:52:33 crc kubenswrapper[4802]: I1206 04:52:33.934198 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3ede1365-588f-4048-9002-034d3dbd2ef3/ceilometer-notification-agent/0.log" Dec 06 04:52:33 crc kubenswrapper[4802]: I1206 04:52:33.966224 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vqlpv"] Dec 06 04:52:33 crc kubenswrapper[4802]: W1206 04:52:33.969640 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0a7427d8_b298_4e7f_bcd4_b36d8787aa07.slice/crio-fe4a65d66b51c4128c0459a5becc02de78a074363d556de635c9f291cdd3c29a WatchSource:0}: Error finding container fe4a65d66b51c4128c0459a5becc02de78a074363d556de635c9f291cdd3c29a: Status 404 returned error can't find the container with id fe4a65d66b51c4128c0459a5becc02de78a074363d556de635c9f291cdd3c29a Dec 06 04:52:34 crc kubenswrapper[4802]: I1206 04:52:34.087291 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3ede1365-588f-4048-9002-034d3dbd2ef3/sg-core/0.log" Dec 06 04:52:34 crc kubenswrapper[4802]: I1206 04:52:34.341905 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3ede1365-588f-4048-9002-034d3dbd2ef3/proxy-httpd/0.log" Dec 06 04:52:34 crc kubenswrapper[4802]: I1206 04:52:34.476242 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_45876a53-aa92-4ca3-bbc2-170e9855b034/cinder-api/0.log" Dec 06 04:52:34 crc kubenswrapper[4802]: I1206 04:52:34.588336 4802 generic.go:334] "Generic (PLEG): container finished" podID="0a7427d8-b298-4e7f-bcd4-b36d8787aa07" containerID="0789dadd2c8066bdf1a99439432fc812a6148f5555b35935e00af23abdc01804" exitCode=0 Dec 06 04:52:34 crc kubenswrapper[4802]: I1206 04:52:34.588377 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vqlpv" event={"ID":"0a7427d8-b298-4e7f-bcd4-b36d8787aa07","Type":"ContainerDied","Data":"0789dadd2c8066bdf1a99439432fc812a6148f5555b35935e00af23abdc01804"} Dec 06 04:52:34 crc kubenswrapper[4802]: I1206 04:52:34.588417 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vqlpv" event={"ID":"0a7427d8-b298-4e7f-bcd4-b36d8787aa07","Type":"ContainerStarted","Data":"fe4a65d66b51c4128c0459a5becc02de78a074363d556de635c9f291cdd3c29a"} Dec 06 04:52:34 crc kubenswrapper[4802]: I1206 04:52:34.591036 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_45876a53-aa92-4ca3-bbc2-170e9855b034/cinder-api-log/0.log" Dec 06 04:52:34 crc kubenswrapper[4802]: I1206 04:52:34.751906 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_92c245ae-d893-44d7-99e3-69a9e58735cf/probe/0.log" Dec 06 04:52:34 crc kubenswrapper[4802]: I1206 04:52:34.757436 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_92c245ae-d893-44d7-99e3-69a9e58735cf/cinder-scheduler/0.log" Dec 06 04:52:35 crc kubenswrapper[4802]: I1206 04:52:35.304150 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-22wfn_7d07f677-e982-45d5-b4c0-92e0db096fd1/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 04:52:35 crc kubenswrapper[4802]: I1206 04:52:35.413588 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-zd97g_3d1e4ffc-0f3b-4bc2-9f36-6193b7f33785/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 04:52:35 crc kubenswrapper[4802]: I1206 04:52:35.599063 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vqlpv" event={"ID":"0a7427d8-b298-4e7f-bcd4-b36d8787aa07","Type":"ContainerStarted","Data":"39c805d0a9518680afbdadb58d3f3a0ecdbad13d6a19fee7df3e0ec65f6eb7fc"} Dec 06 04:52:35 crc kubenswrapper[4802]: I1206 04:52:35.629278 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5d75f767dc-75bb8_ad17cfaf-5662-41ae-9987-766441edb7a4/init/0.log" Dec 06 04:52:35 crc kubenswrapper[4802]: I1206 04:52:35.817509 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5d75f767dc-75bb8_ad17cfaf-5662-41ae-9987-766441edb7a4/init/0.log" Dec 06 04:52:35 crc kubenswrapper[4802]: I1206 04:52:35.909352 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-sgxfp_db0df1b2-ef63-42ae-96ff-cafa00acc8e1/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 04:52:35 crc kubenswrapper[4802]: I1206 04:52:35.912661 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5d75f767dc-75bb8_ad17cfaf-5662-41ae-9987-766441edb7a4/dnsmasq-dns/0.log" Dec 06 04:52:36 crc kubenswrapper[4802]: I1206 04:52:36.165162 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_02888e6b-6185-45b8-84b7-8b3654270b77/glance-log/0.log" Dec 06 04:52:36 crc kubenswrapper[4802]: I1206 04:52:36.340409 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_02888e6b-6185-45b8-84b7-8b3654270b77/glance-httpd/0.log" Dec 06 04:52:36 crc kubenswrapper[4802]: I1206 04:52:36.499757 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_19c7331e-7c80-4aa6-8a9f-2a7f878c79f2/glance-log/0.log" Dec 06 04:52:36 crc kubenswrapper[4802]: I1206 04:52:36.552965 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_19c7331e-7c80-4aa6-8a9f-2a7f878c79f2/glance-httpd/0.log" Dec 06 04:52:37 crc kubenswrapper[4802]: I1206 04:52:37.586241 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-55fc64bfdd-779rm_a86ea592-5d80-4415-8c1f-5471274999d4/heat-engine/0.log" Dec 06 04:52:37 crc kubenswrapper[4802]: I1206 04:52:37.606127 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-bsr6g_79de9688-9efb-44b4-9518-259e1f58b5b4/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 04:52:37 crc kubenswrapper[4802]: I1206 04:52:37.617479 4802 generic.go:334] "Generic (PLEG): container finished" podID="0a7427d8-b298-4e7f-bcd4-b36d8787aa07" containerID="39c805d0a9518680afbdadb58d3f3a0ecdbad13d6a19fee7df3e0ec65f6eb7fc" exitCode=0 Dec 06 04:52:37 crc kubenswrapper[4802]: I1206 04:52:37.617545 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vqlpv" event={"ID":"0a7427d8-b298-4e7f-bcd4-b36d8787aa07","Type":"ContainerDied","Data":"39c805d0a9518680afbdadb58d3f3a0ecdbad13d6a19fee7df3e0ec65f6eb7fc"} Dec 06 04:52:37 crc kubenswrapper[4802]: I1206 04:52:37.836942 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-g9qlw_b6c3bbd2-7993-420e-91d7-ffc72b1d2439/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 04:52:38 crc kubenswrapper[4802]: I1206 04:52:38.033164 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-6ddd7d8889-js75h_71fc6de6-20c6-49fe-9780-3cfdb0002c62/heat-api/0.log" Dec 06 04:52:38 crc kubenswrapper[4802]: I1206 04:52:38.295988 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_5489ef4c-603a-44ce-9a0e-6ad56188e669/kube-state-metrics/0.log" Dec 06 04:52:38 crc kubenswrapper[4802]: I1206 04:52:38.879722 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29416561-rnfxm_613478db-c10d-476b-b66d-4ae3572067ff/keystone-cron/0.log" Dec 06 04:52:38 crc kubenswrapper[4802]: I1206 04:52:38.975077 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-788b9d8f4c-rk4vn_10210aa2-6d6b-4422-ad35-32b04e1444f2/heat-cfnapi/0.log" Dec 06 04:52:38 crc kubenswrapper[4802]: I1206 04:52:38.993511 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-tckw2_18179f2f-334f-48c2-a6be-6d1c1792e600/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 04:52:39 crc kubenswrapper[4802]: I1206 04:52:39.260916 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-75dc574896-4vrbs_bdd2c50b-481a-4ab7-90ce-86cb69efd255/keystone-api/0.log" Dec 06 04:52:39 crc kubenswrapper[4802]: I1206 04:52:39.639656 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9tf9p" Dec 06 04:52:39 crc kubenswrapper[4802]: I1206 04:52:39.639929 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9tf9p" Dec 06 04:52:39 crc kubenswrapper[4802]: I1206 04:52:39.691863 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9tf9p" Dec 06 04:52:40 crc kubenswrapper[4802]: I1206 04:52:40.236539 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-lpc95_64272e53-eccf-4cb7-9176-0bc95c5b7e10/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 04:52:40 crc kubenswrapper[4802]: I1206 04:52:40.258410 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-68bd774b7c-9sjk9_8bf882a4-ac26-40da-8bcc-46052a9db457/neutron-httpd/0.log" Dec 06 04:52:40 crc kubenswrapper[4802]: I1206 04:52:40.362677 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-68bd774b7c-9sjk9_8bf882a4-ac26-40da-8bcc-46052a9db457/neutron-api/0.log" Dec 06 04:52:40 crc kubenswrapper[4802]: I1206 04:52:40.449653 4802 scope.go:117] "RemoveContainer" containerID="4334df6618bec8f6cedfb779435b2c4d4e0f7e149dd426e245ab29892f475208" Dec 06 04:52:40 crc kubenswrapper[4802]: E1206 04:52:40.450237 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:52:40 crc kubenswrapper[4802]: I1206 04:52:40.697358 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9tf9p" Dec 06 04:52:40 crc kubenswrapper[4802]: I1206 04:52:40.899262 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_2c1b64f2-04f9-4fa9-ba9b-852f1383f6ed/nova-cell0-conductor-conductor/0.log" Dec 06 04:52:40 crc kubenswrapper[4802]: I1206 04:52:40.900885 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_add9c7ee-92a7-4cd8-adbf-02ec103cb6dc/nova-api-log/0.log" Dec 06 04:52:41 crc kubenswrapper[4802]: I1206 04:52:41.022171 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_add9c7ee-92a7-4cd8-adbf-02ec103cb6dc/nova-api-api/0.log" Dec 06 04:52:41 crc kubenswrapper[4802]: I1206 04:52:41.381677 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_ede13563-3589-4889-a568-55af534782c5/nova-cell1-novncproxy-novncproxy/0.log" Dec 06 04:52:41 crc kubenswrapper[4802]: I1206 04:52:41.404225 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_eaeadc44-2583-4e10-b0d3-15b754927734/nova-cell1-conductor-conductor/0.log" Dec 06 04:52:41 crc kubenswrapper[4802]: I1206 04:52:41.482887 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9tf9p"] Dec 06 04:52:41 crc kubenswrapper[4802]: I1206 04:52:41.653931 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vqlpv" event={"ID":"0a7427d8-b298-4e7f-bcd4-b36d8787aa07","Type":"ContainerStarted","Data":"017267762f1784908d85c41c76a882a83ebc60cf7772f1eb9aa2341467d45ba4"} Dec 06 04:52:41 crc kubenswrapper[4802]: I1206 04:52:41.664743 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-5plkw_b8e8984b-b60b-48e5-9586-1a1a8d883143/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 04:52:41 crc kubenswrapper[4802]: I1206 04:52:41.682687 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-vqlpv" podStartSLOduration=2.502353025 podStartE2EDuration="8.682667346s" podCreationTimestamp="2025-12-06 04:52:33 +0000 UTC" firstStartedPulling="2025-12-06 04:52:34.591057943 +0000 UTC m=+4347.462967095" lastFinishedPulling="2025-12-06 04:52:40.771372264 +0000 UTC m=+4353.643281416" observedRunningTime="2025-12-06 04:52:41.673703022 +0000 UTC m=+4354.545612174" watchObservedRunningTime="2025-12-06 04:52:41.682667346 +0000 UTC m=+4354.554576508" Dec 06 04:52:41 crc kubenswrapper[4802]: I1206 04:52:41.727281 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_f1350889-6def-43ce-9827-761796bea156/nova-metadata-log/0.log" Dec 06 04:52:42 crc kubenswrapper[4802]: I1206 04:52:42.244216 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_723826fa-70b3-4be3-9380-42561ca2041b/nova-scheduler-scheduler/0.log" Dec 06 04:52:42 crc kubenswrapper[4802]: I1206 04:52:42.308312 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_db5b3080-8b4d-4232-a7da-66e544c09c60/mysql-bootstrap/0.log" Dec 06 04:52:42 crc kubenswrapper[4802]: I1206 04:52:42.502050 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_db5b3080-8b4d-4232-a7da-66e544c09c60/galera/0.log" Dec 06 04:52:42 crc kubenswrapper[4802]: I1206 04:52:42.553793 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_db5b3080-8b4d-4232-a7da-66e544c09c60/mysql-bootstrap/0.log" Dec 06 04:52:42 crc kubenswrapper[4802]: I1206 04:52:42.662266 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9tf9p" podUID="ab924c08-8bf0-4f00-8798-866cbdfb04ba" containerName="registry-server" containerID="cri-o://0e80eb7b72b812173c9701d09e1124a0aa65342071e5d59155e84220b9fa8fae" gracePeriod=2 Dec 06 04:52:42 crc kubenswrapper[4802]: I1206 04:52:42.763503 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_04b6f630-c38a-411e-b48c-99690653972a/mysql-bootstrap/0.log" Dec 06 04:52:43 crc kubenswrapper[4802]: I1206 04:52:43.001216 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_04b6f630-c38a-411e-b48c-99690653972a/mysql-bootstrap/0.log" Dec 06 04:52:43 crc kubenswrapper[4802]: I1206 04:52:43.122667 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_04b6f630-c38a-411e-b48c-99690653972a/galera/0.log" Dec 06 04:52:43 crc kubenswrapper[4802]: I1206 04:52:43.187718 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9tf9p" Dec 06 04:52:43 crc kubenswrapper[4802]: I1206 04:52:43.255533 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_20c275f9-f37c-427a-b8f0-77c35fc880b5/openstackclient/0.log" Dec 06 04:52:43 crc kubenswrapper[4802]: I1206 04:52:43.326303 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab924c08-8bf0-4f00-8798-866cbdfb04ba-catalog-content\") pod \"ab924c08-8bf0-4f00-8798-866cbdfb04ba\" (UID: \"ab924c08-8bf0-4f00-8798-866cbdfb04ba\") " Dec 06 04:52:43 crc kubenswrapper[4802]: I1206 04:52:43.326388 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r2j7n\" (UniqueName: \"kubernetes.io/projected/ab924c08-8bf0-4f00-8798-866cbdfb04ba-kube-api-access-r2j7n\") pod \"ab924c08-8bf0-4f00-8798-866cbdfb04ba\" (UID: \"ab924c08-8bf0-4f00-8798-866cbdfb04ba\") " Dec 06 04:52:43 crc kubenswrapper[4802]: I1206 04:52:43.326508 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab924c08-8bf0-4f00-8798-866cbdfb04ba-utilities\") pod \"ab924c08-8bf0-4f00-8798-866cbdfb04ba\" (UID: \"ab924c08-8bf0-4f00-8798-866cbdfb04ba\") " Dec 06 04:52:43 crc kubenswrapper[4802]: I1206 04:52:43.328233 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ab924c08-8bf0-4f00-8798-866cbdfb04ba-utilities" (OuterVolumeSpecName: "utilities") pod "ab924c08-8bf0-4f00-8798-866cbdfb04ba" (UID: "ab924c08-8bf0-4f00-8798-866cbdfb04ba"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:52:43 crc kubenswrapper[4802]: I1206 04:52:43.349015 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab924c08-8bf0-4f00-8798-866cbdfb04ba-kube-api-access-r2j7n" (OuterVolumeSpecName: "kube-api-access-r2j7n") pod "ab924c08-8bf0-4f00-8798-866cbdfb04ba" (UID: "ab924c08-8bf0-4f00-8798-866cbdfb04ba"). InnerVolumeSpecName "kube-api-access-r2j7n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:52:43 crc kubenswrapper[4802]: I1206 04:52:43.381621 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ab924c08-8bf0-4f00-8798-866cbdfb04ba-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ab924c08-8bf0-4f00-8798-866cbdfb04ba" (UID: "ab924c08-8bf0-4f00-8798-866cbdfb04ba"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:52:43 crc kubenswrapper[4802]: I1206 04:52:43.400440 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-7sljg_11e2072e-0127-48c0-b417-ced9cfa9bf64/openstack-network-exporter/0.log" Dec 06 04:52:43 crc kubenswrapper[4802]: I1206 04:52:43.429140 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r2j7n\" (UniqueName: \"kubernetes.io/projected/ab924c08-8bf0-4f00-8798-866cbdfb04ba-kube-api-access-r2j7n\") on node \"crc\" DevicePath \"\"" Dec 06 04:52:43 crc kubenswrapper[4802]: I1206 04:52:43.429171 4802 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab924c08-8bf0-4f00-8798-866cbdfb04ba-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 04:52:43 crc kubenswrapper[4802]: I1206 04:52:43.429180 4802 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab924c08-8bf0-4f00-8798-866cbdfb04ba-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 04:52:43 crc kubenswrapper[4802]: I1206 04:52:43.430432 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-vqlpv" Dec 06 04:52:43 crc kubenswrapper[4802]: I1206 04:52:43.431530 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-vqlpv" Dec 06 04:52:43 crc kubenswrapper[4802]: I1206 04:52:43.489454 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-vqlpv" Dec 06 04:52:43 crc kubenswrapper[4802]: I1206 04:52:43.603340 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-49cxn_edb79ab5-ba10-4379-96e9-b2de42f6cbc8/ovsdb-server-init/0.log" Dec 06 04:52:43 crc kubenswrapper[4802]: I1206 04:52:43.671213 4802 generic.go:334] "Generic (PLEG): container finished" podID="ab924c08-8bf0-4f00-8798-866cbdfb04ba" containerID="0e80eb7b72b812173c9701d09e1124a0aa65342071e5d59155e84220b9fa8fae" exitCode=0 Dec 06 04:52:43 crc kubenswrapper[4802]: I1206 04:52:43.671269 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9tf9p" event={"ID":"ab924c08-8bf0-4f00-8798-866cbdfb04ba","Type":"ContainerDied","Data":"0e80eb7b72b812173c9701d09e1124a0aa65342071e5d59155e84220b9fa8fae"} Dec 06 04:52:43 crc kubenswrapper[4802]: I1206 04:52:43.671299 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9tf9p" Dec 06 04:52:43 crc kubenswrapper[4802]: I1206 04:52:43.671327 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9tf9p" event={"ID":"ab924c08-8bf0-4f00-8798-866cbdfb04ba","Type":"ContainerDied","Data":"352b2885f0a0289848adad120fa86048b28f6005e007801c7bbd60605db94f80"} Dec 06 04:52:43 crc kubenswrapper[4802]: I1206 04:52:43.671352 4802 scope.go:117] "RemoveContainer" containerID="0e80eb7b72b812173c9701d09e1124a0aa65342071e5d59155e84220b9fa8fae" Dec 06 04:52:43 crc kubenswrapper[4802]: I1206 04:52:43.676039 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_f1350889-6def-43ce-9827-761796bea156/nova-metadata-metadata/0.log" Dec 06 04:52:43 crc kubenswrapper[4802]: I1206 04:52:43.695285 4802 scope.go:117] "RemoveContainer" containerID="3078f49aee039420e278eccf61aa2e77b16c33f8f9dabb2dfb61470af93ead1a" Dec 06 04:52:43 crc kubenswrapper[4802]: I1206 04:52:43.699798 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9tf9p"] Dec 06 04:52:43 crc kubenswrapper[4802]: I1206 04:52:43.719641 4802 scope.go:117] "RemoveContainer" containerID="8c07245b7d76db5731e27d3771e781e7c11d1069114ead43bd42735b752e4bc6" Dec 06 04:52:43 crc kubenswrapper[4802]: I1206 04:52:43.729313 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9tf9p"] Dec 06 04:52:43 crc kubenswrapper[4802]: I1206 04:52:43.762945 4802 scope.go:117] "RemoveContainer" containerID="0e80eb7b72b812173c9701d09e1124a0aa65342071e5d59155e84220b9fa8fae" Dec 06 04:52:43 crc kubenswrapper[4802]: E1206 04:52:43.764710 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e80eb7b72b812173c9701d09e1124a0aa65342071e5d59155e84220b9fa8fae\": container with ID starting with 0e80eb7b72b812173c9701d09e1124a0aa65342071e5d59155e84220b9fa8fae not found: ID does not exist" containerID="0e80eb7b72b812173c9701d09e1124a0aa65342071e5d59155e84220b9fa8fae" Dec 06 04:52:43 crc kubenswrapper[4802]: I1206 04:52:43.764785 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e80eb7b72b812173c9701d09e1124a0aa65342071e5d59155e84220b9fa8fae"} err="failed to get container status \"0e80eb7b72b812173c9701d09e1124a0aa65342071e5d59155e84220b9fa8fae\": rpc error: code = NotFound desc = could not find container \"0e80eb7b72b812173c9701d09e1124a0aa65342071e5d59155e84220b9fa8fae\": container with ID starting with 0e80eb7b72b812173c9701d09e1124a0aa65342071e5d59155e84220b9fa8fae not found: ID does not exist" Dec 06 04:52:43 crc kubenswrapper[4802]: I1206 04:52:43.764822 4802 scope.go:117] "RemoveContainer" containerID="3078f49aee039420e278eccf61aa2e77b16c33f8f9dabb2dfb61470af93ead1a" Dec 06 04:52:43 crc kubenswrapper[4802]: E1206 04:52:43.765193 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3078f49aee039420e278eccf61aa2e77b16c33f8f9dabb2dfb61470af93ead1a\": container with ID starting with 3078f49aee039420e278eccf61aa2e77b16c33f8f9dabb2dfb61470af93ead1a not found: ID does not exist" containerID="3078f49aee039420e278eccf61aa2e77b16c33f8f9dabb2dfb61470af93ead1a" Dec 06 04:52:43 crc kubenswrapper[4802]: I1206 04:52:43.765237 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3078f49aee039420e278eccf61aa2e77b16c33f8f9dabb2dfb61470af93ead1a"} err="failed to get container status \"3078f49aee039420e278eccf61aa2e77b16c33f8f9dabb2dfb61470af93ead1a\": rpc error: code = NotFound desc = could not find container \"3078f49aee039420e278eccf61aa2e77b16c33f8f9dabb2dfb61470af93ead1a\": container with ID starting with 3078f49aee039420e278eccf61aa2e77b16c33f8f9dabb2dfb61470af93ead1a not found: ID does not exist" Dec 06 04:52:43 crc kubenswrapper[4802]: I1206 04:52:43.765272 4802 scope.go:117] "RemoveContainer" containerID="8c07245b7d76db5731e27d3771e781e7c11d1069114ead43bd42735b752e4bc6" Dec 06 04:52:43 crc kubenswrapper[4802]: E1206 04:52:43.766500 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c07245b7d76db5731e27d3771e781e7c11d1069114ead43bd42735b752e4bc6\": container with ID starting with 8c07245b7d76db5731e27d3771e781e7c11d1069114ead43bd42735b752e4bc6 not found: ID does not exist" containerID="8c07245b7d76db5731e27d3771e781e7c11d1069114ead43bd42735b752e4bc6" Dec 06 04:52:43 crc kubenswrapper[4802]: I1206 04:52:43.766531 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c07245b7d76db5731e27d3771e781e7c11d1069114ead43bd42735b752e4bc6"} err="failed to get container status \"8c07245b7d76db5731e27d3771e781e7c11d1069114ead43bd42735b752e4bc6\": rpc error: code = NotFound desc = could not find container \"8c07245b7d76db5731e27d3771e781e7c11d1069114ead43bd42735b752e4bc6\": container with ID starting with 8c07245b7d76db5731e27d3771e781e7c11d1069114ead43bd42735b752e4bc6 not found: ID does not exist" Dec 06 04:52:43 crc kubenswrapper[4802]: I1206 04:52:43.806985 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-49cxn_edb79ab5-ba10-4379-96e9-b2de42f6cbc8/ovs-vswitchd/0.log" Dec 06 04:52:43 crc kubenswrapper[4802]: I1206 04:52:43.832837 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-49cxn_edb79ab5-ba10-4379-96e9-b2de42f6cbc8/ovsdb-server-init/0.log" Dec 06 04:52:43 crc kubenswrapper[4802]: I1206 04:52:43.851246 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-49cxn_edb79ab5-ba10-4379-96e9-b2de42f6cbc8/ovsdb-server/0.log" Dec 06 04:52:44 crc kubenswrapper[4802]: I1206 04:52:44.015742 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-sbt8p_6e785e69-07cd-4c5c-8446-3d2a0d7e7b0b/ovn-controller/0.log" Dec 06 04:52:44 crc kubenswrapper[4802]: I1206 04:52:44.125873 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-9hxgw_e2001db0-90f1-40a7-ba6e-982eabebc117/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 04:52:44 crc kubenswrapper[4802]: I1206 04:52:44.201158 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_dca5fd9f-3aaf-4fd7-96d0-7f78c38c12b6/openstack-network-exporter/0.log" Dec 06 04:52:44 crc kubenswrapper[4802]: I1206 04:52:44.351356 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_dca5fd9f-3aaf-4fd7-96d0-7f78c38c12b6/ovn-northd/0.log" Dec 06 04:52:44 crc kubenswrapper[4802]: I1206 04:52:44.416480 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_56f9d065-29e4-424d-bc84-2657eba9929e/openstack-network-exporter/0.log" Dec 06 04:52:44 crc kubenswrapper[4802]: I1206 04:52:44.543246 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_56f9d065-29e4-424d-bc84-2657eba9929e/ovsdbserver-nb/0.log" Dec 06 04:52:44 crc kubenswrapper[4802]: I1206 04:52:44.634975 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_c45b4e0d-c52c-421b-b748-896ebc517784/openstack-network-exporter/0.log" Dec 06 04:52:44 crc kubenswrapper[4802]: I1206 04:52:44.650953 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_c45b4e0d-c52c-421b-b748-896ebc517784/ovsdbserver-sb/0.log" Dec 06 04:52:45 crc kubenswrapper[4802]: I1206 04:52:45.029170 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-788d784b-hjc7w_5800f6d2-346a-4e9a-8585-e839ddfd035f/placement-api/0.log" Dec 06 04:52:45 crc kubenswrapper[4802]: I1206 04:52:45.030111 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-788d784b-hjc7w_5800f6d2-346a-4e9a-8585-e839ddfd035f/placement-log/0.log" Dec 06 04:52:45 crc kubenswrapper[4802]: I1206 04:52:45.174470 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_cdcf8137-9bbe-467f-bbb3-e30f350760a6/setup-container/0.log" Dec 06 04:52:45 crc kubenswrapper[4802]: I1206 04:52:45.461710 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab924c08-8bf0-4f00-8798-866cbdfb04ba" path="/var/lib/kubelet/pods/ab924c08-8bf0-4f00-8798-866cbdfb04ba/volumes" Dec 06 04:52:45 crc kubenswrapper[4802]: I1206 04:52:45.657962 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_cdcf8137-9bbe-467f-bbb3-e30f350760a6/setup-container/0.log" Dec 06 04:52:45 crc kubenswrapper[4802]: I1206 04:52:45.711155 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_cdcf8137-9bbe-467f-bbb3-e30f350760a6/rabbitmq/0.log" Dec 06 04:52:45 crc kubenswrapper[4802]: I1206 04:52:45.717070 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_cdc37fb9-aeae-4fe0-b86e-e9d41e48314b/setup-container/0.log" Dec 06 04:52:45 crc kubenswrapper[4802]: I1206 04:52:45.899732 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_cdc37fb9-aeae-4fe0-b86e-e9d41e48314b/setup-container/0.log" Dec 06 04:52:45 crc kubenswrapper[4802]: I1206 04:52:45.966414 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-7hgqr_0de67972-73b4-4f71-aca7-d77ae62d73b3/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 04:52:45 crc kubenswrapper[4802]: I1206 04:52:45.994860 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_cdc37fb9-aeae-4fe0-b86e-e9d41e48314b/rabbitmq/0.log" Dec 06 04:52:46 crc kubenswrapper[4802]: I1206 04:52:46.154689 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-tgsk9_63d8761a-c5bb-41d5-a830-f58cc99b0837/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 04:52:46 crc kubenswrapper[4802]: I1206 04:52:46.297126 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-jrmv7_44aaebb1-a5a3-42e6-aacd-04292e80f33d/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 04:52:46 crc kubenswrapper[4802]: I1206 04:52:46.411385 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-4clcr_bcbbe04a-d1ab-4a9b-a4ff-11806f4928c8/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 04:52:46 crc kubenswrapper[4802]: I1206 04:52:46.531019 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-hgx7d_815e5d24-83cc-4bbe-b602-6cd1da82039b/ssh-known-hosts-edpm-deployment/0.log" Dec 06 04:52:46 crc kubenswrapper[4802]: I1206 04:52:46.765466 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-686fbdfd7f-sfrn5_4985762a-1d53-40ad-85da-c6728bdbaa3d/proxy-server/0.log" Dec 06 04:52:46 crc kubenswrapper[4802]: I1206 04:52:46.805231 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-686fbdfd7f-sfrn5_4985762a-1d53-40ad-85da-c6728bdbaa3d/proxy-httpd/0.log" Dec 06 04:52:47 crc kubenswrapper[4802]: I1206 04:52:47.553846 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_508d9d64-dd16-4d21-b492-052d123f1a6a/account-auditor/0.log" Dec 06 04:52:47 crc kubenswrapper[4802]: I1206 04:52:47.686277 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-flhmv_de4cab50-e92b-4282-bd9d-1b9439373aad/swift-ring-rebalance/0.log" Dec 06 04:52:47 crc kubenswrapper[4802]: I1206 04:52:47.732270 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_508d9d64-dd16-4d21-b492-052d123f1a6a/account-reaper/0.log" Dec 06 04:52:47 crc kubenswrapper[4802]: I1206 04:52:47.835472 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_508d9d64-dd16-4d21-b492-052d123f1a6a/account-replicator/0.log" Dec 06 04:52:47 crc kubenswrapper[4802]: I1206 04:52:47.927974 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_508d9d64-dd16-4d21-b492-052d123f1a6a/container-auditor/0.log" Dec 06 04:52:47 crc kubenswrapper[4802]: I1206 04:52:47.932420 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_508d9d64-dd16-4d21-b492-052d123f1a6a/account-server/0.log" Dec 06 04:52:48 crc kubenswrapper[4802]: I1206 04:52:48.019949 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_508d9d64-dd16-4d21-b492-052d123f1a6a/container-replicator/0.log" Dec 06 04:52:48 crc kubenswrapper[4802]: I1206 04:52:48.101912 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_508d9d64-dd16-4d21-b492-052d123f1a6a/container-server/0.log" Dec 06 04:52:48 crc kubenswrapper[4802]: I1206 04:52:48.133927 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_508d9d64-dd16-4d21-b492-052d123f1a6a/container-updater/0.log" Dec 06 04:52:48 crc kubenswrapper[4802]: I1206 04:52:48.191724 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_508d9d64-dd16-4d21-b492-052d123f1a6a/object-auditor/0.log" Dec 06 04:52:48 crc kubenswrapper[4802]: I1206 04:52:48.222140 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_508d9d64-dd16-4d21-b492-052d123f1a6a/object-expirer/0.log" Dec 06 04:52:48 crc kubenswrapper[4802]: I1206 04:52:48.297665 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_508d9d64-dd16-4d21-b492-052d123f1a6a/object-replicator/0.log" Dec 06 04:52:48 crc kubenswrapper[4802]: I1206 04:52:48.336256 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_508d9d64-dd16-4d21-b492-052d123f1a6a/object-server/0.log" Dec 06 04:52:48 crc kubenswrapper[4802]: I1206 04:52:48.376955 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_508d9d64-dd16-4d21-b492-052d123f1a6a/object-updater/0.log" Dec 06 04:52:48 crc kubenswrapper[4802]: I1206 04:52:48.430497 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_508d9d64-dd16-4d21-b492-052d123f1a6a/rsync/0.log" Dec 06 04:52:48 crc kubenswrapper[4802]: I1206 04:52:48.506200 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_508d9d64-dd16-4d21-b492-052d123f1a6a/swift-recon-cron/0.log" Dec 06 04:52:48 crc kubenswrapper[4802]: I1206 04:52:48.668513 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-ns78h_565b62ff-4b9a-4e53-b61c-d6c492d7b253/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 04:52:48 crc kubenswrapper[4802]: I1206 04:52:48.759306 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_63c9a1aa-a443-4071-b60e-317a7b81395e/tempest-tests-tempest-tests-runner/0.log" Dec 06 04:52:48 crc kubenswrapper[4802]: I1206 04:52:48.884500 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_35caa18c-c7c4-4e74-9c6d-053c38327afc/test-operator-logs-container/0.log" Dec 06 04:52:48 crc kubenswrapper[4802]: I1206 04:52:48.983790 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-7nbhk_5fd4f8df-59a8-47cb-b9e7-4162a5811ed7/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 04:52:53 crc kubenswrapper[4802]: I1206 04:52:53.498365 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-vqlpv" Dec 06 04:52:53 crc kubenswrapper[4802]: I1206 04:52:53.564444 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vqlpv"] Dec 06 04:52:53 crc kubenswrapper[4802]: I1206 04:52:53.794056 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-vqlpv" podUID="0a7427d8-b298-4e7f-bcd4-b36d8787aa07" containerName="registry-server" containerID="cri-o://017267762f1784908d85c41c76a882a83ebc60cf7772f1eb9aa2341467d45ba4" gracePeriod=2 Dec 06 04:52:54 crc kubenswrapper[4802]: I1206 04:52:54.450075 4802 scope.go:117] "RemoveContainer" containerID="4334df6618bec8f6cedfb779435b2c4d4e0f7e149dd426e245ab29892f475208" Dec 06 04:52:54 crc kubenswrapper[4802]: E1206 04:52:54.450776 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:52:54 crc kubenswrapper[4802]: I1206 04:52:54.580176 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vqlpv" Dec 06 04:52:54 crc kubenswrapper[4802]: I1206 04:52:54.755554 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a7427d8-b298-4e7f-bcd4-b36d8787aa07-catalog-content\") pod \"0a7427d8-b298-4e7f-bcd4-b36d8787aa07\" (UID: \"0a7427d8-b298-4e7f-bcd4-b36d8787aa07\") " Dec 06 04:52:54 crc kubenswrapper[4802]: I1206 04:52:54.755602 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a7427d8-b298-4e7f-bcd4-b36d8787aa07-utilities\") pod \"0a7427d8-b298-4e7f-bcd4-b36d8787aa07\" (UID: \"0a7427d8-b298-4e7f-bcd4-b36d8787aa07\") " Dec 06 04:52:54 crc kubenswrapper[4802]: I1206 04:52:54.755759 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q6c8r\" (UniqueName: \"kubernetes.io/projected/0a7427d8-b298-4e7f-bcd4-b36d8787aa07-kube-api-access-q6c8r\") pod \"0a7427d8-b298-4e7f-bcd4-b36d8787aa07\" (UID: \"0a7427d8-b298-4e7f-bcd4-b36d8787aa07\") " Dec 06 04:52:54 crc kubenswrapper[4802]: I1206 04:52:54.758619 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a7427d8-b298-4e7f-bcd4-b36d8787aa07-utilities" (OuterVolumeSpecName: "utilities") pod "0a7427d8-b298-4e7f-bcd4-b36d8787aa07" (UID: "0a7427d8-b298-4e7f-bcd4-b36d8787aa07"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:52:54 crc kubenswrapper[4802]: I1206 04:52:54.806350 4802 generic.go:334] "Generic (PLEG): container finished" podID="0a7427d8-b298-4e7f-bcd4-b36d8787aa07" containerID="017267762f1784908d85c41c76a882a83ebc60cf7772f1eb9aa2341467d45ba4" exitCode=0 Dec 06 04:52:54 crc kubenswrapper[4802]: I1206 04:52:54.806427 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vqlpv" event={"ID":"0a7427d8-b298-4e7f-bcd4-b36d8787aa07","Type":"ContainerDied","Data":"017267762f1784908d85c41c76a882a83ebc60cf7772f1eb9aa2341467d45ba4"} Dec 06 04:52:54 crc kubenswrapper[4802]: I1206 04:52:54.806470 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vqlpv" event={"ID":"0a7427d8-b298-4e7f-bcd4-b36d8787aa07","Type":"ContainerDied","Data":"fe4a65d66b51c4128c0459a5becc02de78a074363d556de635c9f291cdd3c29a"} Dec 06 04:52:54 crc kubenswrapper[4802]: I1206 04:52:54.806621 4802 scope.go:117] "RemoveContainer" containerID="017267762f1784908d85c41c76a882a83ebc60cf7772f1eb9aa2341467d45ba4" Dec 06 04:52:54 crc kubenswrapper[4802]: I1206 04:52:54.807214 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vqlpv" Dec 06 04:52:54 crc kubenswrapper[4802]: I1206 04:52:54.814065 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a7427d8-b298-4e7f-bcd4-b36d8787aa07-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0a7427d8-b298-4e7f-bcd4-b36d8787aa07" (UID: "0a7427d8-b298-4e7f-bcd4-b36d8787aa07"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:52:54 crc kubenswrapper[4802]: I1206 04:52:54.838893 4802 scope.go:117] "RemoveContainer" containerID="39c805d0a9518680afbdadb58d3f3a0ecdbad13d6a19fee7df3e0ec65f6eb7fc" Dec 06 04:52:54 crc kubenswrapper[4802]: I1206 04:52:54.857674 4802 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a7427d8-b298-4e7f-bcd4-b36d8787aa07-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 04:52:54 crc kubenswrapper[4802]: I1206 04:52:54.857713 4802 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a7427d8-b298-4e7f-bcd4-b36d8787aa07-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 04:52:55 crc kubenswrapper[4802]: I1206 04:52:55.259061 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a7427d8-b298-4e7f-bcd4-b36d8787aa07-kube-api-access-q6c8r" (OuterVolumeSpecName: "kube-api-access-q6c8r") pod "0a7427d8-b298-4e7f-bcd4-b36d8787aa07" (UID: "0a7427d8-b298-4e7f-bcd4-b36d8787aa07"). InnerVolumeSpecName "kube-api-access-q6c8r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:52:55 crc kubenswrapper[4802]: I1206 04:52:55.264049 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q6c8r\" (UniqueName: \"kubernetes.io/projected/0a7427d8-b298-4e7f-bcd4-b36d8787aa07-kube-api-access-q6c8r\") on node \"crc\" DevicePath \"\"" Dec 06 04:52:55 crc kubenswrapper[4802]: I1206 04:52:55.272987 4802 scope.go:117] "RemoveContainer" containerID="0789dadd2c8066bdf1a99439432fc812a6148f5555b35935e00af23abdc01804" Dec 06 04:52:55 crc kubenswrapper[4802]: I1206 04:52:55.357775 4802 scope.go:117] "RemoveContainer" containerID="017267762f1784908d85c41c76a882a83ebc60cf7772f1eb9aa2341467d45ba4" Dec 06 04:52:55 crc kubenswrapper[4802]: E1206 04:52:55.358188 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"017267762f1784908d85c41c76a882a83ebc60cf7772f1eb9aa2341467d45ba4\": container with ID starting with 017267762f1784908d85c41c76a882a83ebc60cf7772f1eb9aa2341467d45ba4 not found: ID does not exist" containerID="017267762f1784908d85c41c76a882a83ebc60cf7772f1eb9aa2341467d45ba4" Dec 06 04:52:55 crc kubenswrapper[4802]: I1206 04:52:55.358219 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"017267762f1784908d85c41c76a882a83ebc60cf7772f1eb9aa2341467d45ba4"} err="failed to get container status \"017267762f1784908d85c41c76a882a83ebc60cf7772f1eb9aa2341467d45ba4\": rpc error: code = NotFound desc = could not find container \"017267762f1784908d85c41c76a882a83ebc60cf7772f1eb9aa2341467d45ba4\": container with ID starting with 017267762f1784908d85c41c76a882a83ebc60cf7772f1eb9aa2341467d45ba4 not found: ID does not exist" Dec 06 04:52:55 crc kubenswrapper[4802]: I1206 04:52:55.358238 4802 scope.go:117] "RemoveContainer" containerID="39c805d0a9518680afbdadb58d3f3a0ecdbad13d6a19fee7df3e0ec65f6eb7fc" Dec 06 04:52:55 crc kubenswrapper[4802]: E1206 04:52:55.358538 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39c805d0a9518680afbdadb58d3f3a0ecdbad13d6a19fee7df3e0ec65f6eb7fc\": container with ID starting with 39c805d0a9518680afbdadb58d3f3a0ecdbad13d6a19fee7df3e0ec65f6eb7fc not found: ID does not exist" containerID="39c805d0a9518680afbdadb58d3f3a0ecdbad13d6a19fee7df3e0ec65f6eb7fc" Dec 06 04:52:55 crc kubenswrapper[4802]: I1206 04:52:55.358575 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39c805d0a9518680afbdadb58d3f3a0ecdbad13d6a19fee7df3e0ec65f6eb7fc"} err="failed to get container status \"39c805d0a9518680afbdadb58d3f3a0ecdbad13d6a19fee7df3e0ec65f6eb7fc\": rpc error: code = NotFound desc = could not find container \"39c805d0a9518680afbdadb58d3f3a0ecdbad13d6a19fee7df3e0ec65f6eb7fc\": container with ID starting with 39c805d0a9518680afbdadb58d3f3a0ecdbad13d6a19fee7df3e0ec65f6eb7fc not found: ID does not exist" Dec 06 04:52:55 crc kubenswrapper[4802]: I1206 04:52:55.358601 4802 scope.go:117] "RemoveContainer" containerID="0789dadd2c8066bdf1a99439432fc812a6148f5555b35935e00af23abdc01804" Dec 06 04:52:55 crc kubenswrapper[4802]: E1206 04:52:55.359022 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0789dadd2c8066bdf1a99439432fc812a6148f5555b35935e00af23abdc01804\": container with ID starting with 0789dadd2c8066bdf1a99439432fc812a6148f5555b35935e00af23abdc01804 not found: ID does not exist" containerID="0789dadd2c8066bdf1a99439432fc812a6148f5555b35935e00af23abdc01804" Dec 06 04:52:55 crc kubenswrapper[4802]: I1206 04:52:55.359044 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0789dadd2c8066bdf1a99439432fc812a6148f5555b35935e00af23abdc01804"} err="failed to get container status \"0789dadd2c8066bdf1a99439432fc812a6148f5555b35935e00af23abdc01804\": rpc error: code = NotFound desc = could not find container \"0789dadd2c8066bdf1a99439432fc812a6148f5555b35935e00af23abdc01804\": container with ID starting with 0789dadd2c8066bdf1a99439432fc812a6148f5555b35935e00af23abdc01804 not found: ID does not exist" Dec 06 04:52:55 crc kubenswrapper[4802]: I1206 04:52:55.447793 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vqlpv"] Dec 06 04:52:55 crc kubenswrapper[4802]: I1206 04:52:55.463268 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-vqlpv"] Dec 06 04:52:57 crc kubenswrapper[4802]: I1206 04:52:57.152113 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_654f214b-8467-41d2-93f4-3c18377653b9/memcached/0.log" Dec 06 04:52:57 crc kubenswrapper[4802]: I1206 04:52:57.472623 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a7427d8-b298-4e7f-bcd4-b36d8787aa07" path="/var/lib/kubelet/pods/0a7427d8-b298-4e7f-bcd4-b36d8787aa07/volumes" Dec 06 04:53:09 crc kubenswrapper[4802]: I1206 04:53:09.451101 4802 scope.go:117] "RemoveContainer" containerID="4334df6618bec8f6cedfb779435b2c4d4e0f7e149dd426e245ab29892f475208" Dec 06 04:53:09 crc kubenswrapper[4802]: E1206 04:53:09.451844 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:53:15 crc kubenswrapper[4802]: I1206 04:53:15.281252 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_11a37daaca2b5d044ad5e897ef78a9b78d5fd0aec8475f6e59b41a4af1nhddl_e9a7628c-dba2-482e-b66b-e4046bb06ddc/util/0.log" Dec 06 04:53:15 crc kubenswrapper[4802]: I1206 04:53:15.819020 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_11a37daaca2b5d044ad5e897ef78a9b78d5fd0aec8475f6e59b41a4af1nhddl_e9a7628c-dba2-482e-b66b-e4046bb06ddc/util/0.log" Dec 06 04:53:15 crc kubenswrapper[4802]: I1206 04:53:15.847219 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_11a37daaca2b5d044ad5e897ef78a9b78d5fd0aec8475f6e59b41a4af1nhddl_e9a7628c-dba2-482e-b66b-e4046bb06ddc/pull/0.log" Dec 06 04:53:15 crc kubenswrapper[4802]: I1206 04:53:15.857149 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_11a37daaca2b5d044ad5e897ef78a9b78d5fd0aec8475f6e59b41a4af1nhddl_e9a7628c-dba2-482e-b66b-e4046bb06ddc/pull/0.log" Dec 06 04:53:16 crc kubenswrapper[4802]: I1206 04:53:16.010126 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_11a37daaca2b5d044ad5e897ef78a9b78d5fd0aec8475f6e59b41a4af1nhddl_e9a7628c-dba2-482e-b66b-e4046bb06ddc/extract/0.log" Dec 06 04:53:16 crc kubenswrapper[4802]: I1206 04:53:16.023336 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_11a37daaca2b5d044ad5e897ef78a9b78d5fd0aec8475f6e59b41a4af1nhddl_e9a7628c-dba2-482e-b66b-e4046bb06ddc/pull/0.log" Dec 06 04:53:16 crc kubenswrapper[4802]: I1206 04:53:16.032099 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_11a37daaca2b5d044ad5e897ef78a9b78d5fd0aec8475f6e59b41a4af1nhddl_e9a7628c-dba2-482e-b66b-e4046bb06ddc/util/0.log" Dec 06 04:53:16 crc kubenswrapper[4802]: I1206 04:53:16.175232 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-p8kv6_1c3d06b1-b999-4c82-85b5-11bdd996770c/kube-rbac-proxy/0.log" Dec 06 04:53:16 crc kubenswrapper[4802]: I1206 04:53:16.279043 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-p8kv6_1c3d06b1-b999-4c82-85b5-11bdd996770c/manager/0.log" Dec 06 04:53:16 crc kubenswrapper[4802]: I1206 04:53:16.304735 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-xk9cb_309f12a2-5b02-4d5f-9e28-63116ff8bb47/kube-rbac-proxy/0.log" Dec 06 04:53:16 crc kubenswrapper[4802]: I1206 04:53:16.418985 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-xk9cb_309f12a2-5b02-4d5f-9e28-63116ff8bb47/manager/0.log" Dec 06 04:53:16 crc kubenswrapper[4802]: I1206 04:53:16.496846 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-wwgt6_e2bc662f-efe3-4e89-9b0e-e40eba87e41b/manager/0.log" Dec 06 04:53:16 crc kubenswrapper[4802]: I1206 04:53:16.524377 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-wwgt6_e2bc662f-efe3-4e89-9b0e-e40eba87e41b/kube-rbac-proxy/0.log" Dec 06 04:53:16 crc kubenswrapper[4802]: I1206 04:53:16.678418 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-zkp5c_08b2817d-5061-4375-82d4-726108586852/kube-rbac-proxy/0.log" Dec 06 04:53:16 crc kubenswrapper[4802]: I1206 04:53:16.743913 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-zkp5c_08b2817d-5061-4375-82d4-726108586852/manager/0.log" Dec 06 04:53:16 crc kubenswrapper[4802]: I1206 04:53:16.813165 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-55cb9d4b9c-srntd_faf6c2c3-b028-431f-a302-3ad181dd93ae/kube-rbac-proxy/0.log" Dec 06 04:53:16 crc kubenswrapper[4802]: I1206 04:53:16.911475 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-55cb9d4b9c-srntd_faf6c2c3-b028-431f-a302-3ad181dd93ae/manager/0.log" Dec 06 04:53:17 crc kubenswrapper[4802]: I1206 04:53:17.431920 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-gf27t_5f8e1071-aa69-4e21-94c9-f533d55f2cac/manager/0.log" Dec 06 04:53:17 crc kubenswrapper[4802]: I1206 04:53:17.472351 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-gf27t_5f8e1071-aa69-4e21-94c9-f533d55f2cac/kube-rbac-proxy/0.log" Dec 06 04:53:17 crc kubenswrapper[4802]: I1206 04:53:17.607998 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-hpsfv_19f0c68c-32a3-41e2-a8ca-5ab0dc28d748/kube-rbac-proxy/0.log" Dec 06 04:53:17 crc kubenswrapper[4802]: I1206 04:53:17.695826 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-c4sr8_d71b7639-ee30-4323-8be0-a16d844d259f/kube-rbac-proxy/0.log" Dec 06 04:53:17 crc kubenswrapper[4802]: I1206 04:53:17.861720 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-hpsfv_19f0c68c-32a3-41e2-a8ca-5ab0dc28d748/manager/0.log" Dec 06 04:53:17 crc kubenswrapper[4802]: I1206 04:53:17.895308 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-h7629_c2e52d5b-b8cf-41f2-933e-af31359dc69d/kube-rbac-proxy/0.log" Dec 06 04:53:17 crc kubenswrapper[4802]: I1206 04:53:17.897145 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-c4sr8_d71b7639-ee30-4323-8be0-a16d844d259f/manager/0.log" Dec 06 04:53:18 crc kubenswrapper[4802]: I1206 04:53:18.079282 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-h7629_c2e52d5b-b8cf-41f2-933e-af31359dc69d/manager/0.log" Dec 06 04:53:18 crc kubenswrapper[4802]: I1206 04:53:18.097586 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-74c6z_1d3c0479-80bf-46af-a261-42b3834807f1/kube-rbac-proxy/0.log" Dec 06 04:53:18 crc kubenswrapper[4802]: I1206 04:53:18.117864 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-74c6z_1d3c0479-80bf-46af-a261-42b3834807f1/manager/0.log" Dec 06 04:53:18 crc kubenswrapper[4802]: I1206 04:53:18.291301 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-bkfc6_4460e15d-1d9a-4814-a7d2-042c6a84057f/kube-rbac-proxy/0.log" Dec 06 04:53:18 crc kubenswrapper[4802]: I1206 04:53:18.313647 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-bkfc6_4460e15d-1d9a-4814-a7d2-042c6a84057f/manager/0.log" Dec 06 04:53:18 crc kubenswrapper[4802]: I1206 04:53:18.464472 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-hlx24_a076799b-6631-474a-b552-9cf2261c377a/kube-rbac-proxy/0.log" Dec 06 04:53:18 crc kubenswrapper[4802]: I1206 04:53:18.546138 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-4m2vm_ef258927-21c4-4147-8fb5-3db5d78658a3/kube-rbac-proxy/0.log" Dec 06 04:53:18 crc kubenswrapper[4802]: I1206 04:53:18.563023 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-hlx24_a076799b-6631-474a-b552-9cf2261c377a/manager/0.log" Dec 06 04:53:18 crc kubenswrapper[4802]: I1206 04:53:18.727912 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-4m2vm_ef258927-21c4-4147-8fb5-3db5d78658a3/manager/0.log" Dec 06 04:53:18 crc kubenswrapper[4802]: I1206 04:53:18.745796 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-lk8bf_8555a40e-f3d3-4627-8867-6db5ca5ed06a/kube-rbac-proxy/0.log" Dec 06 04:53:18 crc kubenswrapper[4802]: I1206 04:53:18.765628 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-lk8bf_8555a40e-f3d3-4627-8867-6db5ca5ed06a/manager/0.log" Dec 06 04:53:18 crc kubenswrapper[4802]: I1206 04:53:18.890072 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4pwn82_05736c59-c4db-4cfe-addd-1b29c552596e/kube-rbac-proxy/0.log" Dec 06 04:53:18 crc kubenswrapper[4802]: I1206 04:53:18.922037 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4pwn82_05736c59-c4db-4cfe-addd-1b29c552596e/manager/0.log" Dec 06 04:53:19 crc kubenswrapper[4802]: I1206 04:53:19.210256 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-57d4dc878b-qnwf9_31962197-d27b-41af-9469-27ae39fb5d4f/operator/0.log" Dec 06 04:53:19 crc kubenswrapper[4802]: I1206 04:53:19.253550 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-xd8sc_43e371ec-a36d-4082-827e-21518dc7e9c1/registry-server/0.log" Dec 06 04:53:19 crc kubenswrapper[4802]: I1206 04:53:19.404382 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-8q2cg_119bce90-0fad-424d-89cf-2ca06d7126ae/kube-rbac-proxy/0.log" Dec 06 04:53:19 crc kubenswrapper[4802]: I1206 04:53:19.506978 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-8q2cg_119bce90-0fad-424d-89cf-2ca06d7126ae/manager/0.log" Dec 06 04:53:19 crc kubenswrapper[4802]: I1206 04:53:19.633532 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-4x9rb_79efe004-97e7-4586-b48e-08ce9285bb79/kube-rbac-proxy/0.log" Dec 06 04:53:19 crc kubenswrapper[4802]: I1206 04:53:19.706591 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-4x9rb_79efe004-97e7-4586-b48e-08ce9285bb79/manager/0.log" Dec 06 04:53:19 crc kubenswrapper[4802]: I1206 04:53:19.807003 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-b5wv6_a3bed886-a616-4733-947a-288ada88dc74/operator/0.log" Dec 06 04:53:19 crc kubenswrapper[4802]: I1206 04:53:19.946693 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-mmh24_11885e74-7bc3-4207-b046-9e94ab51e499/kube-rbac-proxy/0.log" Dec 06 04:53:20 crc kubenswrapper[4802]: I1206 04:53:20.038684 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-mmh24_11885e74-7bc3-4207-b046-9e94ab51e499/manager/0.log" Dec 06 04:53:20 crc kubenswrapper[4802]: I1206 04:53:20.105183 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-nd42v_18ad32ab-222c-41f9-8409-90fd51178f2d/kube-rbac-proxy/0.log" Dec 06 04:53:20 crc kubenswrapper[4802]: I1206 04:53:20.232840 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-nd42v_18ad32ab-222c-41f9-8409-90fd51178f2d/manager/0.log" Dec 06 04:53:20 crc kubenswrapper[4802]: I1206 04:53:20.238126 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-77f4d6cd-9s876_5f6a5e2f-8db1-4d31-8a2d-5810c5e62083/manager/0.log" Dec 06 04:53:20 crc kubenswrapper[4802]: I1206 04:53:20.312188 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-q6k7m_e9a14170-cafd-433c-83a4-22793b21d376/kube-rbac-proxy/0.log" Dec 06 04:53:20 crc kubenswrapper[4802]: I1206 04:53:20.350225 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-q6k7m_e9a14170-cafd-433c-83a4-22793b21d376/manager/0.log" Dec 06 04:53:20 crc kubenswrapper[4802]: I1206 04:53:20.392245 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-w64dr_91b87773-2f3d-4e35-9a13-708f07cb2bbf/kube-rbac-proxy/0.log" Dec 06 04:53:20 crc kubenswrapper[4802]: I1206 04:53:20.450349 4802 scope.go:117] "RemoveContainer" containerID="4334df6618bec8f6cedfb779435b2c4d4e0f7e149dd426e245ab29892f475208" Dec 06 04:53:20 crc kubenswrapper[4802]: E1206 04:53:20.450640 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:53:20 crc kubenswrapper[4802]: I1206 04:53:20.475191 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-w64dr_91b87773-2f3d-4e35-9a13-708f07cb2bbf/manager/0.log" Dec 06 04:53:31 crc kubenswrapper[4802]: I1206 04:53:31.450691 4802 scope.go:117] "RemoveContainer" containerID="4334df6618bec8f6cedfb779435b2c4d4e0f7e149dd426e245ab29892f475208" Dec 06 04:53:31 crc kubenswrapper[4802]: E1206 04:53:31.451499 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:53:38 crc kubenswrapper[4802]: I1206 04:53:38.340088 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-6d29n_1a34c64b-194e-41a6-8b55-0a87187c4cfc/control-plane-machine-set-operator/0.log" Dec 06 04:53:38 crc kubenswrapper[4802]: I1206 04:53:38.529506 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-9s5r8_eb15a464-d0e8-4e49-b018-14353cc9d6cb/kube-rbac-proxy/0.log" Dec 06 04:53:38 crc kubenswrapper[4802]: I1206 04:53:38.529908 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-9s5r8_eb15a464-d0e8-4e49-b018-14353cc9d6cb/machine-api-operator/0.log" Dec 06 04:53:42 crc kubenswrapper[4802]: I1206 04:53:42.450357 4802 scope.go:117] "RemoveContainer" containerID="4334df6618bec8f6cedfb779435b2c4d4e0f7e149dd426e245ab29892f475208" Dec 06 04:53:42 crc kubenswrapper[4802]: E1206 04:53:42.451122 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:53:49 crc kubenswrapper[4802]: I1206 04:53:49.963694 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-sdbpz_3f3aa4aa-3196-4dd8-802f-73235899e452/cert-manager-controller/0.log" Dec 06 04:53:50 crc kubenswrapper[4802]: I1206 04:53:50.120726 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-7r6gm_5aaa82f6-dd8a-4ca3-8e3d-b574617c240a/cert-manager-cainjector/0.log" Dec 06 04:53:50 crc kubenswrapper[4802]: I1206 04:53:50.159964 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-54qqk_a50fe25c-7be9-42ea-a81b-e6ba817043e8/cert-manager-webhook/0.log" Dec 06 04:53:56 crc kubenswrapper[4802]: I1206 04:53:56.450662 4802 scope.go:117] "RemoveContainer" containerID="4334df6618bec8f6cedfb779435b2c4d4e0f7e149dd426e245ab29892f475208" Dec 06 04:53:56 crc kubenswrapper[4802]: E1206 04:53:56.451954 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:54:02 crc kubenswrapper[4802]: I1206 04:54:02.294768 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-8wwln_db89a256-eef7-476e-ab22-755b4d6bb780/nmstate-console-plugin/0.log" Dec 06 04:54:02 crc kubenswrapper[4802]: I1206 04:54:02.466262 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-swx72_485510d8-9971-4648-bb7c-120875a7c00d/nmstate-handler/0.log" Dec 06 04:54:02 crc kubenswrapper[4802]: I1206 04:54:02.536341 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-d8q2s_75129a1d-9a3e-4efb-b2bf-bf353a4f6ea2/kube-rbac-proxy/0.log" Dec 06 04:54:02 crc kubenswrapper[4802]: I1206 04:54:02.611314 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-d8q2s_75129a1d-9a3e-4efb-b2bf-bf353a4f6ea2/nmstate-metrics/0.log" Dec 06 04:54:02 crc kubenswrapper[4802]: I1206 04:54:02.684549 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-l8t5h_a1e67ef9-99a6-4d92-a5d4-6c00837b8993/nmstate-operator/0.log" Dec 06 04:54:02 crc kubenswrapper[4802]: I1206 04:54:02.803701 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-2trjk_92122afa-11ef-42cf-9b97-86d3a41c6e13/nmstate-webhook/0.log" Dec 06 04:54:11 crc kubenswrapper[4802]: I1206 04:54:11.450550 4802 scope.go:117] "RemoveContainer" containerID="4334df6618bec8f6cedfb779435b2c4d4e0f7e149dd426e245ab29892f475208" Dec 06 04:54:11 crc kubenswrapper[4802]: E1206 04:54:11.451377 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:54:17 crc kubenswrapper[4802]: I1206 04:54:17.128410 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-46vm8_3a078497-d2a1-4be5-87ee-ba151c992ecd/kube-rbac-proxy/0.log" Dec 06 04:54:17 crc kubenswrapper[4802]: I1206 04:54:17.275592 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-46vm8_3a078497-d2a1-4be5-87ee-ba151c992ecd/controller/0.log" Dec 06 04:54:17 crc kubenswrapper[4802]: I1206 04:54:17.342514 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nzpzt_9ff1fa79-a414-498e-9f64-659502eb6ae7/cp-frr-files/0.log" Dec 06 04:54:18 crc kubenswrapper[4802]: I1206 04:54:18.143829 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nzpzt_9ff1fa79-a414-498e-9f64-659502eb6ae7/cp-reloader/0.log" Dec 06 04:54:18 crc kubenswrapper[4802]: I1206 04:54:18.146933 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nzpzt_9ff1fa79-a414-498e-9f64-659502eb6ae7/cp-frr-files/0.log" Dec 06 04:54:18 crc kubenswrapper[4802]: I1206 04:54:18.164856 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nzpzt_9ff1fa79-a414-498e-9f64-659502eb6ae7/cp-reloader/0.log" Dec 06 04:54:18 crc kubenswrapper[4802]: I1206 04:54:18.193074 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nzpzt_9ff1fa79-a414-498e-9f64-659502eb6ae7/cp-metrics/0.log" Dec 06 04:54:18 crc kubenswrapper[4802]: I1206 04:54:18.331436 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nzpzt_9ff1fa79-a414-498e-9f64-659502eb6ae7/cp-metrics/0.log" Dec 06 04:54:18 crc kubenswrapper[4802]: I1206 04:54:18.337100 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nzpzt_9ff1fa79-a414-498e-9f64-659502eb6ae7/cp-frr-files/0.log" Dec 06 04:54:18 crc kubenswrapper[4802]: I1206 04:54:18.374591 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nzpzt_9ff1fa79-a414-498e-9f64-659502eb6ae7/cp-metrics/0.log" Dec 06 04:54:18 crc kubenswrapper[4802]: I1206 04:54:18.388260 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nzpzt_9ff1fa79-a414-498e-9f64-659502eb6ae7/cp-reloader/0.log" Dec 06 04:54:18 crc kubenswrapper[4802]: I1206 04:54:18.532372 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nzpzt_9ff1fa79-a414-498e-9f64-659502eb6ae7/cp-frr-files/0.log" Dec 06 04:54:18 crc kubenswrapper[4802]: I1206 04:54:18.557885 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nzpzt_9ff1fa79-a414-498e-9f64-659502eb6ae7/cp-reloader/0.log" Dec 06 04:54:18 crc kubenswrapper[4802]: I1206 04:54:18.567894 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nzpzt_9ff1fa79-a414-498e-9f64-659502eb6ae7/cp-metrics/0.log" Dec 06 04:54:18 crc kubenswrapper[4802]: I1206 04:54:18.605386 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nzpzt_9ff1fa79-a414-498e-9f64-659502eb6ae7/controller/0.log" Dec 06 04:54:18 crc kubenswrapper[4802]: I1206 04:54:18.705649 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nzpzt_9ff1fa79-a414-498e-9f64-659502eb6ae7/frr-metrics/0.log" Dec 06 04:54:18 crc kubenswrapper[4802]: I1206 04:54:18.725463 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nzpzt_9ff1fa79-a414-498e-9f64-659502eb6ae7/kube-rbac-proxy/0.log" Dec 06 04:54:18 crc kubenswrapper[4802]: I1206 04:54:18.809946 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nzpzt_9ff1fa79-a414-498e-9f64-659502eb6ae7/kube-rbac-proxy-frr/0.log" Dec 06 04:54:18 crc kubenswrapper[4802]: I1206 04:54:18.914360 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nzpzt_9ff1fa79-a414-498e-9f64-659502eb6ae7/reloader/0.log" Dec 06 04:54:19 crc kubenswrapper[4802]: I1206 04:54:19.005884 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-bjbtx_94e7ed3e-c72c-4130-81e6-ef954e2e62dd/frr-k8s-webhook-server/0.log" Dec 06 04:54:19 crc kubenswrapper[4802]: I1206 04:54:19.210471 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5c78fd4b89-5m46c_392ad0d8-35bb-40df-abf2-7a98117f24f7/manager/0.log" Dec 06 04:54:19 crc kubenswrapper[4802]: I1206 04:54:19.326607 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-868c9d9dfb-lxrtk_9b4b22b1-625a-4bd0-9c82-a142a6f7f806/webhook-server/0.log" Dec 06 04:54:19 crc kubenswrapper[4802]: I1206 04:54:19.432983 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-rlwm5_1e43b1ba-4d5f-47a4-8f9c-359875710a1e/kube-rbac-proxy/0.log" Dec 06 04:54:19 crc kubenswrapper[4802]: I1206 04:54:19.986060 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-rlwm5_1e43b1ba-4d5f-47a4-8f9c-359875710a1e/speaker/0.log" Dec 06 04:54:20 crc kubenswrapper[4802]: I1206 04:54:20.356397 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nzpzt_9ff1fa79-a414-498e-9f64-659502eb6ae7/frr/0.log" Dec 06 04:54:24 crc kubenswrapper[4802]: I1206 04:54:24.450972 4802 scope.go:117] "RemoveContainer" containerID="4334df6618bec8f6cedfb779435b2c4d4e0f7e149dd426e245ab29892f475208" Dec 06 04:54:24 crc kubenswrapper[4802]: E1206 04:54:24.451889 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:54:31 crc kubenswrapper[4802]: I1206 04:54:31.815979 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkz86h_7dfbb06c-f867-4d00-aef7-e731fa8579a7/util/0.log" Dec 06 04:54:32 crc kubenswrapper[4802]: I1206 04:54:32.024773 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkz86h_7dfbb06c-f867-4d00-aef7-e731fa8579a7/util/0.log" Dec 06 04:54:32 crc kubenswrapper[4802]: I1206 04:54:32.090849 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkz86h_7dfbb06c-f867-4d00-aef7-e731fa8579a7/pull/0.log" Dec 06 04:54:32 crc kubenswrapper[4802]: I1206 04:54:32.098775 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkz86h_7dfbb06c-f867-4d00-aef7-e731fa8579a7/pull/0.log" Dec 06 04:54:32 crc kubenswrapper[4802]: I1206 04:54:32.270425 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkz86h_7dfbb06c-f867-4d00-aef7-e731fa8579a7/util/0.log" Dec 06 04:54:32 crc kubenswrapper[4802]: I1206 04:54:32.283074 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkz86h_7dfbb06c-f867-4d00-aef7-e731fa8579a7/extract/0.log" Dec 06 04:54:32 crc kubenswrapper[4802]: I1206 04:54:32.504534 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkz86h_7dfbb06c-f867-4d00-aef7-e731fa8579a7/pull/0.log" Dec 06 04:54:32 crc kubenswrapper[4802]: I1206 04:54:32.665223 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838j9dk_207dc360-1c22-429b-b49e-f27bdd02bc03/util/0.log" Dec 06 04:54:32 crc kubenswrapper[4802]: I1206 04:54:32.797796 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838j9dk_207dc360-1c22-429b-b49e-f27bdd02bc03/util/0.log" Dec 06 04:54:32 crc kubenswrapper[4802]: I1206 04:54:32.831478 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838j9dk_207dc360-1c22-429b-b49e-f27bdd02bc03/pull/0.log" Dec 06 04:54:32 crc kubenswrapper[4802]: I1206 04:54:32.845368 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838j9dk_207dc360-1c22-429b-b49e-f27bdd02bc03/pull/0.log" Dec 06 04:54:33 crc kubenswrapper[4802]: I1206 04:54:33.021021 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838j9dk_207dc360-1c22-429b-b49e-f27bdd02bc03/util/0.log" Dec 06 04:54:33 crc kubenswrapper[4802]: I1206 04:54:33.026222 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838j9dk_207dc360-1c22-429b-b49e-f27bdd02bc03/extract/0.log" Dec 06 04:54:33 crc kubenswrapper[4802]: I1206 04:54:33.027324 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838j9dk_207dc360-1c22-429b-b49e-f27bdd02bc03/pull/0.log" Dec 06 04:54:33 crc kubenswrapper[4802]: I1206 04:54:33.239318 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-skt5b_042f382d-836e-4cc2-9065-352a210f6593/extract-utilities/0.log" Dec 06 04:54:33 crc kubenswrapper[4802]: I1206 04:54:33.418334 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-skt5b_042f382d-836e-4cc2-9065-352a210f6593/extract-content/0.log" Dec 06 04:54:33 crc kubenswrapper[4802]: I1206 04:54:33.435507 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-skt5b_042f382d-836e-4cc2-9065-352a210f6593/extract-utilities/0.log" Dec 06 04:54:33 crc kubenswrapper[4802]: I1206 04:54:33.458933 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-skt5b_042f382d-836e-4cc2-9065-352a210f6593/extract-content/0.log" Dec 06 04:54:33 crc kubenswrapper[4802]: I1206 04:54:33.622562 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-skt5b_042f382d-836e-4cc2-9065-352a210f6593/extract-content/0.log" Dec 06 04:54:33 crc kubenswrapper[4802]: I1206 04:54:33.669662 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-skt5b_042f382d-836e-4cc2-9065-352a210f6593/extract-utilities/0.log" Dec 06 04:54:33 crc kubenswrapper[4802]: I1206 04:54:33.860625 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rnrvp_cae8cfbe-5644-45a5-92e6-0ce169cc47c5/extract-utilities/0.log" Dec 06 04:54:34 crc kubenswrapper[4802]: I1206 04:54:34.116357 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-skt5b_042f382d-836e-4cc2-9065-352a210f6593/registry-server/0.log" Dec 06 04:54:34 crc kubenswrapper[4802]: I1206 04:54:34.169555 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rnrvp_cae8cfbe-5644-45a5-92e6-0ce169cc47c5/extract-utilities/0.log" Dec 06 04:54:34 crc kubenswrapper[4802]: I1206 04:54:34.181892 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rnrvp_cae8cfbe-5644-45a5-92e6-0ce169cc47c5/extract-content/0.log" Dec 06 04:54:34 crc kubenswrapper[4802]: I1206 04:54:34.183681 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rnrvp_cae8cfbe-5644-45a5-92e6-0ce169cc47c5/extract-content/0.log" Dec 06 04:54:34 crc kubenswrapper[4802]: I1206 04:54:34.346382 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rnrvp_cae8cfbe-5644-45a5-92e6-0ce169cc47c5/extract-utilities/0.log" Dec 06 04:54:34 crc kubenswrapper[4802]: I1206 04:54:34.354679 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rnrvp_cae8cfbe-5644-45a5-92e6-0ce169cc47c5/extract-content/0.log" Dec 06 04:54:34 crc kubenswrapper[4802]: I1206 04:54:34.622614 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-wxqsx_cffba732-050e-470a-9876-5954b2854ba5/marketplace-operator/0.log" Dec 06 04:54:34 crc kubenswrapper[4802]: I1206 04:54:34.706405 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2l4qq_199a835e-3d62-42ef-8df0-6f087ae5586e/extract-utilities/0.log" Dec 06 04:54:34 crc kubenswrapper[4802]: I1206 04:54:34.884733 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2l4qq_199a835e-3d62-42ef-8df0-6f087ae5586e/extract-content/0.log" Dec 06 04:54:34 crc kubenswrapper[4802]: I1206 04:54:34.952077 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2l4qq_199a835e-3d62-42ef-8df0-6f087ae5586e/extract-utilities/0.log" Dec 06 04:54:35 crc kubenswrapper[4802]: I1206 04:54:35.030247 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2l4qq_199a835e-3d62-42ef-8df0-6f087ae5586e/extract-content/0.log" Dec 06 04:54:35 crc kubenswrapper[4802]: I1206 04:54:35.064086 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rnrvp_cae8cfbe-5644-45a5-92e6-0ce169cc47c5/registry-server/0.log" Dec 06 04:54:35 crc kubenswrapper[4802]: I1206 04:54:35.196396 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2l4qq_199a835e-3d62-42ef-8df0-6f087ae5586e/extract-utilities/0.log" Dec 06 04:54:35 crc kubenswrapper[4802]: I1206 04:54:35.200328 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2l4qq_199a835e-3d62-42ef-8df0-6f087ae5586e/extract-content/0.log" Dec 06 04:54:35 crc kubenswrapper[4802]: I1206 04:54:35.551174 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2l4qq_199a835e-3d62-42ef-8df0-6f087ae5586e/registry-server/0.log" Dec 06 04:54:35 crc kubenswrapper[4802]: I1206 04:54:35.556950 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-ldbdv_a640d4db-76d9-4c21-b394-cbc18a3fe2c4/extract-utilities/0.log" Dec 06 04:54:35 crc kubenswrapper[4802]: I1206 04:54:35.725707 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-ldbdv_a640d4db-76d9-4c21-b394-cbc18a3fe2c4/extract-utilities/0.log" Dec 06 04:54:35 crc kubenswrapper[4802]: I1206 04:54:35.730118 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-ldbdv_a640d4db-76d9-4c21-b394-cbc18a3fe2c4/extract-content/0.log" Dec 06 04:54:35 crc kubenswrapper[4802]: I1206 04:54:35.762198 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-ldbdv_a640d4db-76d9-4c21-b394-cbc18a3fe2c4/extract-content/0.log" Dec 06 04:54:35 crc kubenswrapper[4802]: I1206 04:54:35.942736 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-ldbdv_a640d4db-76d9-4c21-b394-cbc18a3fe2c4/extract-content/0.log" Dec 06 04:54:35 crc kubenswrapper[4802]: I1206 04:54:35.999648 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-ldbdv_a640d4db-76d9-4c21-b394-cbc18a3fe2c4/extract-utilities/0.log" Dec 06 04:54:36 crc kubenswrapper[4802]: I1206 04:54:36.502804 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-ldbdv_a640d4db-76d9-4c21-b394-cbc18a3fe2c4/registry-server/0.log" Dec 06 04:54:38 crc kubenswrapper[4802]: I1206 04:54:38.449899 4802 scope.go:117] "RemoveContainer" containerID="4334df6618bec8f6cedfb779435b2c4d4e0f7e149dd426e245ab29892f475208" Dec 06 04:54:38 crc kubenswrapper[4802]: E1206 04:54:38.450557 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:54:50 crc kubenswrapper[4802]: I1206 04:54:50.450929 4802 scope.go:117] "RemoveContainer" containerID="4334df6618bec8f6cedfb779435b2c4d4e0f7e149dd426e245ab29892f475208" Dec 06 04:54:50 crc kubenswrapper[4802]: E1206 04:54:50.451848 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:55:02 crc kubenswrapper[4802]: I1206 04:55:02.450339 4802 scope.go:117] "RemoveContainer" containerID="4334df6618bec8f6cedfb779435b2c4d4e0f7e149dd426e245ab29892f475208" Dec 06 04:55:02 crc kubenswrapper[4802]: E1206 04:55:02.451096 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:55:17 crc kubenswrapper[4802]: I1206 04:55:17.460311 4802 scope.go:117] "RemoveContainer" containerID="4334df6618bec8f6cedfb779435b2c4d4e0f7e149dd426e245ab29892f475208" Dec 06 04:55:17 crc kubenswrapper[4802]: E1206 04:55:17.461140 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:55:30 crc kubenswrapper[4802]: I1206 04:55:30.451256 4802 scope.go:117] "RemoveContainer" containerID="4334df6618bec8f6cedfb779435b2c4d4e0f7e149dd426e245ab29892f475208" Dec 06 04:55:30 crc kubenswrapper[4802]: E1206 04:55:30.454977 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:55:41 crc kubenswrapper[4802]: I1206 04:55:41.450573 4802 scope.go:117] "RemoveContainer" containerID="4334df6618bec8f6cedfb779435b2c4d4e0f7e149dd426e245ab29892f475208" Dec 06 04:55:41 crc kubenswrapper[4802]: E1206 04:55:41.451301 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:55:52 crc kubenswrapper[4802]: I1206 04:55:52.449935 4802 scope.go:117] "RemoveContainer" containerID="4334df6618bec8f6cedfb779435b2c4d4e0f7e149dd426e245ab29892f475208" Dec 06 04:55:52 crc kubenswrapper[4802]: E1206 04:55:52.450768 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:56:07 crc kubenswrapper[4802]: I1206 04:56:07.456164 4802 scope.go:117] "RemoveContainer" containerID="4334df6618bec8f6cedfb779435b2c4d4e0f7e149dd426e245ab29892f475208" Dec 06 04:56:07 crc kubenswrapper[4802]: E1206 04:56:07.456780 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:56:22 crc kubenswrapper[4802]: I1206 04:56:22.449862 4802 scope.go:117] "RemoveContainer" containerID="4334df6618bec8f6cedfb779435b2c4d4e0f7e149dd426e245ab29892f475208" Dec 06 04:56:22 crc kubenswrapper[4802]: E1206 04:56:22.450638 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:56:27 crc kubenswrapper[4802]: I1206 04:56:27.783101 4802 generic.go:334] "Generic (PLEG): container finished" podID="034bdceb-c945-4869-b588-07cc2af48473" containerID="e8c6bbf7dd78397044bc7b821db5f0e43d3aff5aa973687df30eb73efea3d7a8" exitCode=0 Dec 06 04:56:27 crc kubenswrapper[4802]: I1206 04:56:27.783181 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jw8qc/must-gather-k7q7z" event={"ID":"034bdceb-c945-4869-b588-07cc2af48473","Type":"ContainerDied","Data":"e8c6bbf7dd78397044bc7b821db5f0e43d3aff5aa973687df30eb73efea3d7a8"} Dec 06 04:56:27 crc kubenswrapper[4802]: I1206 04:56:27.784324 4802 scope.go:117] "RemoveContainer" containerID="e8c6bbf7dd78397044bc7b821db5f0e43d3aff5aa973687df30eb73efea3d7a8" Dec 06 04:56:27 crc kubenswrapper[4802]: I1206 04:56:27.962732 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-jw8qc_must-gather-k7q7z_034bdceb-c945-4869-b588-07cc2af48473/gather/0.log" Dec 06 04:56:32 crc kubenswrapper[4802]: I1206 04:56:32.576586 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bfnbz"] Dec 06 04:56:32 crc kubenswrapper[4802]: E1206 04:56:32.577640 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab924c08-8bf0-4f00-8798-866cbdfb04ba" containerName="registry-server" Dec 06 04:56:32 crc kubenswrapper[4802]: I1206 04:56:32.578640 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab924c08-8bf0-4f00-8798-866cbdfb04ba" containerName="registry-server" Dec 06 04:56:32 crc kubenswrapper[4802]: E1206 04:56:32.578679 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab924c08-8bf0-4f00-8798-866cbdfb04ba" containerName="extract-content" Dec 06 04:56:32 crc kubenswrapper[4802]: I1206 04:56:32.578693 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab924c08-8bf0-4f00-8798-866cbdfb04ba" containerName="extract-content" Dec 06 04:56:32 crc kubenswrapper[4802]: E1206 04:56:32.578768 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab924c08-8bf0-4f00-8798-866cbdfb04ba" containerName="extract-utilities" Dec 06 04:56:32 crc kubenswrapper[4802]: I1206 04:56:32.578783 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab924c08-8bf0-4f00-8798-866cbdfb04ba" containerName="extract-utilities" Dec 06 04:56:32 crc kubenswrapper[4802]: E1206 04:56:32.578799 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a7427d8-b298-4e7f-bcd4-b36d8787aa07" containerName="extract-utilities" Dec 06 04:56:32 crc kubenswrapper[4802]: I1206 04:56:32.578807 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a7427d8-b298-4e7f-bcd4-b36d8787aa07" containerName="extract-utilities" Dec 06 04:56:32 crc kubenswrapper[4802]: E1206 04:56:32.578825 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a7427d8-b298-4e7f-bcd4-b36d8787aa07" containerName="registry-server" Dec 06 04:56:32 crc kubenswrapper[4802]: I1206 04:56:32.578834 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a7427d8-b298-4e7f-bcd4-b36d8787aa07" containerName="registry-server" Dec 06 04:56:32 crc kubenswrapper[4802]: E1206 04:56:32.578871 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a7427d8-b298-4e7f-bcd4-b36d8787aa07" containerName="extract-content" Dec 06 04:56:32 crc kubenswrapper[4802]: I1206 04:56:32.578880 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a7427d8-b298-4e7f-bcd4-b36d8787aa07" containerName="extract-content" Dec 06 04:56:32 crc kubenswrapper[4802]: I1206 04:56:32.579146 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a7427d8-b298-4e7f-bcd4-b36d8787aa07" containerName="registry-server" Dec 06 04:56:32 crc kubenswrapper[4802]: I1206 04:56:32.579178 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab924c08-8bf0-4f00-8798-866cbdfb04ba" containerName="registry-server" Dec 06 04:56:32 crc kubenswrapper[4802]: I1206 04:56:32.580943 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bfnbz" Dec 06 04:56:32 crc kubenswrapper[4802]: I1206 04:56:32.591429 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bfnbz"] Dec 06 04:56:32 crc kubenswrapper[4802]: I1206 04:56:32.598726 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxvmb\" (UniqueName: \"kubernetes.io/projected/7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb-kube-api-access-rxvmb\") pod \"redhat-marketplace-bfnbz\" (UID: \"7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb\") " pod="openshift-marketplace/redhat-marketplace-bfnbz" Dec 06 04:56:32 crc kubenswrapper[4802]: I1206 04:56:32.598854 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb-utilities\") pod \"redhat-marketplace-bfnbz\" (UID: \"7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb\") " pod="openshift-marketplace/redhat-marketplace-bfnbz" Dec 06 04:56:32 crc kubenswrapper[4802]: I1206 04:56:32.599011 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb-catalog-content\") pod \"redhat-marketplace-bfnbz\" (UID: \"7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb\") " pod="openshift-marketplace/redhat-marketplace-bfnbz" Dec 06 04:56:32 crc kubenswrapper[4802]: I1206 04:56:32.700886 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb-catalog-content\") pod \"redhat-marketplace-bfnbz\" (UID: \"7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb\") " pod="openshift-marketplace/redhat-marketplace-bfnbz" Dec 06 04:56:32 crc kubenswrapper[4802]: I1206 04:56:32.701117 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxvmb\" (UniqueName: \"kubernetes.io/projected/7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb-kube-api-access-rxvmb\") pod \"redhat-marketplace-bfnbz\" (UID: \"7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb\") " pod="openshift-marketplace/redhat-marketplace-bfnbz" Dec 06 04:56:32 crc kubenswrapper[4802]: I1206 04:56:32.701247 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb-utilities\") pod \"redhat-marketplace-bfnbz\" (UID: \"7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb\") " pod="openshift-marketplace/redhat-marketplace-bfnbz" Dec 06 04:56:32 crc kubenswrapper[4802]: I1206 04:56:32.701678 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb-catalog-content\") pod \"redhat-marketplace-bfnbz\" (UID: \"7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb\") " pod="openshift-marketplace/redhat-marketplace-bfnbz" Dec 06 04:56:32 crc kubenswrapper[4802]: I1206 04:56:32.701887 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb-utilities\") pod \"redhat-marketplace-bfnbz\" (UID: \"7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb\") " pod="openshift-marketplace/redhat-marketplace-bfnbz" Dec 06 04:56:32 crc kubenswrapper[4802]: I1206 04:56:32.856632 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxvmb\" (UniqueName: \"kubernetes.io/projected/7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb-kube-api-access-rxvmb\") pod \"redhat-marketplace-bfnbz\" (UID: \"7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb\") " pod="openshift-marketplace/redhat-marketplace-bfnbz" Dec 06 04:56:32 crc kubenswrapper[4802]: I1206 04:56:32.911495 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bfnbz" Dec 06 04:56:33 crc kubenswrapper[4802]: I1206 04:56:33.415991 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bfnbz"] Dec 06 04:56:33 crc kubenswrapper[4802]: I1206 04:56:33.861951 4802 generic.go:334] "Generic (PLEG): container finished" podID="7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb" containerID="96da8e19084aa689fc3f2617b050bbd172c1180212170b354df5ff32107076e9" exitCode=0 Dec 06 04:56:33 crc kubenswrapper[4802]: I1206 04:56:33.862148 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bfnbz" event={"ID":"7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb","Type":"ContainerDied","Data":"96da8e19084aa689fc3f2617b050bbd172c1180212170b354df5ff32107076e9"} Dec 06 04:56:33 crc kubenswrapper[4802]: I1206 04:56:33.862284 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bfnbz" event={"ID":"7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb","Type":"ContainerStarted","Data":"7d0ce48c8e9027ee2f919951ca521d180a74a1031b8315f0028eb0a165e7508a"} Dec 06 04:56:34 crc kubenswrapper[4802]: I1206 04:56:34.873254 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bfnbz" event={"ID":"7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb","Type":"ContainerStarted","Data":"95df8f05cca2c7bc63fc4528e4b00c648e47d8b7f982454835e504d67afe8eca"} Dec 06 04:56:35 crc kubenswrapper[4802]: I1206 04:56:35.884468 4802 generic.go:334] "Generic (PLEG): container finished" podID="7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb" containerID="95df8f05cca2c7bc63fc4528e4b00c648e47d8b7f982454835e504d67afe8eca" exitCode=0 Dec 06 04:56:35 crc kubenswrapper[4802]: I1206 04:56:35.884692 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bfnbz" event={"ID":"7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb","Type":"ContainerDied","Data":"95df8f05cca2c7bc63fc4528e4b00c648e47d8b7f982454835e504d67afe8eca"} Dec 06 04:56:37 crc kubenswrapper[4802]: I1206 04:56:37.455549 4802 scope.go:117] "RemoveContainer" containerID="4334df6618bec8f6cedfb779435b2c4d4e0f7e149dd426e245ab29892f475208" Dec 06 04:56:37 crc kubenswrapper[4802]: E1206 04:56:37.456220 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:56:37 crc kubenswrapper[4802]: I1206 04:56:37.900895 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bfnbz" event={"ID":"7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb","Type":"ContainerStarted","Data":"91217714980d1f938c694595ae59cd504a8cf52df25ffa9f45faf2ad3ae6e7da"} Dec 06 04:56:37 crc kubenswrapper[4802]: I1206 04:56:37.920214 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bfnbz" podStartSLOduration=3.061044236 podStartE2EDuration="5.920197972s" podCreationTimestamp="2025-12-06 04:56:32 +0000 UTC" firstStartedPulling="2025-12-06 04:56:33.864080972 +0000 UTC m=+4586.735990124" lastFinishedPulling="2025-12-06 04:56:36.723234668 +0000 UTC m=+4589.595143860" observedRunningTime="2025-12-06 04:56:37.918639029 +0000 UTC m=+4590.790548191" watchObservedRunningTime="2025-12-06 04:56:37.920197972 +0000 UTC m=+4590.792107124" Dec 06 04:56:39 crc kubenswrapper[4802]: I1206 04:56:39.324794 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-jw8qc/must-gather-k7q7z"] Dec 06 04:56:39 crc kubenswrapper[4802]: I1206 04:56:39.326671 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-jw8qc/must-gather-k7q7z" podUID="034bdceb-c945-4869-b588-07cc2af48473" containerName="copy" containerID="cri-o://2267a36845a696049e85e413be808c02a8110427e541346563bca457919c365c" gracePeriod=2 Dec 06 04:56:39 crc kubenswrapper[4802]: I1206 04:56:39.332891 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-jw8qc/must-gather-k7q7z"] Dec 06 04:56:39 crc kubenswrapper[4802]: I1206 04:56:39.757543 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-jw8qc_must-gather-k7q7z_034bdceb-c945-4869-b588-07cc2af48473/copy/0.log" Dec 06 04:56:39 crc kubenswrapper[4802]: I1206 04:56:39.758270 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jw8qc/must-gather-k7q7z" Dec 06 04:56:39 crc kubenswrapper[4802]: I1206 04:56:39.890475 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lp8vm\" (UniqueName: \"kubernetes.io/projected/034bdceb-c945-4869-b588-07cc2af48473-kube-api-access-lp8vm\") pod \"034bdceb-c945-4869-b588-07cc2af48473\" (UID: \"034bdceb-c945-4869-b588-07cc2af48473\") " Dec 06 04:56:39 crc kubenswrapper[4802]: I1206 04:56:39.890548 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/034bdceb-c945-4869-b588-07cc2af48473-must-gather-output\") pod \"034bdceb-c945-4869-b588-07cc2af48473\" (UID: \"034bdceb-c945-4869-b588-07cc2af48473\") " Dec 06 04:56:39 crc kubenswrapper[4802]: I1206 04:56:39.903343 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/034bdceb-c945-4869-b588-07cc2af48473-kube-api-access-lp8vm" (OuterVolumeSpecName: "kube-api-access-lp8vm") pod "034bdceb-c945-4869-b588-07cc2af48473" (UID: "034bdceb-c945-4869-b588-07cc2af48473"). InnerVolumeSpecName "kube-api-access-lp8vm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:56:39 crc kubenswrapper[4802]: I1206 04:56:39.942418 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-jw8qc_must-gather-k7q7z_034bdceb-c945-4869-b588-07cc2af48473/copy/0.log" Dec 06 04:56:39 crc kubenswrapper[4802]: I1206 04:56:39.955641 4802 generic.go:334] "Generic (PLEG): container finished" podID="034bdceb-c945-4869-b588-07cc2af48473" containerID="2267a36845a696049e85e413be808c02a8110427e541346563bca457919c365c" exitCode=143 Dec 06 04:56:39 crc kubenswrapper[4802]: I1206 04:56:39.955714 4802 scope.go:117] "RemoveContainer" containerID="2267a36845a696049e85e413be808c02a8110427e541346563bca457919c365c" Dec 06 04:56:39 crc kubenswrapper[4802]: I1206 04:56:39.955986 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jw8qc/must-gather-k7q7z" Dec 06 04:56:39 crc kubenswrapper[4802]: I1206 04:56:39.993061 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lp8vm\" (UniqueName: \"kubernetes.io/projected/034bdceb-c945-4869-b588-07cc2af48473-kube-api-access-lp8vm\") on node \"crc\" DevicePath \"\"" Dec 06 04:56:39 crc kubenswrapper[4802]: I1206 04:56:39.996982 4802 scope.go:117] "RemoveContainer" containerID="e8c6bbf7dd78397044bc7b821db5f0e43d3aff5aa973687df30eb73efea3d7a8" Dec 06 04:56:40 crc kubenswrapper[4802]: I1206 04:56:40.088320 4802 scope.go:117] "RemoveContainer" containerID="2267a36845a696049e85e413be808c02a8110427e541346563bca457919c365c" Dec 06 04:56:40 crc kubenswrapper[4802]: E1206 04:56:40.088889 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2267a36845a696049e85e413be808c02a8110427e541346563bca457919c365c\": container with ID starting with 2267a36845a696049e85e413be808c02a8110427e541346563bca457919c365c not found: ID does not exist" containerID="2267a36845a696049e85e413be808c02a8110427e541346563bca457919c365c" Dec 06 04:56:40 crc kubenswrapper[4802]: I1206 04:56:40.088934 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2267a36845a696049e85e413be808c02a8110427e541346563bca457919c365c"} err="failed to get container status \"2267a36845a696049e85e413be808c02a8110427e541346563bca457919c365c\": rpc error: code = NotFound desc = could not find container \"2267a36845a696049e85e413be808c02a8110427e541346563bca457919c365c\": container with ID starting with 2267a36845a696049e85e413be808c02a8110427e541346563bca457919c365c not found: ID does not exist" Dec 06 04:56:40 crc kubenswrapper[4802]: I1206 04:56:40.088962 4802 scope.go:117] "RemoveContainer" containerID="e8c6bbf7dd78397044bc7b821db5f0e43d3aff5aa973687df30eb73efea3d7a8" Dec 06 04:56:40 crc kubenswrapper[4802]: E1206 04:56:40.089261 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8c6bbf7dd78397044bc7b821db5f0e43d3aff5aa973687df30eb73efea3d7a8\": container with ID starting with e8c6bbf7dd78397044bc7b821db5f0e43d3aff5aa973687df30eb73efea3d7a8 not found: ID does not exist" containerID="e8c6bbf7dd78397044bc7b821db5f0e43d3aff5aa973687df30eb73efea3d7a8" Dec 06 04:56:40 crc kubenswrapper[4802]: I1206 04:56:40.089311 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8c6bbf7dd78397044bc7b821db5f0e43d3aff5aa973687df30eb73efea3d7a8"} err="failed to get container status \"e8c6bbf7dd78397044bc7b821db5f0e43d3aff5aa973687df30eb73efea3d7a8\": rpc error: code = NotFound desc = could not find container \"e8c6bbf7dd78397044bc7b821db5f0e43d3aff5aa973687df30eb73efea3d7a8\": container with ID starting with e8c6bbf7dd78397044bc7b821db5f0e43d3aff5aa973687df30eb73efea3d7a8 not found: ID does not exist" Dec 06 04:56:40 crc kubenswrapper[4802]: I1206 04:56:40.174566 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/034bdceb-c945-4869-b588-07cc2af48473-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "034bdceb-c945-4869-b588-07cc2af48473" (UID: "034bdceb-c945-4869-b588-07cc2af48473"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:56:40 crc kubenswrapper[4802]: I1206 04:56:40.198560 4802 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/034bdceb-c945-4869-b588-07cc2af48473-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 06 04:56:41 crc kubenswrapper[4802]: I1206 04:56:41.462710 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="034bdceb-c945-4869-b588-07cc2af48473" path="/var/lib/kubelet/pods/034bdceb-c945-4869-b588-07cc2af48473/volumes" Dec 06 04:56:42 crc kubenswrapper[4802]: I1206 04:56:42.912225 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bfnbz" Dec 06 04:56:42 crc kubenswrapper[4802]: I1206 04:56:42.912563 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bfnbz" Dec 06 04:56:42 crc kubenswrapper[4802]: I1206 04:56:42.972255 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bfnbz" Dec 06 04:56:43 crc kubenswrapper[4802]: I1206 04:56:43.046532 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bfnbz" Dec 06 04:56:43 crc kubenswrapper[4802]: I1206 04:56:43.152201 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4zzmt"] Dec 06 04:56:43 crc kubenswrapper[4802]: E1206 04:56:43.152691 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="034bdceb-c945-4869-b588-07cc2af48473" containerName="gather" Dec 06 04:56:43 crc kubenswrapper[4802]: I1206 04:56:43.152711 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="034bdceb-c945-4869-b588-07cc2af48473" containerName="gather" Dec 06 04:56:43 crc kubenswrapper[4802]: E1206 04:56:43.152768 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="034bdceb-c945-4869-b588-07cc2af48473" containerName="copy" Dec 06 04:56:43 crc kubenswrapper[4802]: I1206 04:56:43.152780 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="034bdceb-c945-4869-b588-07cc2af48473" containerName="copy" Dec 06 04:56:43 crc kubenswrapper[4802]: I1206 04:56:43.153037 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="034bdceb-c945-4869-b588-07cc2af48473" containerName="copy" Dec 06 04:56:43 crc kubenswrapper[4802]: I1206 04:56:43.153070 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="034bdceb-c945-4869-b588-07cc2af48473" containerName="gather" Dec 06 04:56:43 crc kubenswrapper[4802]: I1206 04:56:43.155111 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4zzmt" Dec 06 04:56:43 crc kubenswrapper[4802]: I1206 04:56:43.174015 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4zzmt"] Dec 06 04:56:43 crc kubenswrapper[4802]: I1206 04:56:43.259544 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9rtz\" (UniqueName: \"kubernetes.io/projected/db6eb539-4ce7-457b-8739-e412cbb56cdd-kube-api-access-j9rtz\") pod \"redhat-operators-4zzmt\" (UID: \"db6eb539-4ce7-457b-8739-e412cbb56cdd\") " pod="openshift-marketplace/redhat-operators-4zzmt" Dec 06 04:56:43 crc kubenswrapper[4802]: I1206 04:56:43.259627 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db6eb539-4ce7-457b-8739-e412cbb56cdd-catalog-content\") pod \"redhat-operators-4zzmt\" (UID: \"db6eb539-4ce7-457b-8739-e412cbb56cdd\") " pod="openshift-marketplace/redhat-operators-4zzmt" Dec 06 04:56:43 crc kubenswrapper[4802]: I1206 04:56:43.259657 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db6eb539-4ce7-457b-8739-e412cbb56cdd-utilities\") pod \"redhat-operators-4zzmt\" (UID: \"db6eb539-4ce7-457b-8739-e412cbb56cdd\") " pod="openshift-marketplace/redhat-operators-4zzmt" Dec 06 04:56:43 crc kubenswrapper[4802]: I1206 04:56:43.361554 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9rtz\" (UniqueName: \"kubernetes.io/projected/db6eb539-4ce7-457b-8739-e412cbb56cdd-kube-api-access-j9rtz\") pod \"redhat-operators-4zzmt\" (UID: \"db6eb539-4ce7-457b-8739-e412cbb56cdd\") " pod="openshift-marketplace/redhat-operators-4zzmt" Dec 06 04:56:43 crc kubenswrapper[4802]: I1206 04:56:43.361653 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db6eb539-4ce7-457b-8739-e412cbb56cdd-catalog-content\") pod \"redhat-operators-4zzmt\" (UID: \"db6eb539-4ce7-457b-8739-e412cbb56cdd\") " pod="openshift-marketplace/redhat-operators-4zzmt" Dec 06 04:56:43 crc kubenswrapper[4802]: I1206 04:56:43.361683 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db6eb539-4ce7-457b-8739-e412cbb56cdd-utilities\") pod \"redhat-operators-4zzmt\" (UID: \"db6eb539-4ce7-457b-8739-e412cbb56cdd\") " pod="openshift-marketplace/redhat-operators-4zzmt" Dec 06 04:56:43 crc kubenswrapper[4802]: I1206 04:56:43.362184 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db6eb539-4ce7-457b-8739-e412cbb56cdd-utilities\") pod \"redhat-operators-4zzmt\" (UID: \"db6eb539-4ce7-457b-8739-e412cbb56cdd\") " pod="openshift-marketplace/redhat-operators-4zzmt" Dec 06 04:56:43 crc kubenswrapper[4802]: I1206 04:56:43.362318 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db6eb539-4ce7-457b-8739-e412cbb56cdd-catalog-content\") pod \"redhat-operators-4zzmt\" (UID: \"db6eb539-4ce7-457b-8739-e412cbb56cdd\") " pod="openshift-marketplace/redhat-operators-4zzmt" Dec 06 04:56:43 crc kubenswrapper[4802]: I1206 04:56:43.387502 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9rtz\" (UniqueName: \"kubernetes.io/projected/db6eb539-4ce7-457b-8739-e412cbb56cdd-kube-api-access-j9rtz\") pod \"redhat-operators-4zzmt\" (UID: \"db6eb539-4ce7-457b-8739-e412cbb56cdd\") " pod="openshift-marketplace/redhat-operators-4zzmt" Dec 06 04:56:43 crc kubenswrapper[4802]: I1206 04:56:43.492383 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4zzmt" Dec 06 04:56:44 crc kubenswrapper[4802]: I1206 04:56:44.622646 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4zzmt"] Dec 06 04:56:45 crc kubenswrapper[4802]: I1206 04:56:44.999831 4802 generic.go:334] "Generic (PLEG): container finished" podID="db6eb539-4ce7-457b-8739-e412cbb56cdd" containerID="ca63fa0525c7464f5ce767d466642f5b053a4e3e6287c2d70e1468abc5bc54d3" exitCode=0 Dec 06 04:56:45 crc kubenswrapper[4802]: I1206 04:56:45.000162 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4zzmt" event={"ID":"db6eb539-4ce7-457b-8739-e412cbb56cdd","Type":"ContainerDied","Data":"ca63fa0525c7464f5ce767d466642f5b053a4e3e6287c2d70e1468abc5bc54d3"} Dec 06 04:56:45 crc kubenswrapper[4802]: I1206 04:56:45.000193 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4zzmt" event={"ID":"db6eb539-4ce7-457b-8739-e412cbb56cdd","Type":"ContainerStarted","Data":"fa1046b1a630b5af2bf9eed54faede89d2602feb0bbe2ccdf937a25d3fe5a9fa"} Dec 06 04:56:45 crc kubenswrapper[4802]: I1206 04:56:45.545174 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bfnbz"] Dec 06 04:56:45 crc kubenswrapper[4802]: I1206 04:56:45.546047 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bfnbz" podUID="7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb" containerName="registry-server" containerID="cri-o://91217714980d1f938c694595ae59cd504a8cf52df25ffa9f45faf2ad3ae6e7da" gracePeriod=2 Dec 06 04:56:45 crc kubenswrapper[4802]: E1206 04:56:45.738332 4802 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7d2c7b3e_fe7b_424b_b4d7_0d497512e9fb.slice/crio-91217714980d1f938c694595ae59cd504a8cf52df25ffa9f45faf2ad3ae6e7da.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7d2c7b3e_fe7b_424b_b4d7_0d497512e9fb.slice/crio-conmon-91217714980d1f938c694595ae59cd504a8cf52df25ffa9f45faf2ad3ae6e7da.scope\": RecentStats: unable to find data in memory cache]" Dec 06 04:56:46 crc kubenswrapper[4802]: I1206 04:56:46.014175 4802 generic.go:334] "Generic (PLEG): container finished" podID="7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb" containerID="91217714980d1f938c694595ae59cd504a8cf52df25ffa9f45faf2ad3ae6e7da" exitCode=0 Dec 06 04:56:46 crc kubenswrapper[4802]: I1206 04:56:46.014260 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bfnbz" event={"ID":"7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb","Type":"ContainerDied","Data":"91217714980d1f938c694595ae59cd504a8cf52df25ffa9f45faf2ad3ae6e7da"} Dec 06 04:56:46 crc kubenswrapper[4802]: I1206 04:56:46.014572 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bfnbz" event={"ID":"7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb","Type":"ContainerDied","Data":"7d0ce48c8e9027ee2f919951ca521d180a74a1031b8315f0028eb0a165e7508a"} Dec 06 04:56:46 crc kubenswrapper[4802]: I1206 04:56:46.014591 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7d0ce48c8e9027ee2f919951ca521d180a74a1031b8315f0028eb0a165e7508a" Dec 06 04:56:46 crc kubenswrapper[4802]: I1206 04:56:46.024506 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bfnbz" Dec 06 04:56:46 crc kubenswrapper[4802]: I1206 04:56:46.118161 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb-catalog-content\") pod \"7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb\" (UID: \"7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb\") " Dec 06 04:56:46 crc kubenswrapper[4802]: I1206 04:56:46.118326 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rxvmb\" (UniqueName: \"kubernetes.io/projected/7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb-kube-api-access-rxvmb\") pod \"7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb\" (UID: \"7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb\") " Dec 06 04:56:46 crc kubenswrapper[4802]: I1206 04:56:46.118389 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb-utilities\") pod \"7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb\" (UID: \"7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb\") " Dec 06 04:56:46 crc kubenswrapper[4802]: I1206 04:56:46.119371 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb-utilities" (OuterVolumeSpecName: "utilities") pod "7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb" (UID: "7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:56:46 crc kubenswrapper[4802]: I1206 04:56:46.119677 4802 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 04:56:46 crc kubenswrapper[4802]: I1206 04:56:46.128914 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb-kube-api-access-rxvmb" (OuterVolumeSpecName: "kube-api-access-rxvmb") pod "7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb" (UID: "7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb"). InnerVolumeSpecName "kube-api-access-rxvmb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:56:46 crc kubenswrapper[4802]: I1206 04:56:46.160429 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb" (UID: "7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:56:46 crc kubenswrapper[4802]: I1206 04:56:46.220990 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rxvmb\" (UniqueName: \"kubernetes.io/projected/7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb-kube-api-access-rxvmb\") on node \"crc\" DevicePath \"\"" Dec 06 04:56:46 crc kubenswrapper[4802]: I1206 04:56:46.221031 4802 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 04:56:47 crc kubenswrapper[4802]: I1206 04:56:47.032244 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bfnbz" Dec 06 04:56:47 crc kubenswrapper[4802]: I1206 04:56:47.032285 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4zzmt" event={"ID":"db6eb539-4ce7-457b-8739-e412cbb56cdd","Type":"ContainerStarted","Data":"af708816762350579c2c14b92c575ccbea1986605b36122debafe77bf8bfe1d2"} Dec 06 04:56:47 crc kubenswrapper[4802]: I1206 04:56:47.073409 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bfnbz"] Dec 06 04:56:47 crc kubenswrapper[4802]: I1206 04:56:47.081925 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bfnbz"] Dec 06 04:56:47 crc kubenswrapper[4802]: I1206 04:56:47.462985 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb" path="/var/lib/kubelet/pods/7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb/volumes" Dec 06 04:56:48 crc kubenswrapper[4802]: I1206 04:56:48.044312 4802 generic.go:334] "Generic (PLEG): container finished" podID="db6eb539-4ce7-457b-8739-e412cbb56cdd" containerID="af708816762350579c2c14b92c575ccbea1986605b36122debafe77bf8bfe1d2" exitCode=0 Dec 06 04:56:48 crc kubenswrapper[4802]: I1206 04:56:48.044360 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4zzmt" event={"ID":"db6eb539-4ce7-457b-8739-e412cbb56cdd","Type":"ContainerDied","Data":"af708816762350579c2c14b92c575ccbea1986605b36122debafe77bf8bfe1d2"} Dec 06 04:56:49 crc kubenswrapper[4802]: I1206 04:56:49.056649 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4zzmt" event={"ID":"db6eb539-4ce7-457b-8739-e412cbb56cdd","Type":"ContainerStarted","Data":"f85495df4ee902654d82fb45ec62904d4dff881eabe34f5222345a969f4616c5"} Dec 06 04:56:49 crc kubenswrapper[4802]: I1206 04:56:49.086121 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4zzmt" podStartSLOduration=2.636307445 podStartE2EDuration="6.08609434s" podCreationTimestamp="2025-12-06 04:56:43 +0000 UTC" firstStartedPulling="2025-12-06 04:56:45.002158453 +0000 UTC m=+4597.874067615" lastFinishedPulling="2025-12-06 04:56:48.451945338 +0000 UTC m=+4601.323854510" observedRunningTime="2025-12-06 04:56:49.072069668 +0000 UTC m=+4601.943978820" watchObservedRunningTime="2025-12-06 04:56:49.08609434 +0000 UTC m=+4601.958003522" Dec 06 04:56:52 crc kubenswrapper[4802]: I1206 04:56:52.450827 4802 scope.go:117] "RemoveContainer" containerID="4334df6618bec8f6cedfb779435b2c4d4e0f7e149dd426e245ab29892f475208" Dec 06 04:56:52 crc kubenswrapper[4802]: E1206 04:56:52.451705 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:56:53 crc kubenswrapper[4802]: I1206 04:56:53.492903 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4zzmt" Dec 06 04:56:53 crc kubenswrapper[4802]: I1206 04:56:53.493525 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4zzmt" Dec 06 04:56:54 crc kubenswrapper[4802]: I1206 04:56:54.551717 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-4zzmt" podUID="db6eb539-4ce7-457b-8739-e412cbb56cdd" containerName="registry-server" probeResult="failure" output=< Dec 06 04:56:54 crc kubenswrapper[4802]: timeout: failed to connect service ":50051" within 1s Dec 06 04:56:54 crc kubenswrapper[4802]: > Dec 06 04:57:03 crc kubenswrapper[4802]: I1206 04:57:03.539406 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4zzmt" Dec 06 04:57:03 crc kubenswrapper[4802]: I1206 04:57:03.590673 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4zzmt" Dec 06 04:57:03 crc kubenswrapper[4802]: I1206 04:57:03.778643 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4zzmt"] Dec 06 04:57:04 crc kubenswrapper[4802]: I1206 04:57:04.450959 4802 scope.go:117] "RemoveContainer" containerID="4334df6618bec8f6cedfb779435b2c4d4e0f7e149dd426e245ab29892f475208" Dec 06 04:57:04 crc kubenswrapper[4802]: E1206 04:57:04.451297 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpxxw_openshift-machine-config-operator(6df38316-e0d3-4018-8d27-3620eba3a68d)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" Dec 06 04:57:05 crc kubenswrapper[4802]: I1206 04:57:05.217283 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-4zzmt" podUID="db6eb539-4ce7-457b-8739-e412cbb56cdd" containerName="registry-server" containerID="cri-o://f85495df4ee902654d82fb45ec62904d4dff881eabe34f5222345a969f4616c5" gracePeriod=2 Dec 06 04:57:05 crc kubenswrapper[4802]: I1206 04:57:05.940062 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4zzmt" Dec 06 04:57:06 crc kubenswrapper[4802]: I1206 04:57:06.034906 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db6eb539-4ce7-457b-8739-e412cbb56cdd-utilities\") pod \"db6eb539-4ce7-457b-8739-e412cbb56cdd\" (UID: \"db6eb539-4ce7-457b-8739-e412cbb56cdd\") " Dec 06 04:57:06 crc kubenswrapper[4802]: I1206 04:57:06.035032 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j9rtz\" (UniqueName: \"kubernetes.io/projected/db6eb539-4ce7-457b-8739-e412cbb56cdd-kube-api-access-j9rtz\") pod \"db6eb539-4ce7-457b-8739-e412cbb56cdd\" (UID: \"db6eb539-4ce7-457b-8739-e412cbb56cdd\") " Dec 06 04:57:06 crc kubenswrapper[4802]: I1206 04:57:06.035192 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db6eb539-4ce7-457b-8739-e412cbb56cdd-catalog-content\") pod \"db6eb539-4ce7-457b-8739-e412cbb56cdd\" (UID: \"db6eb539-4ce7-457b-8739-e412cbb56cdd\") " Dec 06 04:57:06 crc kubenswrapper[4802]: I1206 04:57:06.041251 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db6eb539-4ce7-457b-8739-e412cbb56cdd-kube-api-access-j9rtz" (OuterVolumeSpecName: "kube-api-access-j9rtz") pod "db6eb539-4ce7-457b-8739-e412cbb56cdd" (UID: "db6eb539-4ce7-457b-8739-e412cbb56cdd"). InnerVolumeSpecName "kube-api-access-j9rtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:57:06 crc kubenswrapper[4802]: I1206 04:57:06.046189 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db6eb539-4ce7-457b-8739-e412cbb56cdd-utilities" (OuterVolumeSpecName: "utilities") pod "db6eb539-4ce7-457b-8739-e412cbb56cdd" (UID: "db6eb539-4ce7-457b-8739-e412cbb56cdd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:57:06 crc kubenswrapper[4802]: I1206 04:57:06.137375 4802 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db6eb539-4ce7-457b-8739-e412cbb56cdd-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 04:57:06 crc kubenswrapper[4802]: I1206 04:57:06.137417 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j9rtz\" (UniqueName: \"kubernetes.io/projected/db6eb539-4ce7-457b-8739-e412cbb56cdd-kube-api-access-j9rtz\") on node \"crc\" DevicePath \"\"" Dec 06 04:57:06 crc kubenswrapper[4802]: I1206 04:57:06.147697 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db6eb539-4ce7-457b-8739-e412cbb56cdd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "db6eb539-4ce7-457b-8739-e412cbb56cdd" (UID: "db6eb539-4ce7-457b-8739-e412cbb56cdd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:57:06 crc kubenswrapper[4802]: I1206 04:57:06.226708 4802 generic.go:334] "Generic (PLEG): container finished" podID="db6eb539-4ce7-457b-8739-e412cbb56cdd" containerID="f85495df4ee902654d82fb45ec62904d4dff881eabe34f5222345a969f4616c5" exitCode=0 Dec 06 04:57:06 crc kubenswrapper[4802]: I1206 04:57:06.226761 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4zzmt" event={"ID":"db6eb539-4ce7-457b-8739-e412cbb56cdd","Type":"ContainerDied","Data":"f85495df4ee902654d82fb45ec62904d4dff881eabe34f5222345a969f4616c5"} Dec 06 04:57:06 crc kubenswrapper[4802]: I1206 04:57:06.226793 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4zzmt" event={"ID":"db6eb539-4ce7-457b-8739-e412cbb56cdd","Type":"ContainerDied","Data":"fa1046b1a630b5af2bf9eed54faede89d2602feb0bbe2ccdf937a25d3fe5a9fa"} Dec 06 04:57:06 crc kubenswrapper[4802]: I1206 04:57:06.226811 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4zzmt" Dec 06 04:57:06 crc kubenswrapper[4802]: I1206 04:57:06.226818 4802 scope.go:117] "RemoveContainer" containerID="f85495df4ee902654d82fb45ec62904d4dff881eabe34f5222345a969f4616c5" Dec 06 04:57:06 crc kubenswrapper[4802]: I1206 04:57:06.239467 4802 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db6eb539-4ce7-457b-8739-e412cbb56cdd-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 04:57:06 crc kubenswrapper[4802]: I1206 04:57:06.244460 4802 scope.go:117] "RemoveContainer" containerID="af708816762350579c2c14b92c575ccbea1986605b36122debafe77bf8bfe1d2" Dec 06 04:57:06 crc kubenswrapper[4802]: I1206 04:57:06.257109 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4zzmt"] Dec 06 04:57:06 crc kubenswrapper[4802]: I1206 04:57:06.265048 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-4zzmt"] Dec 06 04:57:06 crc kubenswrapper[4802]: I1206 04:57:06.272044 4802 scope.go:117] "RemoveContainer" containerID="ca63fa0525c7464f5ce767d466642f5b053a4e3e6287c2d70e1468abc5bc54d3" Dec 06 04:57:06 crc kubenswrapper[4802]: I1206 04:57:06.316543 4802 scope.go:117] "RemoveContainer" containerID="f85495df4ee902654d82fb45ec62904d4dff881eabe34f5222345a969f4616c5" Dec 06 04:57:06 crc kubenswrapper[4802]: E1206 04:57:06.317110 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f85495df4ee902654d82fb45ec62904d4dff881eabe34f5222345a969f4616c5\": container with ID starting with f85495df4ee902654d82fb45ec62904d4dff881eabe34f5222345a969f4616c5 not found: ID does not exist" containerID="f85495df4ee902654d82fb45ec62904d4dff881eabe34f5222345a969f4616c5" Dec 06 04:57:06 crc kubenswrapper[4802]: I1206 04:57:06.317157 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f85495df4ee902654d82fb45ec62904d4dff881eabe34f5222345a969f4616c5"} err="failed to get container status \"f85495df4ee902654d82fb45ec62904d4dff881eabe34f5222345a969f4616c5\": rpc error: code = NotFound desc = could not find container \"f85495df4ee902654d82fb45ec62904d4dff881eabe34f5222345a969f4616c5\": container with ID starting with f85495df4ee902654d82fb45ec62904d4dff881eabe34f5222345a969f4616c5 not found: ID does not exist" Dec 06 04:57:06 crc kubenswrapper[4802]: I1206 04:57:06.317181 4802 scope.go:117] "RemoveContainer" containerID="af708816762350579c2c14b92c575ccbea1986605b36122debafe77bf8bfe1d2" Dec 06 04:57:06 crc kubenswrapper[4802]: E1206 04:57:06.317502 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af708816762350579c2c14b92c575ccbea1986605b36122debafe77bf8bfe1d2\": container with ID starting with af708816762350579c2c14b92c575ccbea1986605b36122debafe77bf8bfe1d2 not found: ID does not exist" containerID="af708816762350579c2c14b92c575ccbea1986605b36122debafe77bf8bfe1d2" Dec 06 04:57:06 crc kubenswrapper[4802]: I1206 04:57:06.317540 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af708816762350579c2c14b92c575ccbea1986605b36122debafe77bf8bfe1d2"} err="failed to get container status \"af708816762350579c2c14b92c575ccbea1986605b36122debafe77bf8bfe1d2\": rpc error: code = NotFound desc = could not find container \"af708816762350579c2c14b92c575ccbea1986605b36122debafe77bf8bfe1d2\": container with ID starting with af708816762350579c2c14b92c575ccbea1986605b36122debafe77bf8bfe1d2 not found: ID does not exist" Dec 06 04:57:06 crc kubenswrapper[4802]: I1206 04:57:06.317566 4802 scope.go:117] "RemoveContainer" containerID="ca63fa0525c7464f5ce767d466642f5b053a4e3e6287c2d70e1468abc5bc54d3" Dec 06 04:57:06 crc kubenswrapper[4802]: E1206 04:57:06.317891 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca63fa0525c7464f5ce767d466642f5b053a4e3e6287c2d70e1468abc5bc54d3\": container with ID starting with ca63fa0525c7464f5ce767d466642f5b053a4e3e6287c2d70e1468abc5bc54d3 not found: ID does not exist" containerID="ca63fa0525c7464f5ce767d466642f5b053a4e3e6287c2d70e1468abc5bc54d3" Dec 06 04:57:06 crc kubenswrapper[4802]: I1206 04:57:06.317930 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca63fa0525c7464f5ce767d466642f5b053a4e3e6287c2d70e1468abc5bc54d3"} err="failed to get container status \"ca63fa0525c7464f5ce767d466642f5b053a4e3e6287c2d70e1468abc5bc54d3\": rpc error: code = NotFound desc = could not find container \"ca63fa0525c7464f5ce767d466642f5b053a4e3e6287c2d70e1468abc5bc54d3\": container with ID starting with ca63fa0525c7464f5ce767d466642f5b053a4e3e6287c2d70e1468abc5bc54d3 not found: ID does not exist" Dec 06 04:57:07 crc kubenswrapper[4802]: I1206 04:57:07.464682 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db6eb539-4ce7-457b-8739-e412cbb56cdd" path="/var/lib/kubelet/pods/db6eb539-4ce7-457b-8739-e412cbb56cdd/volumes" Dec 06 04:57:17 crc kubenswrapper[4802]: I1206 04:57:17.455507 4802 scope.go:117] "RemoveContainer" containerID="4334df6618bec8f6cedfb779435b2c4d4e0f7e149dd426e245ab29892f475208" Dec 06 04:57:18 crc kubenswrapper[4802]: I1206 04:57:18.346582 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" event={"ID":"6df38316-e0d3-4018-8d27-3620eba3a68d","Type":"ContainerStarted","Data":"c06e78342e1dde3ef70242489123ba449bf2715fa831392e80666586f7300b06"} Dec 06 04:58:23 crc kubenswrapper[4802]: I1206 04:58:23.616965 4802 scope.go:117] "RemoveContainer" containerID="4546dbf93185c740dcb6d447c4af2e431b1691283328f0b87520f2aa7b441401" Dec 06 04:58:23 crc kubenswrapper[4802]: I1206 04:58:23.675319 4802 scope.go:117] "RemoveContainer" containerID="56125e4880064f244b97c1dcd394f37393054024640494fdfd4c86d4489d152f" Dec 06 04:59:43 crc kubenswrapper[4802]: I1206 04:59:43.283962 4802 patch_prober.go:28] interesting pod/machine-config-daemon-zpxxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:59:43 crc kubenswrapper[4802]: I1206 04:59:43.284509 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 05:00:00 crc kubenswrapper[4802]: I1206 05:00:00.184128 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416620-p7jdk"] Dec 06 05:00:00 crc kubenswrapper[4802]: E1206 05:00:00.201805 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb" containerName="extract-content" Dec 06 05:00:00 crc kubenswrapper[4802]: I1206 05:00:00.201850 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb" containerName="extract-content" Dec 06 05:00:00 crc kubenswrapper[4802]: E1206 05:00:00.201863 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db6eb539-4ce7-457b-8739-e412cbb56cdd" containerName="extract-utilities" Dec 06 05:00:00 crc kubenswrapper[4802]: I1206 05:00:00.201872 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="db6eb539-4ce7-457b-8739-e412cbb56cdd" containerName="extract-utilities" Dec 06 05:00:00 crc kubenswrapper[4802]: E1206 05:00:00.201967 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db6eb539-4ce7-457b-8739-e412cbb56cdd" containerName="registry-server" Dec 06 05:00:00 crc kubenswrapper[4802]: I1206 05:00:00.201976 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="db6eb539-4ce7-457b-8739-e412cbb56cdd" containerName="registry-server" Dec 06 05:00:00 crc kubenswrapper[4802]: E1206 05:00:00.201999 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb" containerName="registry-server" Dec 06 05:00:00 crc kubenswrapper[4802]: I1206 05:00:00.202008 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb" containerName="registry-server" Dec 06 05:00:00 crc kubenswrapper[4802]: E1206 05:00:00.202051 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb" containerName="extract-utilities" Dec 06 05:00:00 crc kubenswrapper[4802]: I1206 05:00:00.202060 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb" containerName="extract-utilities" Dec 06 05:00:00 crc kubenswrapper[4802]: E1206 05:00:00.202095 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db6eb539-4ce7-457b-8739-e412cbb56cdd" containerName="extract-content" Dec 06 05:00:00 crc kubenswrapper[4802]: I1206 05:00:00.202105 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="db6eb539-4ce7-457b-8739-e412cbb56cdd" containerName="extract-content" Dec 06 05:00:00 crc kubenswrapper[4802]: I1206 05:00:00.203247 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="db6eb539-4ce7-457b-8739-e412cbb56cdd" containerName="registry-server" Dec 06 05:00:00 crc kubenswrapper[4802]: I1206 05:00:00.203308 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d2c7b3e-fe7b-424b-b4d7-0d497512e9fb" containerName="registry-server" Dec 06 05:00:00 crc kubenswrapper[4802]: I1206 05:00:00.204780 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416620-p7jdk" Dec 06 05:00:00 crc kubenswrapper[4802]: I1206 05:00:00.208132 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 05:00:00 crc kubenswrapper[4802]: I1206 05:00:00.208586 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 05:00:00 crc kubenswrapper[4802]: I1206 05:00:00.235014 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416620-p7jdk"] Dec 06 05:00:00 crc kubenswrapper[4802]: I1206 05:00:00.299849 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fa55ed76-dfa1-4c6e-9536-d999799c97d3-config-volume\") pod \"collect-profiles-29416620-p7jdk\" (UID: \"fa55ed76-dfa1-4c6e-9536-d999799c97d3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416620-p7jdk" Dec 06 05:00:00 crc kubenswrapper[4802]: I1206 05:00:00.299918 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fa55ed76-dfa1-4c6e-9536-d999799c97d3-secret-volume\") pod \"collect-profiles-29416620-p7jdk\" (UID: \"fa55ed76-dfa1-4c6e-9536-d999799c97d3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416620-p7jdk" Dec 06 05:00:00 crc kubenswrapper[4802]: I1206 05:00:00.300019 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqspf\" (UniqueName: \"kubernetes.io/projected/fa55ed76-dfa1-4c6e-9536-d999799c97d3-kube-api-access-zqspf\") pod \"collect-profiles-29416620-p7jdk\" (UID: \"fa55ed76-dfa1-4c6e-9536-d999799c97d3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416620-p7jdk" Dec 06 05:00:00 crc kubenswrapper[4802]: I1206 05:00:00.401791 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqspf\" (UniqueName: \"kubernetes.io/projected/fa55ed76-dfa1-4c6e-9536-d999799c97d3-kube-api-access-zqspf\") pod \"collect-profiles-29416620-p7jdk\" (UID: \"fa55ed76-dfa1-4c6e-9536-d999799c97d3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416620-p7jdk" Dec 06 05:00:00 crc kubenswrapper[4802]: I1206 05:00:00.402122 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fa55ed76-dfa1-4c6e-9536-d999799c97d3-config-volume\") pod \"collect-profiles-29416620-p7jdk\" (UID: \"fa55ed76-dfa1-4c6e-9536-d999799c97d3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416620-p7jdk" Dec 06 05:00:00 crc kubenswrapper[4802]: I1206 05:00:00.402208 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fa55ed76-dfa1-4c6e-9536-d999799c97d3-secret-volume\") pod \"collect-profiles-29416620-p7jdk\" (UID: \"fa55ed76-dfa1-4c6e-9536-d999799c97d3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416620-p7jdk" Dec 06 05:00:00 crc kubenswrapper[4802]: I1206 05:00:00.403996 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fa55ed76-dfa1-4c6e-9536-d999799c97d3-config-volume\") pod \"collect-profiles-29416620-p7jdk\" (UID: \"fa55ed76-dfa1-4c6e-9536-d999799c97d3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416620-p7jdk" Dec 06 05:00:00 crc kubenswrapper[4802]: I1206 05:00:00.407770 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fa55ed76-dfa1-4c6e-9536-d999799c97d3-secret-volume\") pod \"collect-profiles-29416620-p7jdk\" (UID: \"fa55ed76-dfa1-4c6e-9536-d999799c97d3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416620-p7jdk" Dec 06 05:00:00 crc kubenswrapper[4802]: I1206 05:00:00.424277 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqspf\" (UniqueName: \"kubernetes.io/projected/fa55ed76-dfa1-4c6e-9536-d999799c97d3-kube-api-access-zqspf\") pod \"collect-profiles-29416620-p7jdk\" (UID: \"fa55ed76-dfa1-4c6e-9536-d999799c97d3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416620-p7jdk" Dec 06 05:00:00 crc kubenswrapper[4802]: I1206 05:00:00.534487 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416620-p7jdk" Dec 06 05:00:01 crc kubenswrapper[4802]: I1206 05:00:01.000487 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416620-p7jdk"] Dec 06 05:00:01 crc kubenswrapper[4802]: W1206 05:00:01.005965 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfa55ed76_dfa1_4c6e_9536_d999799c97d3.slice/crio-1325bd1f33c5e8c2f353d55bc2ed5252cc5c804b69ddb8f1f8a777ad74d17c29 WatchSource:0}: Error finding container 1325bd1f33c5e8c2f353d55bc2ed5252cc5c804b69ddb8f1f8a777ad74d17c29: Status 404 returned error can't find the container with id 1325bd1f33c5e8c2f353d55bc2ed5252cc5c804b69ddb8f1f8a777ad74d17c29 Dec 06 05:00:01 crc kubenswrapper[4802]: I1206 05:00:01.192913 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416620-p7jdk" event={"ID":"fa55ed76-dfa1-4c6e-9536-d999799c97d3","Type":"ContainerStarted","Data":"5fa6591e3e9651b9fdb3371d001e3814a930a70d85ee56737d4d4cd80d70fd1d"} Dec 06 05:00:01 crc kubenswrapper[4802]: I1206 05:00:01.192970 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416620-p7jdk" event={"ID":"fa55ed76-dfa1-4c6e-9536-d999799c97d3","Type":"ContainerStarted","Data":"1325bd1f33c5e8c2f353d55bc2ed5252cc5c804b69ddb8f1f8a777ad74d17c29"} Dec 06 05:00:01 crc kubenswrapper[4802]: I1206 05:00:01.212532 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29416620-p7jdk" podStartSLOduration=1.212514192 podStartE2EDuration="1.212514192s" podCreationTimestamp="2025-12-06 05:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:00:01.20843649 +0000 UTC m=+4794.080345662" watchObservedRunningTime="2025-12-06 05:00:01.212514192 +0000 UTC m=+4794.084423344" Dec 06 05:00:02 crc kubenswrapper[4802]: I1206 05:00:02.205287 4802 generic.go:334] "Generic (PLEG): container finished" podID="fa55ed76-dfa1-4c6e-9536-d999799c97d3" containerID="5fa6591e3e9651b9fdb3371d001e3814a930a70d85ee56737d4d4cd80d70fd1d" exitCode=0 Dec 06 05:00:02 crc kubenswrapper[4802]: I1206 05:00:02.206505 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416620-p7jdk" event={"ID":"fa55ed76-dfa1-4c6e-9536-d999799c97d3","Type":"ContainerDied","Data":"5fa6591e3e9651b9fdb3371d001e3814a930a70d85ee56737d4d4cd80d70fd1d"} Dec 06 05:00:03 crc kubenswrapper[4802]: I1206 05:00:03.669042 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416620-p7jdk" Dec 06 05:00:03 crc kubenswrapper[4802]: I1206 05:00:03.788898 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fa55ed76-dfa1-4c6e-9536-d999799c97d3-secret-volume\") pod \"fa55ed76-dfa1-4c6e-9536-d999799c97d3\" (UID: \"fa55ed76-dfa1-4c6e-9536-d999799c97d3\") " Dec 06 05:00:03 crc kubenswrapper[4802]: I1206 05:00:03.789002 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zqspf\" (UniqueName: \"kubernetes.io/projected/fa55ed76-dfa1-4c6e-9536-d999799c97d3-kube-api-access-zqspf\") pod \"fa55ed76-dfa1-4c6e-9536-d999799c97d3\" (UID: \"fa55ed76-dfa1-4c6e-9536-d999799c97d3\") " Dec 06 05:00:03 crc kubenswrapper[4802]: I1206 05:00:03.789177 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fa55ed76-dfa1-4c6e-9536-d999799c97d3-config-volume\") pod \"fa55ed76-dfa1-4c6e-9536-d999799c97d3\" (UID: \"fa55ed76-dfa1-4c6e-9536-d999799c97d3\") " Dec 06 05:00:03 crc kubenswrapper[4802]: I1206 05:00:03.789883 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa55ed76-dfa1-4c6e-9536-d999799c97d3-config-volume" (OuterVolumeSpecName: "config-volume") pod "fa55ed76-dfa1-4c6e-9536-d999799c97d3" (UID: "fa55ed76-dfa1-4c6e-9536-d999799c97d3"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:00:03 crc kubenswrapper[4802]: I1206 05:00:03.795231 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa55ed76-dfa1-4c6e-9536-d999799c97d3-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "fa55ed76-dfa1-4c6e-9536-d999799c97d3" (UID: "fa55ed76-dfa1-4c6e-9536-d999799c97d3"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:00:03 crc kubenswrapper[4802]: I1206 05:00:03.796864 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa55ed76-dfa1-4c6e-9536-d999799c97d3-kube-api-access-zqspf" (OuterVolumeSpecName: "kube-api-access-zqspf") pod "fa55ed76-dfa1-4c6e-9536-d999799c97d3" (UID: "fa55ed76-dfa1-4c6e-9536-d999799c97d3"). InnerVolumeSpecName "kube-api-access-zqspf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:00:03 crc kubenswrapper[4802]: I1206 05:00:03.891975 4802 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fa55ed76-dfa1-4c6e-9536-d999799c97d3-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 05:00:03 crc kubenswrapper[4802]: I1206 05:00:03.892019 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zqspf\" (UniqueName: \"kubernetes.io/projected/fa55ed76-dfa1-4c6e-9536-d999799c97d3-kube-api-access-zqspf\") on node \"crc\" DevicePath \"\"" Dec 06 05:00:03 crc kubenswrapper[4802]: I1206 05:00:03.892031 4802 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fa55ed76-dfa1-4c6e-9536-d999799c97d3-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 05:00:04 crc kubenswrapper[4802]: I1206 05:00:04.248521 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416620-p7jdk" event={"ID":"fa55ed76-dfa1-4c6e-9536-d999799c97d3","Type":"ContainerDied","Data":"1325bd1f33c5e8c2f353d55bc2ed5252cc5c804b69ddb8f1f8a777ad74d17c29"} Dec 06 05:00:04 crc kubenswrapper[4802]: I1206 05:00:04.248594 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1325bd1f33c5e8c2f353d55bc2ed5252cc5c804b69ddb8f1f8a777ad74d17c29" Dec 06 05:00:04 crc kubenswrapper[4802]: I1206 05:00:04.248692 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416620-p7jdk" Dec 06 05:00:04 crc kubenswrapper[4802]: I1206 05:00:04.305808 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416575-fwqzl"] Dec 06 05:00:04 crc kubenswrapper[4802]: I1206 05:00:04.319345 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416575-fwqzl"] Dec 06 05:00:05 crc kubenswrapper[4802]: I1206 05:00:05.474508 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8189b4bc-09c3-4075-b959-bb67bf6ccd34" path="/var/lib/kubelet/pods/8189b4bc-09c3-4075-b959-bb67bf6ccd34/volumes" Dec 06 05:00:13 crc kubenswrapper[4802]: I1206 05:00:13.283736 4802 patch_prober.go:28] interesting pod/machine-config-daemon-zpxxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 05:00:13 crc kubenswrapper[4802]: I1206 05:00:13.284451 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpxxw" podUID="6df38316-e0d3-4018-8d27-3620eba3a68d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 05:00:23 crc kubenswrapper[4802]: I1206 05:00:23.808532 4802 scope.go:117] "RemoveContainer" containerID="be56c3fe487fc382c34351345754aedaa82b6dbfce9e8c1b32005d67202a5a56" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515114734157024455 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015114734160017364 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015114722304016504 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015114722304015454 5ustar corecore